Language of instruction

中文

Current course fees

20,410

Location

Shanghai

X ACADEMY
Global Explorer Summit
-


2024.07.25

Major Course X-Core

Artificial intelligence and natural language processing

Artificial Intelligence and Natural Language Processing

COURSE INTRO

Course Introduction 01

COURSE INTRO

Course Introduction 01

In the past year, we have witnessed the impact of large language models on the world. Models such as ChatGPT, Claude, and Grok have been widely used in various industries, from simplifying daily communication, solving programming problems, accelerating scientific innovation, driving educational reform, to participating in artistic creation and cultural dissemination. Large language models are becoming an important force for social progress. As these models' capabilities continue to advance, we also start to consider: How can we better understand their working principles? How to evaluate their social impact? And how to maximize their positive value while ensuring safety and respecting privacy?

"TechX AI and Natural Language Processing" will uncover the surface for you and explore the application of natural language processing in different fields and the technical support behind it. You will delve into frameworks widely used in the industry (such as autoregressive models) and some product technology analysis currently facing users. At the same time, under the guidance of academic leaders, you will together explore the application of natural language processing in fields such as social sciences, education, and healthcare, applying the knowledge gained to discuss and research the latest technology in natural language processing.

In the past year, we have witnessed the impact of large language models on the world. Models such as ChatGPT, Claude, and Grok have been widely used in various industries, from simplifying daily communication, solving programming problems, accelerating scientific innovation, driving educational reform, to participating in artistic creation and cultural dissemination. Large language models are becoming an important force for social progress. As these models' capabilities continue to advance, we also start to consider: How can we better understand their working principles? How to evaluate their social impact? And how to maximize their positive value while ensuring safety and respecting privacy?

"TechX AI and Natural Language Processing" will uncover the surface for you and explore the application of natural language processing in different fields and the technical support behind it. You will delve into frameworks widely used in the industry (such as autoregressive models) and some product technology analysis currently facing users. At the same time, under the guidance of academic leaders, you will together explore the application of natural language processing in fields such as social sciences, education, and healthcare, applying the knowledge gained to discuss and research the latest technology in natural language processing.

COURSE OUTCOME

Course Outcome 02

COURSE OUTCOME

Course Outcome 02

According to discussion and research, complete and display your research posters or summary writing, or complete your chatbot or natural language processing-based agent applications.


*The final type and content shall be subject to the finished product of the explorer after participating in the vertex project X-Capstone

According to discussion and research, complete and display your research posters or summary writing, or complete your chatbot or natural language processing-based agent applications.


*The final type and content shall be subject to the finished product of the explorer after participating in the vertex project X-Capstone

COURSE CONTENTS

Course Contents 03

COURSE CONTENTS

Course Contents 03

01 Introduction to Natural Language Processing & Machine Learning Basics

Applications of natural language processing, examples; Basics of machine learning and deep learning


02 Static Word Representation & Text Classification

Sparse representation: one-hot, ppmi, SVD.; Dense representation: word2vec; Generative classifier: generative classifier; Naive Bayes: Naive Bayes; Discriminative classifier: discriminative classifier; Sparse features BoW/N-gram; Dense features

Exercises: Word2vec, text classification


03 Language Model Language Modeling (I)

n-gram language model; RNN recurrent neural network and language model; RNN, RNN-LM - LSTM / GRU -

Exercises: RNN language model


04 Language Model Language Modeling & Attention Mechanism (II)

Attention mechanism Attention; RNN LM with Attention

Exercises: Attention mechanism


05 Self-attention mechanism Self-attention & Transformer

Transformer Encoder; Transformer Decoder; Sequence-to-sequence Seq2Seq, concepts of encoder-decoder

Exercises: huggingface tutorial


06 Pretrained Language Model Pretrained LM

Encoder LM (BERT); Decoder-only LM (GPT); Seq2seq LM (T5) -

Exercises: huggingface tutorial


07 Large Language Model (Large Language Model) -

Prompting and prompt learning; Contextual learning In-context Learning; Evolutionary trajectory of GPT series; Basic introduction (GPT-3, GPT3.5, ChatGPT, GPT4...); Instruction learning (Super-Inst, Self-Inst, Flan …); Retrieval enhancement; Introduction to applications of large language models

Exercises: Using large language models


08 Advances

Efficient language models, linear attention mechanism; Integration of knowledge in natural language processing; Multimodal; Introduction to research in natural language processing

01 Introduction to Natural Language Processing & Machine Learning Basics

Applications of natural language processing, examples; Basics of machine learning and deep learning


02 Static Word Representation & Text Classification

Sparse representation: one-hot, ppmi, SVD.; Dense representation: word2vec; Generative classifier: generative classifier; Naive Bayes: Naive Bayes; Discriminative classifier: discriminative classifier; Sparse features BoW/N-gram; Dense features

Exercises: Word2vec, text classification


03 Language Model Language Modeling (I)

n-gram language model; RNN recurrent neural network and language model; RNN, RNN-LM - LSTM / GRU -

Exercises: RNN language model


04 Language Model Language Modeling & Attention Mechanism (II)

Attention mechanism Attention; RNN LM with Attention

Exercises: Attention mechanism


05 Self-attention mechanism Self-attention & Transformer

Transformer Encoder; Transformer Decoder; Sequence-to-sequence Seq2Seq, concepts of encoder-decoder

Exercises: huggingface tutorial


06 Pretrained Language Model Pretrained LM

Encoder LM (BERT); Decoder-only LM (GPT); Seq2seq LM (T5) -

Exercises: huggingface tutorial


07 Large Language Model (Large Language Model) -

Prompting and prompt learning; Contextual learning In-context Learning; Evolutionary trajectory of GPT series; Basic introduction (GPT-3, GPT3.5, ChatGPT, GPT4...); Instruction learning (Super-Inst, Self-Inst, Flan …); Retrieval enhancement; Introduction to applications of large language models

Exercises: Using large language models


08 Advances

Efficient language models, linear attention mechanism; Integration of knowledge in natural language processing; Multimodal; Introduction to research in natural language processing

PREREQUISITES

Prerequisites 04

PREREQUISITES

Prerequisites 04

Linear algebra, calculus, basic statistics, familiar with Python syntax, foundation in machine learning.


*During the pre-study phase, mentors and academic leaders will help you complete the prerequisite learning

Linear algebra, calculus, basic statistics, familiar with Python syntax, foundation in machine learning.


*During the pre-study phase, mentors and academic leaders will help you complete the prerequisite learning

"What XA has brought to me, the improvement of CS literacy goes without saying, but what I am more grateful for is being able to meet the top and most outstanding peers of my age. I have seen living MIT, CIT students, consulted TAs who can do nothing but debug, and also gained a bunch of board game friends. Perhaps saying that my dream is to change the world will be laughed at by people, but XA has given me the confidence to turn my dreams into reality."

This road is long, and the places to go are filled with poetry and ideals. The tears and sweat left behind yesterday make me more tenacious and courageous now. This is what XA has taught me: we can always be a little braver."

Haiyi Jin Explorer
Fudan University

"What XA has brought to me, the improvement of CS literacy goes without saying, but what I am more grateful for is being able to meet the top and most outstanding peers of my age. I have seen living MIT, CIT students, consulted TAs who can do nothing but debug, and also gained a bunch of board game friends. Perhaps saying that my dream is to change the world will be laughed at by people, but XA has given me the confidence to turn my dreams into reality."

This road is long, and the places to go are filled with poetry and ideals. The tears and sweat left behind yesterday make me more tenacious and courageous now. This is what XA has taught me: we can always be a little braver."

Haiyi Jin Explorer
Fudan University

"What XA has brought to me, the improvement of CS literacy goes without saying, but what I am more grateful for is being able to meet the top and most outstanding peers of my age. I have seen living MIT, CIT students, consulted TAs who can do nothing but debug, and also gained a bunch of board game friends. Perhaps saying that my dream is to change the world will be laughed at by people, but XA has given me the confidence to turn my dreams into reality."

This road is long, and the places to go are filled with poetry and ideals. The tears and sweat left behind yesterday make me more tenacious and courageous now. This is what XA has taught me: we can always be a little braver."

Haiyi Jin Explorer
Fudan University

MENTORS

Previous Course Mentors 05

MENTORS

Previous Course Mentors 05

Jiang Chengyue

Ph.D. in Computer Science, ShanghaiTech University


Research interests in natural language processing, focusing on the integration of knowledge representation and symbolized knowledge with neural networks. Published multiple research papers in top academic conferences such as EMNLP 2020 (High Score), Findings of EMNLP 2020, and CoNLL 2019. Served as a teaching assistant for the Artificial Intelligence course for three consecutive years, with extensive experience in hackathons and innovation competitions, including winning the first prize in the Challenge Cup 2019 and Saike 2017 Hackathon. Involved in inventing multiple pending technical patents and about to embark on a research internship at Alibaba DAMO Academy.

Jiang Chengyue

Ph.D. in Computer Science, ShanghaiTech University


Research interests in natural language processing, focusing on the integration of knowledge representation and symbolized knowledge with neural networks. Published multiple research papers in top academic conferences such as EMNLP 2020 (High Score), Findings of EMNLP 2020, and CoNLL 2019. Served as a teaching assistant for the Artificial Intelligence course for three consecutive years, with extensive experience in hackathons and innovation competitions, including winning the first prize in the Challenge Cup 2019 and Saike 2017 Hackathon. Involved in inventing multiple pending technical patents and about to embark on a research internship at Alibaba DAMO Academy.

ACADEMIC LEADS

Previous Course TAs 06

ACADEMIC LEADS

Previous Course TAs 06

Shenyang Art Yangyi

Stanford University Symbol System Major


Stanford University Symbol System Major, graduated from UWC Changshu, with strong enthusiasm for linguistics and natural language processing. Once ranked first in the International Linguistics Olympiad China region, selected for the Chinese team and won first place in the Asia-Pacific International Linguistics Olympiad team, selected for the Chinese Linguistics Olympiad academic group. The image captioning team he formed was once selected for the Iowa SSTP lab, YSI, Young Aurora, and built an automatic traffic light recognizer for the visually impaired, and built a model for the automatic annotation of Chinese painting in museums.



Jiang Nan Will

University of California, San Diego, Computer Science/Bioinformatics Major


Previously engaged in immunology-related research at Greiff Lab and ImmunoMind and exploring the application of machine learning in single-cell analysis. Participated in the development of the WeChat mini-program Hungry24 for the China Poverty Alleviation Foundation and ultimately obtained 17,800 registered users. Previously conducted research on few-shot learning in natural language processing at Baidu Research Institute, and participated in the development of the PaddlePaddle deep learning framework's few-shot learning library. Currently interested in network security, recommendation systems, and NLP.


Shenyang Art Yangyi

Stanford University Symbol System Major


Stanford University Symbol System Major, graduated from UWC Changshu, with strong enthusiasm for linguistics and natural language processing. Once ranked first in the International Linguistics Olympiad China region, selected for the Chinese team and won first place in the Asia-Pacific International Linguistics Olympiad team, selected for the Chinese Linguistics Olympiad academic group. The image captioning team he formed was once selected for the Iowa SSTP lab, YSI, Young Aurora, and built an automatic traffic light recognizer for the visually impaired, and built a model for the automatic annotation of Chinese painting in museums.



Jiang Nan Will

University of California, San Diego, Computer Science/Bioinformatics Major


Previously engaged in immunology-related research at Greiff Lab and ImmunoMind and exploring the application of machine learning in single-cell analysis. Participated in the development of the WeChat mini-program Hungry24 for the China Poverty Alleviation Foundation and ultimately obtained 17,800 registered users. Previously conducted research on few-shot learning in natural language processing at Baidu Research Institute, and participated in the development of the PaddlePaddle deep learning framework's few-shot learning library. Currently interested in network security, recommendation systems, and NLP.


Language of instruction

中文

Current course fees

20,410

Location

Shanghai

X ACADEMY
Global Explorer Summit
-


2024.07.25

Location

Shanghai

Language of instruction

中文

Current course fees

20,410

X ACADEMY
Global Explorer Summit
-


2024.07.25

Frequently Asked Questions

How is the student-teacher ratio in the major courses?

In each year's X ACADEMY summer program, the ratio of mentors to explorers is 1:5, with 1-2 mentors and 2-5 course assistants (academic leaders) in each course. Compared to traditional large classes, you will have more abundant opportunities for communication and learning.

I heard that the major courses of X ACADEMY are very hardcore. Will I fall behind when the course starts?

Can I only study one major course at X ACADEMY?

© 2024 | Shanghai Des Education Technology Co., Ltd. | Established in 2016. | 沪ICP备2021004001号-6

OUR CIVILIZATION WAS BUILT ON TECHNOLOGY. OUR CIVILIZATION IS BUILT ON TECHNOLOGY. TECHNOLOGY IS THE GLORY OF HUMAN AMBITION AND ACHIEVEMENT, THE SPEARHEAD OF PROGRESS, AND THE REALIZATION OF OUR POTENTIAL. FOR HUNDREDS OF YEARS, WE PROPERLY GLORIFIED THIS – UNTIL RECENTLY. WE, AT X ACADEMY, ARE HERE TO REVITALIZE THAT SPIRIT. TO US, THE FUTURE IS A CANVAS OF VAST POSSIBILITY, PAINTED WITH THE BRUSHES OF INNOVATION, CREATIVITY, AND OPTIMISM. WE BELIEVE IN A FUTURE WHERE TECHNOLOGY CONTINUES TO BE THE BACKBONE OF ADVANCEMENT, NOT JUST IN MATERIAL TERMS BUT AS A MEANS TO ELEVATE HUMAN EXPERIENCE AND CONSCIOUSNESS. WE ARE THE FUTURE OPTIMISTS, THE DREAMERS, THE MAKERS, AND THE DOERS. WE STAND READY TO COLLABORATE, TO INNOVATE, AND TO BUILD A BETTER TOMORROW WITH YOU. JOIN US, AS WE EMBARK ON THIS JOURNEY TO IGNITE THE FLAMES OF PROGRESS AND TO CHERISH THE TECHNOLOGY THAT MAKES IT POSSIBLE. IT IS TIME, ONCE AGAIN, TO RAISE THE TECHNOLOGY FLAG. IT IS TIME TO BE TECHNO-OPTIMISTS. WITH X ACADEMY, LET'S BUILD A BRIGHTER FUTURE TOGETHER.

© 2024 | Shanghai Des Education Technology Co., Ltd. | Established in 2016. |

沪ICP备2021004001号-6

OUR CIVILIZATION WAS BUILT ON TECHNOLOGY. OUR CIVILIZATION IS BUILT ON TECHNOLOGY. TECHNOLOGY IS THE GLORY OF HUMAN AMBITION AND ACHIEVEMENT, THE SPEARHEAD OF PROGRESS, AND THE REALIZATION OF OUR POTENTIAL. FOR HUNDREDS OF YEARS, WE PROPERLY GLORIFIED THIS – UNTIL RECENTLY. WE, AT X ACADEMY, ARE HERE TO REVITALIZE THAT SPIRIT. TO US, THE FUTURE IS A CANVAS OF VAST POSSIBILITY, PAINTED WITH THE BRUSHES OF INNOVATION, CREATIVITY, AND OPTIMISM. WE BELIEVE IN A FUTURE WHERE TECHNOLOGY CONTINUES TO BE THE BACKBONE OF ADVANCEMENT, NOT JUST IN MATERIAL TERMS BUT AS A MEANS TO ELEVATE HUMAN EXPERIENCE AND CONSCIOUSNESS. WE ARE THE FUTURE OPTIMISTS, THE DREAMERS, THE MAKERS, AND THE DOERS. WE STAND READY TO COLLABORATE, TO INNOVATE, AND TO BUILD A BETTER TOMORROW WITH YOU. JOIN US, AS WE EMBARK ON THIS JOURNEY TO IGNITE THE FLAMES OF PROGRESS AND TO CHERISH THE TECHNOLOGY THAT MAKES IT POSSIBLE. IT IS TIME, ONCE AGAIN, TO RAISE THE TECHNOLOGY FLAG. IT IS TIME TO BE TECHNO-OPTIMISTS. WITH X ACADEMY, LET'S BUILD A BRIGHTER FUTURE TOGETHER.

© 2024 | Shanghai Des Education Technology Co., Ltd. | Established in 2016. |

沪ICP备2021004001号-6

OUR CIVILIZATION WAS BUILT ON TECHNOLOGY. OUR CIVILIZATION IS BUILT ON TECHNOLOGY. TECHNOLOGY IS THE GLORY OF HUMAN AMBITION AND ACHIEVEMENT, THE SPEARHEAD OF PROGRESS, AND THE REALIZATION OF OUR POTENTIAL. FOR HUNDREDS OF YEARS, WE PROPERLY GLORIFIED THIS – UNTIL RECENTLY. WE, AT X ACADEMY, ARE HERE TO REVITALIZE THAT SPIRIT. TO US, THE FUTURE IS A CANVAS OF VAST POSSIBILITY, PAINTED WITH THE BRUSHES OF INNOVATION, CREATIVITY, AND OPTIMISM. WE BELIEVE IN A FUTURE WHERE TECHNOLOGY CONTINUES TO BE THE BACKBONE OF ADVANCEMENT, NOT JUST IN MATERIAL TERMS BUT AS A MEANS TO ELEVATE HUMAN EXPERIENCE AND CONSCIOUSNESS. WE ARE THE FUTURE OPTIMISTS, THE DREAMERS, THE MAKERS, AND THE DOERS. WE STAND READY TO COLLABORATE, TO INNOVATE, AND TO BUILD A BETTER TOMORROW WITH YOU. JOIN US, AS WE EMBARK ON THIS JOURNEY TO IGNITE THE FLAMES OF PROGRESS AND TO CHERISH THE TECHNOLOGY THAT MAKES IT POSSIBLE. IT IS TIME, ONCE AGAIN, TO RAISE THE TECHNOLOGY FLAG. IT IS TIME TO BE TECHNO-OPTIMISTS. WITH X ACADEMY, LET'S BUILD A BRIGHTER FUTURE TOGETHER.