Overview
Explore a thought-provoking lecture examining the impact of pre-training on natural language processing, domain transfer, and distribution shift through the lens of modern language models. Delve into the progression from ELMo and BERT models to contemporary large language models, questioning whether pre-training genuinely enhances task comprehension or merely improves pattern recognition. Investigate the complexities of transfer learning in models pre-trained on vast internet datasets, and examine research findings that present contrasting evidence about language models' true understanding capabilities. Learn from Robin Jia of the University of Southern California as he shares insights from various studies, some suggesting the limitations of pre-training alone for genuine task comprehension, while others indicate pre-training's success in developing appropriate representations.
Syllabus
The Evolution of Domain Transfer in the Era of Pre-trained Language Models
Taught by
Simons Institute