LLMs and Transformers Demystified: Introduction to AI Engineering - Lecture 1
Data Centric via YouTube
Overview
Dive into the world of artificial intelligence with this comprehensive lecture on Large Language Models (LLMs) and Transformers. Explore the groundbreaking 'Attention is All You Need' research paper and its impact on AI development. Learn about the transformer architecture, a key component in many LLMs powering applications like Chat-GPT. Gain essential context for engaging with LLMs through an intuitive weather analogy, making complex concepts accessible. Discover the transformative power of the transformer architecture and its profound influence on AI advancements. Follow along as the lecture covers topics such as the history of transformers, the structure of "Transformer City," detailed explanations of encoder and decoder towers, and an analogy for model training. Perfect for AI engineers, researchers, and enthusiasts looking to demystify the complexities of transformers and enhance their understanding of cutting-edge AI technologies.
Syllabus
Introduction – Intro to Transformers & LLMs
Attention is All you Need – History of Transformers
Welcome to Transformer City
Geographic Map of Transformer City
Encoder Tower in Detail
Encoder & Decoder Towers Working Together
Decoder Tower in Detail
Training our Scientists to forecast model training analogy
Outro
Taught by
Data Centric