The Machine Learning Behind Apple Intelligence - Understanding Modern LLM Architecture
Neural Breakdown with AVB via YouTube
Overview
Explore a detailed 20-minute technical video breakdown of Apple's Foundation Large Language Models (AFM-server and AFM-on-device) as revealed in their latest technical report. Learn about advanced machine learning concepts including Next Word Prediction with Transformer Decoders, Reinforcement Learning with Human Feedback, Low Rank Adaptation (LoRA), Knowledge Distillation, and Structured Pruning. Dive into technical implementations like Quantization with Palettization and Mirror Descent Policy Optimization with Leave One Out (MDLOO). Through structured chapters covering pretraining, adapters, and RLHF, gain comprehensive insights into the algorithmic advancements powering Apple's language models. Access supplementary materials including write-ups, slides, and notebooks through membership options, while benefiting from relevant research paper references and links to related content about NLP history and transformer architectures.
Syllabus
- Intro
- Chapter 1 - Overview
- Pretraining
- Structured Pruning
- Knowledge Distillation
- Post Training
- Iterative Teaching Committee
- Chapter 2 - Adapters
- LoRA Low Rank Adapters
- Quantization Palettization
- Chapter 3 - RLHF
- Reward Modelling
- Leave One Out
- Mirror Descent Policy and MDLOO
- Results
Taught by
Neural Breakdown with AVB