Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Making LLMs Fully Utilize Context - A Data-Driven Approach

AI Bites via YouTube

Overview

Explore a 14-minute video explanation of Microsoft's research paper on improving Large Language Models' context utilization through data-driven solutions, contrasting with Google's architectural approach in the infini-attention paper. Learn about the 'Lost in the Middle Challenge,' Information Intensive Training (IN2), and Various Long Context Probing (VAL) methodologies. Dive into mathematical representations, training settings, experimental results, and real-world performance data that demonstrate how LLMs can better process and utilize extended context. Presented by an experienced Machine Learning researcher with 15 years of software engineering background, the video breaks down complex concepts into digestible segments, complete with detailed timestamps for easy navigation through specific topics.

Syllabus

- Intro
- Lost in the Middle Challenge in Context
- Related work in Long Context LLMs
- Information Intensive Training IN2 Training
- Fine-grained Information awareness
- Integration and Reasoning of Information
- Mathematical Representation
- Trainnig setting/Details
- VArious Long Context Probing VAL Probing
- Needle in a Haystack for Long Context LLMs
- Experimental Results
- Quantitative Results
- Real-world data performance
- Summary and Extro

Taught by

AI Bites

Reviews

Start your review of Making LLMs Fully Utilize Context - A Data-Driven Approach

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.