- Computer Science
- Artificial Intelligence
- Natural Language Processing (NLP)
- LLM (Large Language Model)
- ChatGPT
- Computer Science
- Artificial Intelligence
- Natural Language Processing (NLP)
- LLM (Large Language Model)
- GPT-2
- Computer Science
- Artificial Intelligence
- Natural Language Processing (NLP)
- LLM (Large Language Model)
- GPT-4
Inside GPT - Large Language Models Demystified
Overview
Dive deep into the architecture and inner workings of GPT algorithms and ChatGPT in this comprehensive conference talk from GOTO Amsterdam 2024. Explore fundamental concepts of natural language processing, including word embedding, vectorization, and tokenization. Follow along with hands-on demonstrations of training a GPT2 model to generate song lyrics, showcasing the internals of word sequence prediction. Examine larger language models like ChatGPT and GPT4, understanding their capabilities and limitations. Learn about hyperparameters such as temperature and frequency penalty, and see their effects on generated output. Gain practical insights into harnessing GPT algorithms for your own solutions through multiple demos covering ChatGPT2, Word2Vec dimensionality reduction, GPT2 input embedding, multi-head attention, and next token prediction. Discover how to leverage these powerful tools to create engaging and useful applications for your business.
Syllabus
Intro
GPT sequence prediction
Prompt engineering
Demo: ChatGPT2
Processing text
Demo: Word2Vec dimensionality reduction
Transformer architecture
Demo: GPT2 input embedding
Self attention
Demo: GPT2 multi-head attention
Attention example
Demo: GPT2 next token prediction
Parameters
Thanks for explanations & inspiration
Outro
Taught by
GOTO Conferences