Overview
Syllabus
intro
Bengio et al. 2003 MLP language model paper walkthrough
re-building our training dataset
implementing the embedding lookup table
implementing the hidden layer + internals of torch.Tensor: storage, views
implementing the output layer
implementing the negative log likelihood loss
summary of the full network
introducing F.cross_entropy and why
implementing the training loop, overfitting one batch
training on the full dataset, minibatches
finding a good initial learning rate
splitting up the dataset into train/val/test splits and why
experiment: larger hidden layer
visualizing the character embeddings
experiment: larger embedding size
summary of our final code, conclusion
sampling from the model
google collab new!! notebook advertisement
Taught by
Andrej Karpathy