Completed
IBM Model 1
Class Central Classrooms beta
YouTube videos curated by Class Central.
Classroom Contents
Applied Natural Language Processing
Automatically move to the next video in the Classroom when playback concludes
- 1 Operations on a Corpus
- 2 Probability and NLP
- 3 Machine Translation
- 4 Statistical Properties of Words - Part 01
- 5 Statistical Properties of Words - Part 02
- 6 Statistical Properties of Words - Part 03
- 7 Vector Space Models for NLP
- 8 Document Similarity - Demo, Inverted index, Exercise
- 9 Contextual understanding of text
- 10 Collocations, Dense word Vectors
- 11 Query Processing
- 12 Topic Modeling
- 13 Introduction
- 14 Sequence Learning
- 15 Vector Representation of words
- 16 Co-occurence matrix, n-grams
- 17 SVD, Dimensionality reduction, Demo
- 18 Vector Space models
- 19 Preprocessing
- 20 Introduction to Probability in the context of NLP
- 21 Joint and conditional probabilities, independence with examples
- 22 The definition of probabilistic language model
- 23 Chain rule and Markov assumption
- 24 Out of vocabulary words and curse of dimensionality
- 25 Exercise
- 26 Examples for word prediction
- 27 Generative Models
- 28 Bigram and Trigram Language models -peeking indide the model building
- 29 Naive-Bayes, classification
- 30 Machine learning, perceptron, linearly separable
- 31 Linear Models for Claassification
- 32 Biological Neural Network
- 33 Perceptron
- 34 Perceptron Learning
- 35 Logical XOR
- 36 Activation Functions
- 37 Gradient Descent
- 38 Feedforward and Backpropagation Neural Network
- 39 Why Word2Vec?
- 40 What are CBOW and Skip-Gram Models?
- 41 One word learning architecture
- 42 Forward pass for Word2Vec
- 43 Matrix Operations Explained
- 44 CBOW and Skip Gram Models
- 45 Binay tree, Hierarchical softmax
- 46 Updating the weights using hierarchical softmax
- 47 Sequence Learning and its applications
- 48 ANN as a LM and its limitations
- 49 Discussion on the results obtained from word2vec
- 50 Recap and Introduction
- 51 Mapping the output layer to Softmax
- 52 Reduction of complexity - sub-sampling, negative sampling
- 53 Building Skip-gram model using Python
- 54 GRU
- 55 Truncated BPTT
- 56 LSTM
- 57 BPTT - Exploding and vanishing gradient
- 58 BPTT - Derivatives for W,V and U
- 59 BPTT - Forward Pass
- 60 RNN - Based Language Model
- 61 Unrolled RNN
- 62 Introuduction to Recurrent Neural Network
- 63 IBM Model 2
- 64 IBM Model 1
- 65 Alignments again!
- 66 Translation Model, Alignment Variables
- 67 Noisy Channel Model, Bayes Rule, Language Model
- 68 What is SMT?
- 69 Introduction and Historical Approaches to Machine Translation
- 70 BLEU Demo using NLTK and other metrics
- 71 BLEU - "A short Discussion of the seminal paper"
- 72 Introduction to evaluation of Machine Translation
- 73 Extraction of Phrases
- 74 Introduction to Phrase-based translation
- 75 Symmetrization of alignments
- 76 Learning/estimating the phrase probabilities using another Symmetrization example
- 77 mod10lec79-Recap and Connecting Bloom Taxonomy with Machine Learning
- 78 mod10lec80-Introduction to Attention based Translation
- 79 mod10lec81- Neural machine translation by jointly learning to align and translate
- 80 mod10lec82-Typical NMT architecture architecture and models for multi-language translation
- 81 mod10lec77-Encoder-Decoder model for Neural Machine Translation
- 82 mod10lec78-RNN Based Machine Translation
- 83 mod10lec83-Beam Search
- 84 mod10lec84-Variants of Gradient Descend
- 85 mod11lec85-Introduction to Conversation Modeling
- 86 mod11lec86-A few examples in Conversation Modeling
- 87 mod11lec87-Some ideas to Implement IR-based Conversation Modeling
- 88 mod11lec88-Discussion of some ideas in Question Answering
- 89 mod12lec89-Hyperspace Analogue to Language - HAL
- 90 mod12lec90-Correlated Occurence Analogue to Lexical Semantic - COALS
- 91 mod12lec91-Global Vectors - Glove
- 92 mod12lec92-Evaluation of Word vectors