Overview
Class Central Tips
This course will provide you a foundational understanding of machine learning models (logistic regression, multilayer perceptrons, convolutional neural networks, natural language processing, etc.) as well as demonstrate how these models can solve complex problems in a variety of industries, from medical diagnostics to image recognition to text prediction. In addition, we have designed practice exercises that will give you hands-on experience implementing these data science models on data sets. These practice exercises will teach you how to implement machine learning algorithms with PyTorch, open source libraries used by leading tech companies in the machine learning field (e.g., Google, NVIDIA, CocaCola, eBay, Snapchat, Uber and many more).
Syllabus
- Simple Introduction to Machine Learning
- The focus of this module is to introduce the concepts of machine learning with as little mathematics as possible. We will introduce basic concepts in machine learning, including logistic regression, a simple but widely employed machine learning (ML) method. Also covered is multilayered perceptron (MLP), a fundamental neural network. The concept of deep learning is discussed, and also related to simpler models.
- Basics of Model Learning
- In this module we will be discussing the mathematical basis of learning deep networks. We’ll first work through how we define the issue of learning deep networks as a minimization problem of a mathematical function. After defining our mathematical goal, we will introduce validation methods to estimate real-world performance of the learned deep networks. We will then discuss how gradient descent, a classical technique in optimization, can be used to achieve this mathematical goal. Finally, we will discuss both why and how stochastic gradient descent is used in practice to learn deep networks.
- Image Analysis with Convolutional Neural Networks
- This week will cover model training, as well as transfer learning and fine-tuning. In addition to learning the fundamentals of a CNN and how it is applied, careful discussion is provided on the intuition of the CNN, with the goal of providing a conceptual understanding.
- Recurrent Neural Networks for Natural Language Processing
- This week will cover the application of neural networks to natural language processing (NLP), from simple neural models to the more complex. The fundamental concept of word embeddings is discussed, as well as how such methods are employed within model learning and usage for several NLP applications. A wide range of neural NLP models are also discussed, including recurrent neural networks, and specifically long short-term memory (LSTM) models.
- The Transformer Network for Natural Language Processing
- This week we'll cover an Introduction to the Transformer Network, a deep machine learning model designed to be more flexible and robust than Recurrent Neural Network (RNN). We'll start by reviewing several machine learning building blocks of a Transformer Network: the Inner products of word vectors, attention mechanisms, and sequence-to-sequence encoders and decoders. Then, we'll put all of these components together to explore the complete Transformer Network.
- Introduction to Reinforcement Learning
- This week will cover Reinforcement Learning, a fundamental concept in machine learning that is concerned with taking suitable actions to maximize rewards in a particular situation. After learning the initial steps of Reinforcement Learning, we'll move to Q Learning, as well as Deep Q Learning. We'll discuss the difference between the concepts of Exploration and Exploitation and why they are important.
Taught by
Lawrence Carin
Tags
Reviews
3.5 rating, based on 2 Class Central reviews
4.7 rating at Coursera based on 3608 ratings
Showing Class Central Sort
-
The theory is ok and videos are well created. The practice is also fine, but not working properly. The page sends you to a Duke U site that is not working right now :-P). But even if it worked, since is a site outside of coursera, the activity neve…
-
Excellent theoretical introduction, useful for those who know something but still do not have a broader picture.
There are also Jupiter notebooks where you can find the implementation in Python, but unfortunately it is additional material and is not explained in the videos.