Overview
Syllabus
⌨️ Introduction
⌨️ Question 1: What is Deep Learning?
⌨️ Question 2: How does Deep Learning differ from traditional Machine Learning?
⌨️ Question 3: What is a Neural Network?
⌨️ Question 4: Explain the concept of a neuron in Deep Learning
⌨️ Question 5: Explain architecture of Neural Networks in simple way
⌨️ Question 6: What is an activation function in a Neural Network?
⌨️ Question 7: Name few popular activation functions and describe them
⌨️ Question 8: What happens if you do not use any activation functions in a neural network?
⌨️ Question 9: Describe how training of basic Neural Networks works
⌨️ Question 10: What is Gradient Descent?
⌨️ Question 11: What is the function of an optimizer in Deep Learning?
⌨️ Question 12: What is backpropagation, and why is it important in Deep Learning?
⌨️ Question 13: How is backpropagation different from gradient descent?
⌨️ Question 14: Describe what Vanishing Gradient Problem is and it’s impact on NN
⌨️ Question 15: Describe what Exploding Gradients Problem is and it’s impact on NN
⌨️ Question 16: There is a neuron in the hidden layer that always results in an error. What could be the reason?
⌨️ Question 17: What do you understand by a computational graph?
⌨️ Question 18: What is Loss Function and what are various Loss functions used in Deep Learning?
⌨️ Question 19: What is Cross Entropy loss function and how is it called in industry?
⌨️ Question 20: Why is Cross-entropy preferred as the cost function for multi-class classification problems?
⌨️ Question 21: What is SGD and why it’s used in training Neural Networks?
⌨️ Question 22: Why does stochastic gradient descent oscillate towards local minima?
⌨️ Question 23: How is GD different from SGD?
⌨️ Question 24: How can optimization methods like gradient descent be improved? What is the role of the momentum term?
⌨️ Question 25: Compare batch gradient descent, minibatch gradient descent, and stochastic gradient descent.
⌨️ Question 26: How to decide batch size in deep learning considering both too small and too large sizes?
⌨️ Question 27: Batch Size vs Model Performance: How does the batch size impact the performance of a deep learning model?
⌨️ Question 28: What is Hessian, and how can it be used for faster training? What are its disadvantages?
⌨️ Question 29: What is RMSProp and how does it work?
⌨️ Question 30: Discuss the concept of an adaptive learning rate. Describe adaptive learning methods
⌨️ Question 31: What is Adam and why is it used most of the time in NNs?
⌨️ Question 32: What is AdamW and why it’s preferred over Adam?
⌨️ Question 33: What is Batch Normalization and why it’s used in NN?
⌨️ Question 34: What is Layer Normalization, and why it’s used in NN?
⌨️ Question 35: What are Residual Connections and their function in NN?
⌨️ Question 36: What is Gradient clipping and their impact on NN?
⌨️ Question 37: What is Xavier Initialization and why it’s used in NN?
⌨️ Question 38: What are different ways to solve Vanishing gradients?
⌨️ Question 39: What are ways to solve Exploding Gradients?
⌨️ Question 40: What happens if the Neural Network is suffering from Overfitting relate to large weights?
⌨️ Question 41: What is Dropout and how does it work?
⌨️ Question 42: How does Dropout prevent overfitting in NN?
⌨️ Question 43: Is Dropout like Random Forest?
⌨️ Question 44: What is the impact of Drop Out on the training vs testing?
⌨️ Question 45: What are L2/L1 Regularizations and how do they prevent overfitting in NN?
⌨️ Question 46: What is the difference between L1 and L2 regularisations in NN?
⌨️ Question 47: How do L1 vs L2 Regularization impact the Weights in a NN?
⌨️ Question 48: What is the curse of dimensionality in ML or AI?
⌨️ Question 49: How deep learning models tackle the curse of dimensionality?
⌨️ Question 50: What are Generative Models, give examples?
Taught by
freeCodeCamp.org