Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Indian Institute of Technology Madras

Deep Learning Part 1 (IITM)

Indian Institute of Technology Madras and NPTEL via Swayam


Deep Learning has received a lot of attention over the past few years and has been employed successfully by companies like Google, Microsoft, IBM, Facebook, Twitter etc. to solve a wide range of problems in Computer Vision and Natural Language Processing. In this course we will learn about the building blocks used in these Deep Learning based solutions. Specifically, we will learn about feedforward neural networks, convolutional neural networks, recurrent neural networks and attention mechanisms. We will also look at various optimization algorithms such as Gradient Descent, Nesterov Accelerated Gradient Descent, Adam, AdaGrad and RMSProp which are used for training such deep neural networks. At the end of this course students would have knowledge of deep architectures used for solving various Vision and NLP tasks
INTENDED AUDIENCE: Any Interested LearnersPREREQUISITES: Working knowledge of Linear Algebra, Probability Theory. It would be beneficial if the participants have
done a course on Machine Learning.



Week 1: (Partial) History of Deep Learning, Deep Learning Success Stories, McCulloch Pitts Neuron, Thresholding Logic,
Perceptrons, Perceptron Learning Algorithm
Week 2: Multilayer Perceptrons (MLPs), Representation Power of MLPs, Sigmoid Neurons, Gradient Descent, Feedforward
Neural Networks, Representation Power of Feedforward Neural Networks
Week 3: FeedForward Neural Networks, Backpropagation
Week 4: Gradient Descent (GD), Momentum Based GD, Nesterov Accelerated GD, Stochastic GD, AdaGrad, RMSProp, Adam,
Eigenvalues and eigenvectors, Eigenvalue Decomposition, Basis
Week 5: Principal Component Analysis and its interpretations, Singular Value Decomposition
Week 6: Autoencoders and relation to PCA, Regularization in autoencoders, Denoising autoencoders, Sparse autoencoders,
Contractive autoencoders
Week 7: Regularization: Bias Variance Tradeoff, L2 regularization, Early stopping, Dataset augmentation, Parameter sharing
and tying, Injecting noise at input, Ensemble methods, Dropout
Week 8: Greedy Layerwise Pre-training, Better activation functions, Better weight initialization methods, Batch Normalization
Week 9: Learning Vectorial Representations Of Words
Week 10: Convolutional Neural Networks, LeNet, AlexNet, ZF-Net, VGGNet, GoogLeNet, ResNet, Visualizing Convolutional
Neural Networks, Guided Backpropagation, Deep Dream, Deep Art, Fooling Convolutional Neural Networks
Week 11: Recurrent Neural Networks, Backpropagation through time (BPTT), Vanishing and Exploding Gradients, Truncated BPTT, GRU, LSTMs
Week 12: Encoder Decoder Models, Attention Mechanism, Attention over images

Taught by

Prof. Sudarshan Iyengar & Prof.Mitesh M. Khapra

Related Courses


5.0 rating, based on 1 reviews

Start your review of Deep Learning Part 1 (IITM)

  • Anonymous

    Anonymous completed this course.

    The entire learning experience was great. The course is a package of models that helps us to build different networks. It gives an expanded view of FFNN, CNN and RNN. The course makes you get a thorough understanding of how each and every type of network works . It also lets you know the different tricks and tips that one should be aware of while designing a neural network to get top notch performance. It also provides at the end the various applications of what is being learnt from the course. It is an amazing course and it is worth taking it to explore the DEEP DEPTHS OF DEEP LEARNING.

    On top of everything, THE TEACHING IS GREAT.

    Happy learning!!!

Never Stop Learning!

Get personalized course recommendations, track subjects and courses with reminders, and more.

Sign up for free