DTSA 5511 Introduction to Deep Learning
Same as CSCA 5642
- Specialization: Machine Learning: Theory and Hands-on Practice with Python Specialization
- Instructor: Daniel Acuna
- Prior knowledge needed: Calculus, Linear algebra, Python
Learning Outcomes
- Explain what multilayer perceptrons, convolutional neural networks and recurrent neural networks are and how they work.
- Build and train neural network models using deep learning libraries such as Keras.
- Tune hyperparameters of a neural network model and an optimizer to improve performance.
- Choose and/or design neural network architectures.
- Explain how optimization and weight update works in neural network training
Course Content
Duration: 5h
Welcome to Introduction to Deep Learning. This module builds the mathematical foundations of neural networks. Starting from linear models, you will learn about the artificial neuron and develop the mathematics of gradient descent and backpropagation. The focus is on understanding how and why neural networks work through the underlying math—covering the forward pass, loss functions, and the chain rule to show how information flows through networks and how they learn from data.
Duration: 4h
This module focuses on training neural networks effectively. Topics include optimization algorithms, hyperparameter tuning, and regularization techniques to prevent overfitting and achieve good generalization. You will compare different optimizers like SGD, momentum, and Adam, understand how learning rate and batch size affect training dynamics, and apply weight decay, dropout, early stopping, and batch normalization.
Duration: 3h
This module introduces you to convolutional neural networks (CNNs), the foundation of modern computer vision. Topics include how convolutional and pooling layers work, CNN architecture design, and practical techniques like data augmentation and transfer learning. The module covers classic architectures like VGG and ResNet and explains why CNNs outperform fully-connected networks on image data.
Duration: 3h
This module covers sequence modeling, starting with recurrent neural networks (RNNs) and long short-term memory networks (LSTMs), then progressing to the attention mechanism—the key innovation that led to transformers. Topics include how RNNs maintain hidden states across time steps, why the vanishing gradient problem motivated LSTMs, and how attention allows models to focus on relevant parts of their input.
Duration: 3h
This final module covers the transformer architecture, which has revolutionized deep learning across domains. Topics include BERT and GPT as encoder-only and decoder-only variants, Vision Transformers (ViT) that apply attention to images, and CLIP for multimodal learning connecting vision and language. The module emphasizes applying pre-trained models to real tasks.
Duration: 1h
You will complete a final exam 20% of your grade. You must attempt the final in order to earn a grade in the course. If you've upgraded to the for-credit version of this course, please make sure you review the additional for-credit materials in the Introductory module and anywhere else they may be found.
Note: This page is periodically updated. Course information on the Coursera platform supersedes the information on this page. Click View on Coursera button above for the most up-to-date information.