Deep Learning Express

A quick head start for deep learning starting from MLP to Transformers, BERT, and GPT-2!

Through these 6 Google Colab notebooks you will learn necessary deep learning concepts and you will learn how to code deep learning models using PyTorch


Notebook 1 (Key Concepts): In this notebook you will learn the required deep learning (and ML) concepts before diving into the model architectures. These topics include: dataset spliting, forward propagation and backward propagation, optimizers, loss functions, activation functions, weight initialization, regularization, and normalization.

Open In Colab


Notebook 2 (Multi-Layer Perceptron (MLP)): In this notebook you will learn about the simplest deep learning model, the Multi-Layer Perceptron (MLP) network.

If you want to do data visualization, data compression, encryption, or simple classification, MLP might be a good choice!

Open In Colab


Notebook 3 (Convolutional Neural Networks (CNN)): In this notebook you will learn about Convolutional Neural Networks (CNN).

If you want to do work with images, such as image recognition, object detection, face recognition, etc tasks CNN might be a good choice!

Open In Colab


Notebook 4 (Recurrent Neural Networks (RNN)): In this nottebook you will learn about Recurrent Neural Networks (RNN).

If you are interested in sequence modeling tasks such as prediction problems, language modelling RNNs can be a good choice! Machine Translation.

Open In Colab


Notebook 5 (Transformers): In this notebook you will learn about transformer networks. Transformers are the state of the art architecture used for NLP applications and sequnce modeling.

Transformers are similar to RNNs but can solve sequence-to-sequence tasks while handling long-range dependencies, thus more powerful than RNNs!

Open In Colab


Notebook 5 extras (BERT & GPT-2): In these two notebooks you will learn how to fine-tune the two leading language models out there, BERT, and GPT-2. They are the same in that they are both based on the transformer architecture, but they are fundamentally different in that BERT has just the encoder blocks from the transformer, whilst GPT-2 has just the decoder blocks from the transformer.

BERT is mostly used for text classification tasks, while GPT-2 is great for text generation! Keep in mind that these models are very large in size!

BERT   Open In Colab

GPT-2   Open In Colab


Notebook 6 (Additional Concepts): In this notebook you will learn about some methods to enhance your model architectures, and some considerations that will help boost the training (depending on the application). These topics include: Residual Networks (ResNet), dimensionality reduction, gradient clipping, learning rate warmup.

Open In Colab