Fundamentals in Neural Networks

Build up your intuition of the fundamental building blocks of Neural Networks.

Deep learning (also known as deep structured learning) is part of a broader family of machine learning methods based on artificial neural networks with representation learning. Learning can be supervised, semi-supervised or unsupervised. Deep-learning architectures such as deep neural networks, deep belief networks, deep reinforcement learning, recurrent neural networks and convolutional neural networks have been applied to fields including computer vision, speech recognition, natural language processing, machine translation, bioinformatics, drug design, medical image analysis, material inspection and board game programs, where they have produced results comparable to and in some cases surpassing human expert performance.

What you’ll learn

  • Understand the intuition behind Artificial Neural Networks.
  • Understand the intuition behind Convolutional Neural Networks.
  • Understand the intuition behind Recurrent Neural Networks.
  • Apply Artificial Neural Networks in practice.
  • Apply Convolutional Neural Networks in practice.
  • Apply Recurrent Neural Networks in practice.

Course Content

  • Welcome! –> 2 lectures • 3min.
  • Artificial Neural Networks –> 13 lectures • 2hr 52min.
  • Convolutional Neural Network –> 12 lectures • 2hr 3min.
  • Recurrent Neural Network –> 10 lectures • 1hr 44min.

Fundamentals in Neural Networks

Requirements

  • There is no prior coding or programming experience required. This course assumes you have your own laptop and the code will be done using Colab..

Deep learning (also known as deep structured learning) is part of a broader family of machine learning methods based on artificial neural networks with representation learning. Learning can be supervised, semi-supervised or unsupervised. Deep-learning architectures such as deep neural networks, deep belief networks, deep reinforcement learning, recurrent neural networks and convolutional neural networks have been applied to fields including computer vision, speech recognition, natural language processing, machine translation, bioinformatics, drug design, medical image analysis, material inspection and board game programs, where they have produced results comparable to and in some cases surpassing human expert performance.

This course covers the following three sections: (1) Neural Networks, (2) Convolutional Neural Networks, and (3) Recurrent Neural Networks. You will be receiving around 4 hours of materials on detailed discussion, mathematical description, and code walkthroughs of the three common families of neural networks. The descriptions of each section is summarized below.

 

Section 1 – Neural Network

1.1 Linear Regression

1.2 Logistic Regression

1.3 Purpose of Neural Network

1.4 Forward Propagation

1.5 Backward Propagation

1.6 Activation Function (Relu, Sigmoid, Softmax)

1.7 Cross-entropy Loss Function

1.8 Gradient Descent

 

Section 2 – Convolutional Neural Network

2.1 Image Data

2.2 Tensor and Matrix

2.3 Convolutional Operation

2.4 Padding

2.5 Stride

2.6 Convolution in 2D and 3D

2.7 VGG16

2.8 Residual Network

 

Section 3 – Recurrent Neural Network

3.1 Welcome

3.2 Why use RNN

3.3 Language Processing

3.4 Forward Propagation in RNN

3.5 Backpropagation through Time

3.6 Gated Recurrent Unit (GRU)

3.7 Long Short Term Memory (LSTM)

3.8 Bidirectional RNN (bi-RNN)

Get Tutorial