Neural-Networks-From-Scratch/README.md

57 lines
1.5 KiB
Markdown

# Purpose
Following along with the playlist created by Vizuara on Youtube (https://www.youtube.com/playlist?list=PLPTV0NXA_ZSj6tNyn_UadmUeU3Q3oR-hu).
The primary objective is to gain a foundational understanding of simple neural networks including forward propagation, activation layers, backward propagation, and gradient descent.
# Lectures
## Lectures 1-2
Neurons and layers with matrices and numpy.
## Lectures 3-6
Multiple layer neural networks with matrices and numpy.
ReLU and softmax activation functions.
## Lectures 7-11
Cross entropy loss and optimizing a single layer with gradient descent.
## Lecture 12
Backpropagation for a single neuron.
## Lectures 13-17
Backpropagation for neuron layers and activation layers.
## Lecture 18
Backpropagation on the cross entropy loss function.
## Lectures 19-21
Combined backpropagation for softmax and cross entropy loss.
Entire backpropagation pipeline for neural networks.
Entire forward pass pipeline for neural networks.
## Lecture 22
Gradient descent for entire neural network.
## Lectures 23-24
Learing rate decay in optimization.
Momentum in training neural networks.
## Lectures 25-26
Coding the ADAGRAD optimizer for neural networks.
Coding the RMSprop optimizer for neural networks.
## Lecture 27
Coding the ADAM optimizer for neural networks.
## Lectures 28-31
Neural network testing, generilization, and overfitting.
K-Fold cross validation.
L1/L2 regularization to avoid overfitting.
Dropout layers in neural networks.