Neural Network Advanced - utkaln/machine-learning GitHub Wiki

Shorcut to Neural-Network-Basics

Gradient Descent replaced with Adam Optimizer

  • Gradient Descent can be further optimized to automatically choose better Learning Rate aka alpha
  • This is done by choosing Adam Optimizer instead of traditional Gradient Descent equation with a manual input for learning rate
  • This is represented in Tensorflow with a minor modification in the compile step that computes the loss (cost) function
model.compile(
  optimizer=tf.keras.optimizers.Adam(learning_rate=1e-3), 
  loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True))

Convolutional Neural Network

  • Instead of each layer of Neural Network trying to read into the all the training data or output of the previous layer, another alternative is to only look into a subset of data to make more efficient prediction and reduce the risk of overfitting. This type of layers are called Convolutional Layers.
  • Convolutional is a strong alternative to Dense layer