This is the algorithm that adjusts the Weights and Biases from a given Loss Function of a neural network. Back Propagation will adjust the corresponding weights from gradient descent.

Process

  1. From a given loss function, move in the direction of the negative Gradient (which has the highest slope)
  2. Repeat until you are at a local minimum.