<aside> 💡

Backpropagation optimizes Weights and Biases in Neural Networks

</aside>

  1. Use Chain Rule to calculate derivative
  2. Putting derivatives into Gradient Descent to optimize parameters

image.png

We start with the last parameter (Bias) and work our way backwards to estimate all of the other parameters.