Add backpropogation and chain rule relation. (#18667)

This commit is contained in:
Ajay Arasanipalai
2018-10-14 02:32:53 +05:30
committed by Quincy Larson
parent bf5cc9279f
commit 41f2f10e94

View File

@ -10,6 +10,8 @@ Backprogapation is a subtopic of [neural networks](../neural-networks/index.md).
**Method:** This is done by calculating the gradients of each node in the network. These gradients measure the "error" each node contributes to the output layer, so in training a neural network, these gradients are minimized.
Backpropogation can be thought of as using the chain rule to compute gradients with respect to different parameters in a neural network in order to perform iterative updates to those parameters.
Note: Backpropagation, and machine learning in general, requires significant familiarity with linear algebra and matrix manipulation. Coursework or reading on this topic is highly recommended before trying to understand the contents of this article.
### Computation