You are viewing a single comment's thread from:

RE: Backpropagation Step by Step - [The Mathematical Approach]

in #deep-learning7 years ago

I always use chain rule of differentiation to understand backpropagation. Since neural network can be represented as f(g(h(...(input)...))) where each function is a layer of the neural network, while trying to minimize the error function which is another function of (predicted-actual), we differentiate this function of function representation, resulting in usage of chain rule.

Coin Marketplace

STEEM 0.21
TRX 0.20
JST 0.034
BTC 90598.76
ETH 3112.87
USDT 1.00
SBD 2.99