Backpropagation derivation using Leibniz notation

From Machinelearning
Revision as of 22:28, 8 November 2018 by IssaRice (talk | contribs)

The cost function C depends on wjkl only through the activation of the jth neuron in the lth layer, i.e. on the value of ajl. Thus we can use the chain rule to expand:

Cwjkl=Cajlajlwjkl

We know that ajlwjkl=σ(zjl)akl1 because ajl=σ(zjl)=σ(kwjklakl1+bjl). We have used the chain rule again here.

In turn, C depends on ajl only through the activations of the (l+1)th layer. Thus we can write:

Cajl=i{1,,n(l+1)}Cail+1ail+1ajl

where n(l+1) is the number of neurons in the (l+1)th layer.