Backpropagation: Difference between revisions

Content deleted Content added
m Derivation: cite repair;
Yoderj (talk | contribs)
Finding the derivative of the error: Comment on differentiable activation functions
Line 162:
: <math>\frac{\partial o_j}{\partial\text{net}_j} = \frac {\partial}{\partial \text{net}_j} \varphi(\text{net}_j) = \varphi(\text{net}_j)(1-\varphi(\text{net}_j))</math>
 
This is the reason why backpropagation requires the activation function to be [[Differentiable function|differentiable]]. (Nevertheless, the non-differentiable [[ReLU]] activation function has become quite popular recently, e.g. in [[AlexNet]])
 
The first factor is straightforward to evaluate if the neuron is in the output layer, because then <math>o_j = y</math> and