Content deleted Content added
→Savage loss: Made the parentheses look better. |
Added plots of Cross-entropy and other loss functions used to train ANNs. Tags: Reverted Visual edit |
||
Line 98:
|<math>\arctan(v)+\frac{1}{2}</math>
|<math>\tan(\eta-\frac{1}{2})</math>
|}
|}<br />The sole minimizer of the expected risk, <math>f^*_{\phi}</math>, associated with the above generated loss functions can be directly found from equation (1) and shown to be equal to the corresponding <math>▼
[[File:Plot of Loss Functions.png|thumb|A plot of different loss functions that can be used to train ANNs for classification.]]
<br />
▲
f(\eta)
</math>. This holds even for the nonconvex loss functions, which means that gradient descent based algorithms such as [[gradient boosting]] can be used to construct the minimizer.
|