Loss functions for classification: Difference between revisions

Content deleted Content added
Savage loss: Made the parentheses look better.
Added plots of Cross-entropy and other loss functions used to train ANNs.
Tags: Reverted Visual edit
Line 98:
|<math>\arctan(v)+\frac{1}{2}</math>
|<math>\tan(\eta-\frac{1}{2})</math>
|}
|}<br />The sole minimizer of the expected risk, <math>f^*_{\phi}</math>, associated with the above generated loss functions can be directly found from equation (1) and shown to be equal to the corresponding <math>
[[File:Plot of Loss Functions.png|thumb|A plot of different loss functions that can be used to train ANNs for classification.]]
<br />
 
|}<br />The sole minimizer of the expected risk, <math>f^*_{\phi}</math>, associated with the above generated loss functions can be directly found from equation (1) and shown to be equal to the corresponding <math>
f(\eta)
</math>. This holds even for the nonconvex loss functions, which means that gradient descent based algorithms such as [[gradient boosting]] can be used to construct the minimizer.