Loss functions for classification: Difference between revisions

Content deleted Content added
See also: Removed links already present in the article, per WP:SEEALSO.
add exponential loss
Line 96:
 
The cross entropy loss is closely related to the [[Kullback-Leibler divergence]] between the empirical distribution and the predicted distribution. This function is not naturally represented as a product of the true label and the predicted value, but is convex and can be minimized using [[stochastic gradient descent]] methods. The cross entropy loss is ubiquitous in modern [[deep learning|deep neural networks]].
 
== Exponential loss ==
The exponential loss function is defined as
 
:<math>V(f(\vec{x}),y) = e^{-yf(\vec{x})}</math>
 
It penalizes incorrect predictions more than Hinge loss and has larger gradient.
 
== References ==