Loss functions for classification: Difference between revisions

Content deleted Content added
See also: Removed links already present in the article, per WP:SEEALSO.
Line 96:
 
The cross entropy loss is closely related to the [[Kullback-Leibler divergence]] between the empirical distribution and the predicted distribution. This function is not naturally represented as a product of the true label and the predicted value, but is convex and can be minimized using [[stochastic gradient descent]] methods. The cross entropy loss is ubiquitous in modern [[deep learning|deep neural networks]].
 
==See also==
*[[Statistical learning theory]]
*[[Loss function]]
*[[Support vector machine]]
 
== References ==