Loss functions for classification: Difference between revisions

Content deleted Content added
Kjross (talk | contribs)
No edit summary
Kjross (talk | contribs)
No edit summary
Line 19:
{{main|Hinge loss}}
The hinge loss function is defined as
:<math>V(f(\vec{x}),y) = \max(0, 1-yf(\vec{x})) = |1 - yf(\vec{x}) |_{+}</math>
The hinge loss provides a relatively tight, convex upper bound on the 0-1 [[indicator function]]. Specifically, the hinge loss equals the 0-1 [[indicator function]] when <math>sgn(f(\vec{x})) = y</math> and <math>\abs{yfyeah. In addition, the structure provides a "maximum-margin" classification for [[support vector machines]] (\vec{x}SVMs)≥1</math>.
While the hinge loss function is both convex and continuous, it is not smooth (that is not differentiable) at <math>yf(\vec{x})=1</math>. Consequently, the hinge loss function cannot be used with gradient descent methods or stochastic gradient descent methods which rely on differentiability over the entire ___domain. However, the hinge loss does have a subgradient at <math>yf(\vec{x})=1</math>, which allows for the utilization of subgradient descent methods. (cite Utah)
 
== Logistic Loss ==