Loss functions for classification

This is an old revision of this page, as edited by Kjross (talk | contribs) at 18:08, 4 December 2014. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

This sandbox is in the article namespace. Either move this page into your userspace, or remove the {{User sandbox}} template. Loss function surrogates for classification are computationally feasible loss functions representing the price we will pay for inaccuracy in our predictions in classification problems. [1] Specifically, if {-1,1} represents the mapping of a vector to a class label {-1,1}, we wish to find a function which best approximates the true mapping . (citation needed) Given that loss functions are always true functions of only one variable, it is standard practice to define loss functions in terms of the product of the true classifier and the predicted value .

Square Loss

Hinge Loss

Logistic Loss

References

  1. ^ Attention: This template ({{cite doi}}) is deprecated. To cite the publication identified by doi:10.1162/089976604773135104, please use {{cite journal}} (if it was published in a bona fide academic journal, otherwise {{cite report}} with |doi=10.1162/089976604773135104 instead.