Radial basis function kernel

This is an old revision of this page, as edited by 24.130.2.80 (talk) at 22:03, 13 May 2013 (Closed paren around "(when '''x''' = '''x'''')"). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In machine learning, the (Gaussian) radial basis function kernel, or RBF kernel, is a popular kernel function. It is the most popular kernel function used in support vector machine classification[1]

The RBF kernel on two samples x and x', represented as feature vectors in some input space, is defined as[2]

may be recognized as the squared Euclidean distance between the two feature vectors. is a free parameter. An equivalent, but simpler, definition involves a parameter :

Since the value of the RBF kernel decreases with distance and ranges between zero (in the limit) and one (when x = x'), it has a ready interpretation as a similarity measure.[2] The feature space of the kernel has an infinite number of dimensions; for , its expansion is:[3]

Approximations

Because support vector machines and other models employing the kernel trick do not scale well to large numbers of training samples or large numbers of features in the input space, several approximations to the RBF kernel (and similar kernels) have been devised.[4] Typically, these take the form z(x), i.e. a function transforming a single vector independently of other vectors (e.g. the support vectors in an SVM), such that

 

where   is the implicit mapping embedded in the RBF kernel.

One way to construct such a z is to randomly sample from the Fourier transformation of the kernel.[5]

See also

References

  1. ^ Yin-Wen Chang, Cho-Jui Hsieh, Kai-Wei Chang, Michael Ringgaard and Chih-Jen Lin (2010). Training and testing low-degree polynomial data mappings via linear SVM. J. Machine Learning Research 11:1471–1490.
  2. ^ a b Vert, Jean-Philippe, Koji Tsuda, and Bernhard Schölkopf (2004). "A primer on kernel methods." Kernel Methods in Computational Biology.
  3. ^ Shashua, Amnon (2009). "Introduction to Machine Learning: Class Notes 67577". 1. arXiv:0904.3664 [cs.LG]. {{cite arXiv}}: |access-date= requires |url= (help); Unknown parameter |accessdate= ignored (help); Unknown parameter |version= ignored (help)
  4. ^ Andreas Müller (2012). Kernel Approximations for Efficient SVMs (and other feature extraction methods).
  5. ^ Ali Rahimi and Benjamin Recht (2007). Random features for large-scale kernel machines. Neural Information Processing Systems.