Content deleted Content added
Revert to revision 750282630 dated 2016-11-18 20:47:34 by Marcocapelle using popups per WP:REFSPAM |
Skbrownlee (talk | contribs) m grammer correction |
||
(29 intermediate revisions by 26 users not shown) | |||
Line 1:
{{Short description|Category of regression analysis}}
{{Regression bar}}
'''Nonparametric regression''' is a
== Definition ==
==Gaussian process regression or Kriging==▼
Nonparametric regression assumes the following relationship, given the random variables <math>X</math> and <math>Y</math>:
:<math>
\mathbb{E}[Y\mid X=x] = m(x),
</math>
where <math>m(x)</math> is some deterministic function. [[Linear regression]] is a restricted case of nonparametric regression where <math>m(x)</math> is assumed to be a linear function of the data.
Sometimes a slightly stronger assumption of additive noise is used:
:<math>
Y = m(X) + U,
</math>
where the random variable <math>U</math> is the `noise term', with mean 0.
Without the assumption that <math>m</math> belongs to a specific parametric family of functions it is impossible to get an unbiased estimate for <math>m</math>, however most estimators are [[Consistency_(statistics)|consistent]] under suitable conditions.
== Common nonparametric regression algorithms ==
This is a non-exhaustive list of non-parametric models for regression.
* [[nearest neighbor smoothing]] (see also [[k-nearest neighbors algorithm]])
* [[regression tree|regression trees]]
* [[kernel regression]]
* [[local regression]]
* [[multivariate adaptive regression splines]]
* [[smoothing spline|smoothing splines]]
* [[Artificial neural network|neural networks]]<ref>{{Cite journal |last=Cherkassky |first=Vladimir |last2=Mulier |first2=Filip |date=1994 |editor-last=Cheeseman |editor-first=P. |editor2-last=Oldford |editor2-first=R. W. |title=Statistical and neural network techniques for nonparametric regression |url=https://link.springer.com/chapter/10.1007/978-1-4612-2660-4_39 |journal=Selecting Models from Data |series=Lecture Notes in Statistics |language=en |___location=New York, NY |publisher=Springer |pages=383–392 |doi=10.1007/978-1-4612-2660-4_39 |isbn=978-1-4612-2660-4|url-access=subscription }}</ref>
== Examples ==
▲=== Gaussian process regression or Kriging ===
{{Main|Gaussian process regression}}
In Gaussian process regression, also known as Kriging, a Gaussian prior is assumed for the regression curve. The errors are assumed to have a [[multivariate normal distribution]] and the regression curve is estimated by its [[posterior mode]]. The Gaussian prior may depend on unknown hyperparameters, which are usually estimated via [[empirical Bayes]].
The hyperparameters typically specify a prior covariance kernel. In case the kernel should also be inferred nonparametrically from the data, the [[Information_field_theory#Critical_filter|critical filter]] can be used.
[[Smoothing splines]] have an interpretation as the posterior mode of a Gaussian process regression.
=== Kernel regression ===
{{Main|Kernel regression}}
[[File:NonparRegrGaussianKernel.png|thumb| Example of a curve (red line) fit to a small data set (black points) with nonparametric regression using a Gaussian kernel smoother. The pink shaded area illustrates the kernel function applied to obtain an estimate of y for a given value of x. The kernel function defines the weight given to each data point in producing the estimate for a target point.]]{{Unreferenced section|date=August 2020}}
Kernel regression estimates the continuous dependent variable from a limited set of data points by [[Convolution|convolving]] the data points' locations with a [[kernel function]]—approximately speaking, the kernel function specifies how to "blur" the influence of the data points so that their values can be used to predict the value for nearby locations.
=== Regression trees ===▼
▲==Regression trees==
{{Main|Decision tree learning}}
Line 62 ⟶ 52:
|___location=Monterey, CA
|isbn=978-0-412-04841-8
}}</ref> Although the original Classification And Regression Tree (CART) formulation applied only to predicting univariate data, the framework can be used to predict multivariate data, including time series.<ref>{{Cite journal
| last = Segal
| first = M.R.
Line 73 ⟶ 63:
| jstor =2290271
| doi = 10.2307/2290271
| publisher = American Statistical Association, Taylor & Francis
}}</ref>
==See also==
* [[Lasso (statistics)]]
* [[Local regression]]
* [[Non-parametric statistics]]
Line 87 ⟶ 79:
==Further reading==
* {{cite book |last=Bowman |first=A. W. |first2=A. |last2=Azzalini |year=1997 |title=Applied Smoothing Techniques for Data Analysis |publisher=Clarendon Press |___location=Oxford |isbn=0-19-852396-3 |url=https://books.google.com/books?id=7WBMrZ9umRYC }}
* {{cite book |last=Fan |first=J. |first2=I. |last2=Gijbels|author2-link= Irène Gijbels |year=1996 |title=Local Polynomial Modelling and its Applications |___location=Boca Raton |publisher=Chapman and Hall |isbn=0-412-98321-4 |url=https://books.google.com/books?id=BM1ckQKCXP8C }}
* {{cite book |last=Henderson |first=D. J. |first2=C. F. |last2=Parmeter |title=Applied Nonparametric Econometrics |___location=New York |publisher=Cambridge University Press |year=2015 |isbn=978-1-107-01025-3 |url=https://books.google.com/books?id=hD3WBQAAQBAJ }}
* {{cite book |last=Li |first=Q. |first2=J. |last2=Racine |year=2007 |title=Nonparametric Econometrics: Theory and Practice |___location=Princeton |publisher=Princeton University Press |isbn=978-0-691-12161-1 |url=https://books.google.com/books?id=BI_PiWazY0YC }}
* {{cite book |last=Pagan |first=A. |
==External links==
{{Commonscat}}
*
*[http://www.cs.tut.fi/~lasip Scale-adaptive nonparametric regression] (with Matlab software).
{{statistics}}
|