Content deleted Content added
Fgnievinski (talk | contribs) →Nonparametric multiplicative regression: split longest sec in this article |
Skbrownlee (talk | contribs) m grammer correction |
||
(43 intermediate revisions by 35 users not shown) | |||
Line 1:
{{Short description|Category of regression analysis}}
{{Regression bar}}
'''Nonparametric regression''' is a form of [[regression analysis]]
== Definition ==
==Gaussian process regression or Kriging==▼
Nonparametric regression assumes the following relationship, given the random variables <math>X</math> and <math>Y</math>:
{{Main|Gaussian process regression}}▼
:<math>
In Gaussian process regression, also known as Kriging, a Gaussian prior is assumed for the regression curve. The errors are assumed to have a [[multivariate normal distribution]] and the regression curve is estimated by its [[posterior mode]]. The Gaussian prior may depend on unknown hyperparameters, which are usually estimated via [[empirical Bayes]].▼
\mathbb{E}[Y\mid X=x] = m(x),
</math>
where <math>m(x)</math> is some deterministic function. [[Linear regression]] is a restricted case of nonparametric regression where <math>m(x)</math> is assumed to be a linear function of the data.
Sometimes a slightly stronger assumption of additive noise is used:
:<math>
Y = m(X) + U,
</math>
where the random variable <math>U</math> is the `noise term', with mean 0.
Without the assumption that <math>m</math> belongs to a specific parametric family of functions it is impossible to get an unbiased estimate for <math>m</math>, however most estimators are [[Consistency_(statistics)|consistent]] under suitable conditions.
== Common nonparametric regression algorithms ==
This is a non-exhaustive list of non-parametric models for regression.
* [[nearest neighbor smoothing]] (see also [[k-nearest neighbors algorithm]])
* [[regression tree|regression trees]]
* [[local regression]]
* [[multivariate adaptive regression splines]]
* [[smoothing spline|smoothing splines]]
* [[Artificial neural network|neural networks]]<ref>{{Cite journal |last=Cherkassky |first=Vladimir |last2=Mulier |first2=Filip |date=1994 |editor-last=Cheeseman |editor-first=P. |editor2-last=Oldford |editor2-first=R. W. |title=Statistical and neural network techniques for nonparametric regression |url=https://link.springer.com/chapter/10.1007/978-1-4612-2660-4_39 |journal=Selecting Models from Data |series=Lecture Notes in Statistics |language=en |___location=New York, NY |publisher=Springer |pages=383–392 |doi=10.1007/978-1-4612-2660-4_39 |isbn=978-1-4612-2660-4|url-access=subscription }}</ref>
== Examples ==
▲=== Gaussian process regression or Kriging ===
▲{{Main|Gaussian process regression}}
▲In Gaussian process regression, also known as Kriging, a Gaussian prior is assumed for the regression curve. The errors are assumed to have a [[multivariate normal distribution]] and the regression curve is estimated by its [[posterior mode]]. The Gaussian prior may depend on unknown hyperparameters, which are usually estimated via [[empirical Bayes]].
The hyperparameters typically specify a prior covariance kernel. In case the kernel should also be inferred nonparametrically from the data, the [[Information_field_theory#Critical_filter|critical filter]] can be used.
===
{{Main|Lasso (statistics)}}▼
▲==Kernel regression==
{{Main|Kernel regression}}
[[File:NonparRegrGaussianKernel.png|thumb| Example of a curve (red line) fit to a small data set (black points) with nonparametric regression using a Gaussian kernel smoother. The pink shaded area illustrates the kernel function applied to obtain an estimate of y for a given value of x. The kernel function defines the weight given to each data point in producing the estimate for a target point.]]{{Unreferenced section|date=August 2020}}
Kernel regression estimates the continuous dependent variable from a limited set of data points by [[Convolution|convolving]] the data points' locations with a [[kernel function]]—approximately speaking, the kernel function specifies how to "blur" the influence of the data points so that their values can be used to predict the value for nearby locations.
=== Regression trees ===▼
▲==Regression trees==
{{Main|Decision tree learning}}
Line 71 ⟶ 52:
|___location=Monterey, CA
|isbn=978-0-412-04841-8
}}</ref> Although the original Classification And Regression Tree (CART) formulation applied only to predicting univariate data, the framework can be used to predict multivariate data, including time series.<ref>{{Cite journal
| last = Segal
| first = M.R.
Line 82 ⟶ 63:
| jstor =2290271
| doi = 10.2307/2290271
| publisher = American Statistical Association, Taylor & Francis
}}</ref>
==See also==
* [[Local regression]]
* [[Non-parametric statistics]]
* [[Semiparametric regression]]
Line 94 ⟶ 78:
==Further reading==
* {{cite book |last=Bowman |first=A. W. |first2=A. |last2=Azzalini |year=1997 |title=Applied Smoothing Techniques for Data Analysis |publisher=Clarendon Press |___location=Oxford |isbn=
* {{cite book |last=Fan |first=J. |first2=I. |last2=Gijbels|author2-link= Irène Gijbels |year=1996 |title=Local Polynomial Modelling and its Applications |___location=Boca Raton |publisher=Chapman and Hall |isbn=0-412-98321-4 |url=https://books.google.com/books?id=BM1ckQKCXP8C }}
* {{cite book |last=Henderson |first=D. J. |first2=C. F. |last2=Parmeter |title=Applied Nonparametric Econometrics |___location=New York |publisher=Cambridge University Press |year=2015 |isbn=978-1-107-01025-3 |url=https://books.google.com/books?id=hD3WBQAAQBAJ }}
* {{cite book |last=Li |first=Q. |first2=J. |last2=Racine |year=2007 |title=Nonparametric Econometrics: Theory and Practice |___location=Princeton |publisher=Princeton University Press |isbn=978-0-691-12161-1 |url=https://books.google.com/books?id=BI_PiWazY0YC }}
* {{cite book |last=Pagan |first=A. |author-link=Adrian Pagan |first2=A. |last2=Ullah |year=1999 |title=Nonparametric Econometrics |___location=New York |publisher=Cambridge University Press |isbn=0-521-35564-8 |url=https://archive.org/details/nonparametriceco00paga |url-access=registration }}
==External links==
{{Commonscat}}
*
*[http://www.cs.tut.fi/~lasip Scale-adaptive nonparametric regression] (with Matlab software).
{{statistics}}
[[Category:Nonparametric regression| ]]
|