Content deleted Content added
Fgnievinski (talk | contribs) No edit summary |
Added short description Tags: Mobile edit Mobile app edit Android app edit App description add |
||
(6 intermediate revisions by one other user not shown) | |||
Line 1:
{{Short description|Statistics concept}}
In [[statistics]] the '''mean squared prediction error''' ('''MSPE'''), also known as '''mean squared error of the predictions''', of a [[smoothing]], [[curve fitting]], or [[regression (statistics)|regression]] procedure is the [[expected value]] of the [[Square (algebra)|squared]] '''prediction errors''' ('''PE'''), the [[squared deviation|square difference]] between the fitted values implied by the predictive function <math>\widehat{g}</math> and the values of the (unobservable) [[true value]] ''g''. It is an inverse measure of the
==Formulation==
▲In [[statistics]] the '''mean squared prediction error''' ('''MSPE'''), also known as '''mean squared error of the predictions''', of a [[smoothing]], [[curve fitting]], or [[regression (statistics)|regression]] procedure is the expected value of the [[squared deviation|square difference]] between the fitted values implied by the predictive function <math>\widehat{g}</math> and the values of the (unobservable) [[true value]] ''g''. It is an inverse measure of the [[explanatory power]] of <math>\widehat{g},</math> and can be used in the process of [[cross-validation (statistics)|cross-validation]] of an estimated model.
If the smoothing or fitting procedure has [[projection matrix]] (i.e., hat matrix) ''L'', which maps the observed values vector <math>y</math> to [[predicted value]]s vector <math>\hat{y}=Ly,</math> then PE and MSPE are formulated as:▼
:<math>\operatorname{PE_i}=g(x_i)-\widehat{g}(x_i),</math>
▲If the smoothing or fitting procedure has [[projection matrix]] (i.e., hat matrix) ''L'', which maps the observed values vector <math>y</math> to [[predicted value]]s vector <math>\hat{y}=Ly,</math> then
:<math>\operatorname{MSPE}
The MSPE can be decomposed into two terms: the squared [[bias (statistics)|bias]] (mean error) of the fitted values and the [[variance]]
:<math>
:<math>\operatorname{ME}=\operatorname{E}\left[ \widehat{g}(x_i) - g(x_i)\right]</math>
:<math>\operatorname{VAR}=\operatorname{E}\left[\left(\widehat{g}(x_i) - \operatorname{E}\left[{g}(x_i)\right]\right)^2\right].</math>
The quantity {{math|SSPE{{=}}''n''MSPE}} is called '''sum squared prediction error'''
The '''root mean squared prediction error'''
==Computation of MSPE over out-of-sample data==
Line 56:
==See also==
* [[Akaike information criterion]]
* [[Bias-variance tradeoff]]
* [[Mean squared error]]
* [[Errors and residuals in statistics]]
* [[Law of total variance]]
* [[Mallows's Cp|Mallows's ''C<sub>p</sub>'']]
{{Machine learning evaluation metrics}}▼
* [[Model selection]]
==
{{reflist}}
▲*{{cite book |first1=Robert S. |last1=Pindyck |authorlink=Robert Pindyck |first2=Daniel L. |last2=Rubinfeld |authorlink2=Daniel L. Rubinfeld |title=Econometric Models & Economic Forecasts |___location=New York |publisher=McGraw-Hill |edition=3rd |year=1991 |isbn=0-07-050098-3 |chapter=Forecasting with Time-Series Models |pages=[https://archive.org/details/econometricmodel00pind/page/516 516–535] |chapter-url=https://archive.org/details/econometricmodel00pind/page/516 }}
▲{{Machine learning evaluation metrics}}
{{DEFAULTSORT:Mean Squared Prediction Error}}
|