Content deleted Content added
Undid revision 1216542234 by 95.83.136.159 (talk) Tags: Undo Mobile edit Mobile web edit Advanced mobile edit |
|||
(60 intermediate revisions by 32 users not shown) | |||
Line 1:
{{Short description|Statistical concept}}
{{More footnotes|date=November 2010}}
In [[statistics]], the '''coefficient of
The coefficient of multiple correlation takes values between 0 and 1. Higher values indicate higher predictability of the [[dependent and independent variables|dependent variable]] from the [[dependent and independent variables|independent variables]], with a value of 1 indicating that the predictions are exactly correct and a value of 0 indicating that no linear combination of the independent variables is a better predictor than is the fixed [[mean]] of the dependent variable.<ref>[http://mtweb.mtsu.edu/stats/regression/level3/multicorrel/multicorrcoef.htm Multiple correlation coefficient]</ref>
{| class="wikitable"
|Correlation Coefficient (r)
|Direction and Strength of Correlation
|-
|1
|Perfectly positive
|-
|0.8
|Strongly positive
|-
|0.5
|Moderately positive
|-
|0.2
|Weakly positive
|-
|0
|No association
|-
| -0.2
|Weakly negative
|-
| -0.5
|Moderately negative
|-
| -0.8
|Strongly negative
|-
| -1
|Perfectly negative
|}
The coefficient of multiple correlation is known as the square root of the [[coefficient of determination]], but under the particular assumptions that an intercept is included and that the best possible linear predictors are used, whereas the coefficient of determination is defined for more general cases, including those of nonlinear prediction and those in which the predicted values have not been derived from a model-fitting procedure.
==Definition==
The coefficient of multiple correlation, denoted ''R''
==Computation==
The square of the coefficient of multiple correlation can be computed using the [[Euclidean space|vector]]
::<math>R^2 = \mathbf{c}^\top R_{xx}^{-1}\, \mathbf{c},</math>
where
::<math>R_{xx} = \left(\begin{array}{cccc}
▲where ''c'' ' is the [[transpose]] of ''c'', and ''R''<sub>''xx''</sub><sup>−1</sup> is [[Matrix inversion|inverse]] of the matrix ''R''<sub>''xx''</sub>.
r_{x_1 x_1} & r_{x_1 x_2} & \dots & r_{x_1 x_N} \\
r_{x_2 x_1} & \ddots & & \vdots \\
\vdots & & \ddots & \\
r_{x_N x_1} & \dots & & r_{x_N x_N}
\end{array}\right).</math>
If all the predictor variables are uncorrelated, the matrix
The squared coefficient of multiple correlation can also be computed as the fraction of variance of the dependent variable that is explained by the independent variables, which in turn is 1 minus the unexplained fraction. The unexplained fraction can be computed as the [[sum of
==Properties==
==References==
{{Reflist}}
==Further reading==
*
*
* Crown, William H. (1998). ''Statistical Models for the Social and Behavioral Sciences: Multiple Regression and Limited-Dependent Variable Models''. {{ISBN|0275953165}}
*
*
* Fred N. Kerlinger, Elazar J. Pedhazur (1973). ''Multiple Regression in Behavioral Research.'' New York: Holt Rinehart Winston. {{ISBN|9780030862113}}
* Stanton, Jeffrey M. (2001). [http://www.amstat.org/publications/jse/v9n3/stanton.html "Galton, Pearson, and the Peas: A Brief History of Linear Regression for Statistics Instructors"], ''Journal of Statistics Education'', 9 (3).
{{DEFAULTSORT:Multiple Correlation}}
[[Category:Correlation indicators]]
[[Category:Regression analysis]]
|