Content deleted Content added
Undid revision 1227126801 by 128.153.179.53 (talk) rv error |
|||
(17 intermediate revisions by 16 users not shown) | |||
Line 1:
{{Short description|Variance of a random variable given value of other variables}}
In [[probability theory]] and [[statistics]], a '''conditional variance''' is the [[variance]] of a [[random variable]] given the value(s) of one or more other variables.
Particularly in [[econometrics]], the conditional variance is also known as the '''scedastic function''' or '''skedastic function'''.<ref>{{cite book |first=Aris |last=Spanos |chapter=Conditioning and regression |title=Probability Theory and Statistical Inference |___location=New York |publisher=Cambridge University Press |year=1999 |isbn=0-521-42408-9 |pages=339–356 [p. 342] |url=https://books.google.com/books?id=G0_HxBubGAwC&pg=PA342 }}</ref> Conditional variances are important parts of [[autoregressive conditional heteroskedasticity]] (ARCH) models.
Line 5 ⟶ 6:
The conditional variance of a [[random variable]] ''Y'' given another random variable ''X'' is
:<math>\operatorname{Var}(Y
The conditional variance tells us how much variance is left if we use <math>\operatorname{E}(Y\mid X)</math> to "predict" ''Y''.
Here, as usual, <math>\operatorname{E}(Y\mid X)</math> stands for the [[conditional expectation]] of ''Y'' given ''X'',
which we may recall, is a random variable itself (a function of ''X'', determined up to probability one).
As a result, <math>\operatorname{Var}(Y
==Explanation, relation to
{{main|least-squares}}
Recall that variance is the expected squared deviation between a random variable (say, ''Y'') and its expected value.
The expected value can be thought of as a reasonable prediction of the outcomes of the random experiment (in particular, the expected value is the best constant prediction when predictions are assessed by expected squared prediction error). Thus, one interpretation of variance is that it gives the smallest possible expected squared prediction error. If we have the knowledge of another random variable (''X'') that we can use to predict ''Y'', we can potentially use this knowledge to reduce the expected squared error. As it turns out, the best prediction of ''Y'' given ''X'' is the conditional expectation. In particular, for any <math>f: \mathbb{R} \to \mathbb{R}</math> measurable,
Line 21 ⟶ 23:
&= \operatorname{E}[ (Y-\operatorname{E}(Y|X)\,\,+\,\, \operatorname{E}(Y|X)-f(X) )^2 ] \\
&= \operatorname{E}[ \operatorname{E}\{ (Y-\operatorname{E}(Y|X)\,\,+\,\, \operatorname{E}(Y|X)-f(X) )^2|X\} ] \\
&= \operatorname{E}[\operatorname{Var}( Y| X )] + \operatorname{E}[(\operatorname{E}(Y|X)-f(X))^2]\,.
\end{align}
</math>
Line 31 ⟶ 33:
==Special cases, variations==
===Conditioning on discrete random variables===
When ''X'' takes on countable many values <math>S = \{x_1,
:<math>\operatorname{Var}(Y|X=x) = \operatorname{E}((Y - \operatorname{E}(Y\mid X=x))^{2}\mid X=x)=\operatorname{E}(Y^2|X=x)-\operatorname{E}(Y|X=x)^2,</math>
where recall that <math>\operatorname{E}(Z\mid X=x)</math> is the [[Conditional_expectation#Conditional_expectation_with_respect_to_a_random_variable|conditional expectation of ''Z'' given that ''X=x'']], which is well-defined for <math>x\in S</math>.
Line 49 ⟶ 51:
In particular, letting <math>P_{Y|X}</math> be the (regular) [[conditional distribution]] <math>P_{Y|X}</math> of ''Y'' given ''X'', i.e., <math>P_{Y|X}:\mathcal{B} \times \mathbb{R}\to [0,1]</math> (the intention is that <math>P_{Y|X}(U,x) = P(Y\in U|X=x)</math> almost surely over the support of ''X''), we can define
<math> \operatorname{Var}(Y|X=x) = \int \left(y- \int y' P_{Y|X}(dy'|x)\right)^2 P_{Y|X}(dy|x). </math>
This can, of course, be specialized to when ''Y'' is discrete itself (replacing the integrals with sums), and also when the [[conditional density]] of ''Y'' given ''X=x'' with respect to some underlying distribution exists.
Line 58 ⟶ 60:
<math>\operatorname{Var}(Y) = \operatorname{E}(\operatorname{Var}(Y\mid X))+\operatorname{Var}(\operatorname{E}(Y\mid X)).</math>
In words: the variance of ''Y'' is the sum of the expected conditional variance of ''Y'' given ''X'' and the variance of the conditional expectation of ''Y'' given ''X''. The first term captures the variation left after "using ''X'' to predict ''Y''", while the second term captures the variation due to the mean of the prediction of ''Y'' due to the randomness of ''X''.
==See also==
*[[Mixed model]]
*[[Random effects model]]
==References==
Line 64 ⟶ 70:
==Further reading==
* {{cite book |first=George |last=Casella |first2=Roger L. |last2=Berger |title=Statistical Inference
[[Category:Statistical deviation and dispersion]]
[[Category:Theory of probability distributions]]
[[Category:Conditional probability]]
{{statistics-stub}}
|