Content deleted Content added
m Disambiguated: support → support (mathematics) |
User357269 (talk | contribs) cleaned up based on feedback |
||
Line 1:
In [[probability theory]], '''regular conditional probability''' is a concept that formalizes the notion of conditioning on the outcome of a [[random variable]]. The resulting '''conditional probability distribution''' is a parametrized family of probability measures called a [[Markov kernel]].
==
=== Conditional probability distribution ===
:<math>P(Y \in A | X = x) = \frac{P(Y \in A, X = x)}{P(X=x)}.</math>▼
:<math>\nu(x, A) = P(A | X =x)</math>▼
The conditional probability of ''Y'' being in ''A'' is given by▼
:<math>P(Y \in A | X = x) = \frac{\int_A f_{X,Y}(x, y) \mathrm{d}y}{\int_\mathbb{R} f_{X,Y}(x, y) \mathrm{d}y}.</math>▼
If the random variable ''X'' is discrete
:<math>\kappa_{Y|X}(x, A) = P(Y \in A | X = x) = \begin{cases}
\frac{P(Y \in A, X = x)}{P(X=x)} & \text{ if } P(X = x) > 0 \\
\text{arbitrary value} & \text{ otherwise}.
\end{cases}</math>
If the random variables ''X'', ''Y'' are continuous with density <math>f_{X,Y}(x,y)</math>.
==Relation to conditional expectation==▼
:<math>\kappa_{Y|X}(x, A) = \begin{cases}
▲
\text{ if } \int_\mathbb{R} f_{X,Y}(x, y) \mathrm{d}y > 0 \\
\text{arbitrary value} & \text{ otherwise}.
\end{cases}</math>
A more general definition can be given in terms of [[conditional expectation]]. Consider a function <math> e_{Y \in A} : \mathbb{R} \to [0,1]</math> satisfying
In probability theory, the theory of [[conditional expectation]] is developed before that of regular conditional distributions.<ref>{{cite book |last1=Durrett |first1=Richard |title=Probability : theory and examples |date=2010 |publisher=Cambridge University Press |___location=Cambridge |isbn=9780521765398 |edition=4th}}</ref><ref>{{cite book |last1=Klenke |first1=Achim |title=Probability theory : a comprehensive course |___location=London |isbn=978-1-4471-5361-0 |edition=Second}}</ref>▼
▲:<math>
for almost all <math>\omega</math>.
As with conditional expectation, this can be further generalized to conditioning on a sigma algebra <math>\mathcal{F}</math>. In that case the conditional distribution is a function <math>\Omega \times \mathcal{B}(\mathbb{R}) \to [0, 1]</math>:
For discrete and continuous random variables, the conditional expectation is given by▼
:<math> \
=== Regularity ===
For working with <math>\kappa_{Y|X}</math>, it is important that it be ''regular'', that is:
* For almost all ''x'', <math>A \mapsto \kappa_{Y|X}(x, A)</math> is a probability measure
* For all ''A'', <math>x \mapsto \kappa_{Y|X}(x, A)</math> is a measurable function
The former is trivial, but the proof of the latter is more involved. It can be shown that if ''Y'' is a random element <math>\Omega \to S</math> in a [[Radon space]] ''S'', there exists a <math>\kappa_{Y|X}</math> that satisfies the measurability condition.<ref>{{cite book |last1=Klenke |first1=Achim |title=Probability theory : a comprehensive course |___location=London |isbn=978-1-4471-5361-0 |edition=Second}}</ref> It is possible to construct more general spaces where a regular conditional probability distribution does not exist.<ref>Faden, A.M., 1985. The existence of regular conditional probabilities: necessary and sufficient conditions. ''The Annals of Probability'', 13(1), pp.288-298.</ref>
▲=== Relation to conditional expectation ===
▲In probability theory, the theory of [[conditional expectation]] is developed before that of regular conditional distributions.<ref>{{cite book |last1=Durrett |first1=Richard |title=Probability : theory and examples |date=2010 |publisher=Cambridge University Press |___location=Cambridge |isbn=9780521765398 |edition=4th
▲For discrete and continuous random variables, the conditional expectation
:<math>
\begin{aligned}
\mathbb{E}[
\mathbb{E}[Y|X
\end{aligned}
</math>
where <math>f_{
:<math>\mathbb{E}[
== Formal definition==
▲:<math>\nu(\omega, A) = \mathbb{E}[1_{X \in A} | Y](\omega).</math>
Let <math>(\Omega, \mathcal F, P)</math> be a [[probability space]], and let <math>T:\Omega\rightarrow E</math> be a [[random variable]], defined as a [[Borel measure|Borel-]][[measurable function]] from <math>\Omega</math> to its [[Probability space#Random variables|state space]] <math>(E, \mathcal E)</math>.
One should think of <math>T</math> as a way to "disintegrate" the sample space <math>\Omega</math> into <math>\{ T^{-1}(x) \}_{x \in E}</math>.
Line 48 ⟶ 64:
:<math>P\big(A\cap T^{-1}(B)\big) = \int_B \nu(x,A) \,P\big(T^{-1}(d x)\big).</math>
where <math>P\circ T^{-1}</math> is the [[pushforward measure]] <math>T_*P</math> of the distribution of the random element <math>T</math>,
<math>x\in\mathrm{supp}\,T,</math> i.e. the [[Support (measure theory)|
Specifically, if we take <math>B=E</math>, then <math>A \cap T^{-1}(E) = A</math>, and so
:<math>P(A) = \int_E \nu(x,A) \,P\big(T^{-1}(d x)\big)</math>,
where <math>\nu(x, A)</math> can be denoted, using more familiar terms <math>P(A\ |\ T=x)</math>.
==Alternate definition==
Line 71 ⟶ 80:
:<math>\left|\frac {P(A\cap V)}{P(V)}-L\right| < \epsilon,</math>
where <math>L = P (A|T=t)</math> is the limit.
==See also==
|