Conditional probability distribution: Difference between revisions

Content deleted Content added
Jt68 (talk | contribs)
No edit summary
Undid revision 339337438 by Jt68 (talk)See Talk
Line 8:
Similarly for [[continuous random variable]]s, the conditional [[probability density function]] can be written as ''f''<sub>''Y''</sub>(''y'' | ''X=x'') and this is
 
:<math>f_{Yf_Y(y \mid X=x}(y) = \frac{f_{X, Y}(x, y)}{f_X(x)}= \frac{f_{Xf_X(x \mid Y=y}(x)f_Y(y)}{f_X(x)}, </math>
 
where ''f''<sub>''X'',''Y''</sub>(x, y) gives the [[joint distribution|joint density]] of ''X'' and ''Y'', while ''f''<sub>''X''</sub>(''x'') gives the [[marginal density]] for ''X''.
Line 14:
The concept of the conditional distribution of a continuous random variable is not as intuitive as it might seem: [[Borel's paradox]] shows that conditional probability density functions need not be invariant under coordinate transformations.
 
If for discrete random variables ''P''(''Y'' = ''y'' | ''X'' = ''x'') = ''P''(''Y'' = ''y'') for all ''x'' and ''y'', or for continuous random variables ''f''<sub>''Y'' | ''X=x''</sub>(''y'' | ''X=x'') = ''f''<sub>''Y''</sub>(''y'') for all x and y, then ''Y'' is said to be [[Statistical independence|independent]] of ''X'' (and this implies that ''X'' is also independent of ''Y'').
 
Seen as a function of ''y'' for given ''x'', ''P''(''Y'' = ''y'' | ''X'' = ''x'') is a probability and so the sum over all ''y'' (or integral if it is a conditional probability density) is 1. Seen as a function of ''x'' for given ''y'', it is a [[likelihood function]], so that the sum over all ''x'' need not be 1.