Content deleted Content added
More verbose definition with an added reference |
MichaelMaggs (talk | contribs) Adding short description: "Concept in information processing" |
||
(12 intermediate revisions by 10 users not shown) | |||
Line 1:
{{Short description|Concept in information processing}}
The '''
==
Let three random variables form the [[Markov chain]] <math>X \rightarrow Y \rightarrow Z</math>, implying that the conditional distribution of <math>Z</math> depends only on <math>Y</math> and is [[Conditional independence|conditionally independent]] of <math>X</math>. Specifically, we have such a Markov chain if the joint probability mass function can be written as
:<math>p(x,y,z) = p(x)p(y|x)p(z|y)=p(y)p(x|y)p(z|y)</math>
In this setting, no processing of <math>Y
:<math> I(
==Proof==
One can apply the [[Conditional_mutual_information#Chain_rule_for_mutual_information chain rule for mutual information|chain rule for mutual information]] to obtain two different decompositions of <math>I(X;Y,Z)</math>:
:<math>
I(X;Z) + I(X;Y\mid Z) = I(X;Y,Z) = I(X;Y) + I(X;Z\mid Y)
</math>
By the relationship <math>X \rightarrow Y \rightarrow Z</math>, we know that <math>X</math> and <math>Z</math> are conditionally independent, given <math>Y</math>, which means the [[conditional mutual information]], <math>I(X;Z\mid Y)=0</math>. The data processing inequality then follows from the non-negativity of <math>I(X;Y\mid Z)\ge0</math>.
==See also==
|