Content deleted Content added
No edit summary |
m Reverted 1 edit by 2A14:B85:9641:9E1:7975:DDF0:AD23:C858 (talk) to last revision by PanHaspadar |
||
(39 intermediate revisions by 30 users not shown) | |||
Line 1:
{{short description|Probability distribution of the sum of random variables}}
The '''convolution of probability distributions''' arises in [[probability theory]] and [[statistics]] as the operation in terms of [[probability distribution]]s that corresponds to the addition of [[statistically independent|independent]] [[random variable]]s and, by extension, to forming linear combinations of random variables. The operation here is a special case of [[convolution]] in the context of probability distributions.▼
▲The '''convolution/sum of probability distributions''' arises in [[probability theory]] and [[statistics]] as the operation in terms of [[probability distribution]]s that corresponds to the addition of [[statistically independent|independent]] [[random variable]]s and, by extension, to forming linear combinations of random variables. The operation here is a special case of [[convolution]] in the context of probability distributions.
==Introduction==
The [[probability distribution]] of the sum of two or more [[independent (probability)|independent]] [[random variable]]s is the convolution of their individual distributions. The term is motivated by the fact that the [[probability mass function]] or [[probability density function]] of a sum of independent random variables is the [[convolution]] of their corresponding probability mass functions or probability density functions respectively. Many well known distributions have simple convolutions: see [[List of convolutions of probability distributions]]
The general formula for the distribution of the sum <math>Z=X+Y</math> of two independent integer-valued (and hence discrete) random variables is<ref>[[Susan P. Holmes|Susan Holmes ]](1998). Sums of Random Variables:
Statistics 116. Stanford. https://web.archive.org/web/20210413200454/http://statweb.stanford.edu/~susan/courses/s116/node114.html</ref>
:<math>P(Z=z) = \sum_{k=-\infty}^\infty P(X=k)P(Y=z-k)</math>
For independent, continuous random variables with [[probability density function]]s (PDF) <math>f,g</math> and [[cumulative distribution function]]s (CDF) <math>F,G</math> respectively, we have that the CDF of the sum is:
:<math>H(z)=\int_{-\infty}^\infty F(z-t)g(t) dt = \int_{-\infty}^\infty G(t)f(z-t) dt</math>
If we start with random variables <math>X</math> and <math>Y</math>, related by <math>Z = X + Y</math>, and with no information about their possible independence, then:
:<math>f_Z(z) = \int \limits_{-\infty}^{\infty} f_{XY}(x, z-x)~dx</math>
However, if <math>X</math> and <math>Y</math> are independent, then:
:<math>f_{XY}(x,y) = f_X(x) f_Y(y)</math>
and this formula becomes the convolution of probability distributions:
:<math>f_Z(z) = \int \limits_{-\infty}^{\infty} f_{X}(x)~f_Y(z-x)~dx</math>
== Example derivation ==
There are several ways of
One of the straightforward techniques is to use [[Characteristic function (probability theory)|characteristic functions]], which always exists and are unique to a given distribution.{{
=== Convolution of Bernoulli distributions ===
The convolution of two
:<math> \sum_{i=1}^2 \mathrm{Bernoulli}(p) \sim \mathrm{Binomial}(2,p)
To show this let
:<math>X_i \sim \mathrm{Bernoulli}(p), \quad 0<p<1, \quad 1 \le i \le 2</math>
and define
:<math>Y=\sum_{i=1}^2 X_i
Also, let ''Z'' denote a generic binomial random variable:
:<math>Z \sim \mathrm{Binomial}(2,p) \,\!
====Using probability mass functions====
Line 31 ⟶ 52:
&=p^n\left(1-p\right)^{2-n}\sum_{m\in\mathbb{Z}}\binom{1}{m}\binom{1}{n-m} \\
&=p^n\left(1-p\right)^{2-n}\left[\binom{1}{0}\binom{1}{n}+\binom{1}{1}\binom{1}{n-1}\right]\\
&=\binom{2}{n}p^n\left(1-p\right)^{2-n}=\mathbb{P}[Z=n]
\end{align}</math>
Here,
==== Using characteristic functions ====
Line 53 ⟶ 74:
== References ==
{{Reflist}}
* {{cite book | last1=Hogg | first1=Robert V. |authorlink1=Robert V. Hogg | last2=McKean | first2=Joseph W. | last3=Craig | first3=Allen T. | title=Introduction to mathematical statistics | edition=6th | publisher=Prentice Hall | url=http://www.pearsonhighered.com/educator/product/Introduction-to-Mathematical-Statistics/9780130085078.page | ___location=Upper Saddle River, New Jersey | year=2004 | pages=692 | ISBN=978-0-13-008507-8|MR=467974
[[Category:Theory of probability distributions]]
|