Content deleted Content added
+ "," to break verbose phrases. Minor C/E fixes. + inline refs. - dup empty lines. "occurrencies" -> "occurrences". Typeset inline math properly. Lc "categorical distribution". + cm on geometric distribution. Tags: Reverted Visual edit |
Reverted 1 edit by Miaumee (talk): Per User talk:Miaumee, this is apparently the preferred response to poor editing |
||
Line 1:
{{Short description|Discrete-variable probability distribution}}
[[Image:Discrete probability distrib.svg|right|thumb|The graph of a probability mass function. All the values of this function must be non-negative and sum up to 1.]]
In [[probability theory|probability]] and [[statistics]], a '''probability mass function''' ('''PMF''')
A probability mass function differs from a [[probability density function]] (PDF) in that the latter is associated with continuous rather than discrete random variables. A PDF must be [[integration (mathematics)|integrated]] over an interval to yield a probability.<ref name=":0">{{Cite book|title=A modern introduction to probability and statistics : understanding why and how|date=2005|publisher=Springer|others=Dekking, Michel, 1946-|isbn=978-1-85233-896-1|___location=London|oclc=262680588}}</ref>
The value of the random variable
==Formal definition==
Probability mass function is the probability distribution of a discrete random variable, and provides the possible values and their associated probabilities. It is the function <math>p:\mathbb{\R}</math> <math>\rightarrow [0,1]</math> defined by
{{Equation box 1
|indent =
Line 20:
|background colour=#F5FFFA}}
for <math>-\infin < x < \infin</math>,<ref name=":0" /> where <math>P</math> is a [[probability measure]]. <math>p_X(x)</math> can also be simplified as <math>p(x)</math>.<ref>{{Cite book|title=Engineering optimization : theory and practice|last=Rao, Singiresu S., 1944-|date=1996|publisher=Wiley|isbn=0-471-55034-5|edition=3rd|___location=New York|oclc=62080932}}</ref
The probabilities associated with each possible values must be positive and sum up to 1. For all other values, the probabilities need to be 0.
Line 28:
:<math>p(x) = 0</math> for all other x
Thinking of probability as mass
==Measure theoretic formulation==
Line 36:
Suppose that <math>(A, \mathcal A, P)</math> is a [[probability space]]
and that <math>(B, \mathcal B)</math> is a measurable space whose underlying [[sigma algebra|σ-algebra]] is discrete, so in particular contains singleton sets of <math>B</math>. In this setting, a random variable <math> X \colon A \to B</math> is discrete
The [[pushforward measure]] <math>X_{*}(P)</math>—called a distribution of <math>X</math> in this context—is a probability measure on <math>B</math> whose restriction to singleton sets induces a probability mass function <math>f_X \colon B \to \mathbb R</math>
Now
:<math>P(X=b)=P(X^{-1}( \{ b \} )) := \int_{X^{-1}(\{b \})} dP =</math><math>\int_{ \{b \}} f d \mu = f(b),</math>
demonstrating that <math>f</math> is in fact a probability mass function.
When there is a natural order among the potential outcomes <math>x</math>, it may be convenient to assign numerical values to them (or ''n''-tuples in case of a discrete [[multivariate random variable]]), and to consider also values not in the [[Image (mathematics)|image]] of <math>X</math>. That is, <math>f_X</math> may be defined for all [[real number]]s and <math>f_X(x)=0</math> for all <math>x \notin X(S)</math> as shown in the figure.▼
▲When there is a natural order among the potential outcomes <math>x</math>, it may be convenient to assign numerical values to them (or ''n''-tuples in case of a discrete [[multivariate random variable]])
The image of <math>X</math> has a [[countable]] subset on which the probability mass function <math>f_X(x)</math> is one. Consequently, the probability mass function is zero for all but a countable number of values of <math>x</math>.
The discontinuity of probability mass functions is related to the fact that the [[cumulative distribution function]] of a discrete random variable is also discontinuous. If <math>X</math> is a discrete random variable, then <math> P(X = x) = 1</math> means that the casual event <math>(X = x)</math> is certain (it is true in the 100% of the
==Examples==
Line 54 ⟶ 56:
===Finite===
*[[Bernoulli distribution]], Ber(p),
:<math>p_X(x) = \begin{cases} p, & \text{if }x\text{ is 1} \\ 1-p, & \text{if }x\text{ is 0} \end{cases}</math>
Line 62 ⟶ 64:
::<math>p_X(x) = \begin{cases}\frac{1}{2}, &x \in \{0, 1\},\\0, &x \notin \{0, 1\}.\end{cases}</math>
*[[Binomial distribution]], Bin(n,p),
:An example of the Binomial distribution is the probability of getting exactly one 6
*
:An example is tossing the coin until the first head appears.
::
Other distributions that can be modeled using a probability mass function are the [[
* If the discrete distribution has two or more categories
* An example of a [[Joint probability distribution|multivariate discrete distribution]], and of its probability mass function, is provided by the [[multinomial distribution]]. Here
===Infinite
*The following exponentially declining distribution is an example of a distribution with an infinite number of possible outcomes—all the positive integers:
Line 88 ⟶ 91:
{{Main|Joint probability distribution}}
Two or more discrete random variables have a joint probability mass function, which gives the probability of each possible combination of realizations for the random variables.
==References==
|