Content deleted Content added
No edit summary Tag: Reverted |
whether to call it PMF or discrete PDF is a non-generally accepted convention. Hence, the differentiation must be against a continuous PDF and not against a PDF in general. |
||
(42 intermediate revisions by 26 users not shown) | |||
Line 1:
{{Short description|Discrete-variable probability distribution}}
[[Image:Discrete probability distrib.svg|right|thumb|The graph of a probability mass function. All the values of this function must be non-negative and sum up to 1.]]
In [[probability theory|probability
A probability mass function differs from a [[probability density function|continuous probability density function]] (PDF) in that the latter is associated with continuous rather than discrete random variables. A continuous PDF must be [[integration (mathematics)|integrated]] over an interval to yield a probability.<ref name=":0">{{Cite book|title=A modern introduction to probability and statistics : understanding why and how|date=2005|publisher=Springer|others=Dekking, Michel, 1946-|isbn=978-1-85233-896-1|___location=London|oclc=262680588}}</ref>
The value of the random variable having the largest probability mass is called the [[mode (statistics)|mode]].
==Formal definition==
Probability mass function is the probability distribution of a [[discrete random variable]], and provides the possible values and their associated probabilities. It is the function <math>p:
▲Probability mass function is the probability distribution of a discrete random variable, and provides the possible values and their associated probabilities. It is the function <math>p:\mathbb{\R}</math> <math>\rightarrow [0,1]</math> defined by
{{Equation box 1
|indent =
|title=
|equation = <math>p_X(
|cellpadding= 6
|border
Line 20 ⟶ 18:
|background colour=#F5FFFA}}
for <math>-\infin < x < \infin</math>,<ref name=":0" /> where <math>P</math> is a [[probability measure]]. <math>p_X(x)</math> can also be simplified as <math>p(x)</math>.<ref>{{Cite book|title=Engineering optimization : theory and practice| last=Rao
The probabilities associated with
Thinking of probability as mass helps to avoid mistakes since the physical mass is [[Conservation of mass|conserved]] as is the total probability for all hypothetical outcomes <math>x</math>.
==Measure theoretic formulation==
A probability mass function of a discrete random variable <math>X</math> can be seen as a special case of two more general measure theoretic constructions:
the [[probability distribution|distribution]] of <math>X</math> and the [[probability density function]] of <math>X</math> with respect to the [[counting measure]]. We make this more precise below.
Line 37 ⟶ 32:
Suppose that <math>(A, \mathcal A, P)</math> is a [[probability space]]
and that <math>(B, \mathcal B)</math> is a measurable space whose underlying [[sigma algebra|σ-algebra]] is discrete, so in particular contains singleton sets of <math>B</math>. In this setting, a random variable <math> X \colon A \to B</math> is discrete provided its image is countable.
The [[pushforward measure]] <math>X_{*}(P)</math>—called
Now suppose that <math>(B, \mathcal B, \mu)</math> is a [[measure space]] equipped with the counting measure
demonstrating that <math>f</math> is in fact a probability mass function.
When there is a natural order among the potential outcomes <math>x</math>, it may be convenient to assign numerical values to them (or ''n''-tuples in case of a discrete [[multivariate random variable]]) and to consider also values not in the [[Image (mathematics)|image]] of <math>X</math>. That is, <math>f_X</math> may be defined for all [[real number]]s and <math>f_X(x)=0</math> for all <math>x \notin X(S)</math> as shown in the figure.
Line 50 ⟶ 43:
The image of <math>X</math> has a [[countable]] subset on which the probability mass function <math>f_X(x)</math> is one. Consequently, the probability mass function is zero for all but a countable number of values of <math>x</math>.
The discontinuity of probability mass functions is related to the fact that the [[cumulative distribution function]] of a discrete random variable is also discontinuous. If <math>X</math> is a discrete random variable, then <math> P(X = x) = 1</math> means that the casual event <math>(X = x)</math> is certain (it is true in
==Examples==
Line 56 ⟶ 49:
===Finite===
There are three major distributions associated, the [[Bernoulli distribution]], the [[
*[[Bernoulli distribution]], Ber(p), is used to model an experiment with only two possible outcomes. The two outcomes are often encoded as 1 and 0. ▼
:An example of the Bernoulli distribution is tossing a coin. Suppose that <math>S</math> is the sample space of all outcomes of a single toss of a fair coin, and <math>X</math> is the random variable defined on <math>S</math> assigning 0 to the category "tails" and 1 to the category "heads". Since the coin is fair, the probability mass function is▼
*[[Binomial distribution]], Bin(n,p), models the number of successes when someone draws n times with replacement. Each draw or experiment is independent, with two possible outcomes. The associated probability mass function is<math>\binom{n}{k}p^k (1-p)^{n-k}</math>. [[Image:Fair dice probability distribution.svg|right|thumb|The probability mass function of a [[Dice|fair die]]. All the numbers on the {{dice}} have an equal chance of appearing on top when the die stops rolling.]]▼
▲*
p, & \text{if }x\text{ is 1} \\
1-p, & \text{if }x\text{ is 0}
▲
\frac{1}{2}, &x = 0,\\
\frac{1}{2}, &x = 1,\\
0, &x \notin \{0, 1\}.
\end{cases}</math>
▲*
* Geometric distribution describes the number of trials needed to get one success. Its probability mass function is <math display="inline">p_X(k) = (1-p)^{k-1} p</math>.{{pb}}An example is tossing a coin until the first "heads" appears. <math>p</math> denotes the probability of the outcome "heads", and <math>k</math> denotes the number of necessary coin tosses. {{pb}}Other distributions that can be modeled using a probability mass function are the [[categorical distribution]] (also known as the generalized Bernoulli distribution) and the [[multinomial distribution]].
* If the discrete distribution has two or more categories one of which may occur, whether or not these categories have a natural ordering, when there is only a single trial (draw) this is a categorical distribution.
* An example of a [[Joint probability distribution|multivariate discrete distribution]], and of its probability mass function, is provided by the [[multinomial distribution]]. Here the multiple random variables are the numbers of successes in each of the categories after a given number of trials, and each non-zero probability mass gives the probability of a certain combination of numbers of successes in the various categories.
{{clear}}
===Infinite===
The following exponentially declining distribution is an example of a distribution with an infinite number of possible outcomes—all the positive integers: <math display="block">\text{Pr}(X=i)= \frac{1}{2^i}\qquad \text{for } i=1, 2, 3, \dots </math> Despite the infinite number of possible outcomes, the total probability mass is 1/2 + 1/4 + 1/8 +
▲:Despite the infinite number of possible outcomes, the total probability mass is 1/2 + 1/4 + 1/8 + ... = 1, satisfying the unit total probability requirement for a probability distribution.
==Multivariate case==
Line 100 ⟶ 80:
{{Theory of probability distributions}}
{{Authority control}}
[[Category:Types of probability distributions]]
|