Content deleted Content added
Magioladitis (talk | contribs) moved long comment in talk page |
swap out deadlink |
||
(32 intermediate revisions by 25 users not shown) | |||
Line 1:
{{Short description|Mathematical method of risk analysis}}
{{Use dmy dates|date=October 2022}}
'''Probability bounds analysis''' ('''PBA''') is a collection of methods of uncertainty propagation for making qualitative and quantitative calculations in the face of uncertainties of various kinds. It is used to project partial information about random variables and other quantities through mathematical expressions. For instance, it computes sure bounds on the distribution of a sum, product, or more complex function, given only sure bounds on the distributions of the inputs. Such bounds are called [[probability box]]es, and constrain [[cumulative distribution function|cumulative probability distributions]] (rather than [[probability density function|densities]] or [[probability mass function|mass functions]]).
This [[upper and lower bounds|bounding]] approach permits analysts to make calculations without requiring overly precise assumptions about parameter values, dependence among variables, or even distribution shape. Probability bounds analysis is essentially a combination of the methods of standard [[interval analysis]] and classical [[probability theory]]. Probability bounds analysis gives the same answer as interval analysis does when only range information is available. It also gives the same answers as [[Monte Carlo simulation]] does when information is abundant enough to precisely specify input distributions and their dependencies. Thus, it is a generalization of both interval analysis and probability theory.
The diverse methods comprising probability bounds analysis provide algorithms to evaluate mathematical expressions when there is uncertainty about the input values, their dependencies, or even the form of mathematical expression itself. The calculations yield results that are guaranteed to enclose all possible distributions of the output variable if the input [[probability box|p-boxes]] were also sure to enclose their respective distributions. In some cases, a calculated p-box will also be best-possible in the sense that the bounds could be no tighter without excluding some of the possible distributions.
P-boxes are usually merely bounds on possible distributions. The bounds often also enclose distributions that are not themselves possible. For instance, the set of probability distributions that could result from adding random values without the independence assumption from two (precise) distributions is generally a proper [[subset]] of all the distributions enclosed by the p-box computed for the sum. That is, there are distributions within the output p-box that could not arise under any dependence between the two input distributions. The output p-box will, however, always contain all distributions that are possible, so long as the input p-boxes were sure to enclose their respective underlying distributions. This property often suffices for use in [[Probabilistic risk assessment|risk analysis]] and other fields requiring calculations under uncertainty.
==History of bounding probability==
The idea of bounding probability has a very long tradition throughout the history of probability theory. Indeed, in 1854 [[George Boole]] used the notion of interval bounds on probability in his ''[[The Laws of Thought]]''.<ref name="BOOLE1854">{{cite book|url= https://www.gutenberg.org/ebooks/15114 |last=Boole |first=George |title=An Investigation of the Laws of Thought on which are Founded the Mathematical Theories of Logic and Probabilities |publisher=Walton and Maberly |year=1854 |___location=London}}</ref><ref name=Hailperin86>{{cite book |last=Hailperin |first=Theodore |title=Boole's Logic and Probability |publisher=North-Holland |year=1986 |___location=Amsterdam |isbn=978-0-444-11037-4 }}</ref> Also dating from the latter half of the 19th century, the [[Chebyshev inequality|inequality]] attributed to [[Chebyshev]] described bounds on a distribution when only the mean and variance of the variable are known, and the related [[Markov inequality|inequality]] attributed to [[Andrey Markov|Markov]] found bounds on a positive variable when only the mean is known. [[Henry E. Kyburg, Jr.|Kyburg]]<ref name="kyburg99">Kyburg, H.E., Jr. (1999). [https://sipta.org/documentation/interval_prob/kyburgnew.pdf Interval valued probabilities]. SIPTA Documentation on Imprecise Probability.</ref> reviewed the history of interval probabilities and traced the development of the critical ideas through the 20th century, including the important notion of incomparable probabilities favored by [[John Maynard Keynes|Keynes]].
Of particular note is [[Maurice René Fréchet|Fréchet]]'s derivation in the 1930s of bounds on calculations involving total probabilities without dependence assumptions. Bounding probabilities has continued to the present day (e.g., Walley's theory of [[imprecise probability]].<ref name="WALLEY1991">{{cite book|url= https://archive.org/details/statisticalreaso0000wall |last=Walley |first=Peter |title=Statistical Reasoning with Imprecise Probabilities |url-access=registration |publisher=Chapman and Hall |year=1991 |___location=London |isbn=978-0-412-28660-5 }}</ref>)
The methods of probability bounds analysis that could be routinely used in
risk assessments were developed in the 1980s. Hailperin<ref name=Hailperin86 /> described a computational scheme for bounding logical calculations extending the ideas of Boole. Yager<ref name=Yager>Yager, R.R. (1986). Arithmetic and other operations on Dempster–Shafer structures. ''International Journal of Man-machine Studies'' '''25''': 357–366.</ref> described the elementary procedures by which bounds on [[convolution of probability distributions|convolutions]] can be computed under an assumption of independence. At about the same time, Makarov,
==Arithmetic expressions==
Arithmetic expressions involving operations such as additions, subtractions, multiplications, divisions, minima, maxima, powers, exponentials, logarithms, square roots, absolute values, etc., are commonly used in [[Probabilistic risk assessment|risk analyses]] and uncertainty modeling. Convolution is the operation of finding the probability distribution of a sum of independent random variables specified by probability distributions. We can extend the term to finding distributions of other mathematical functions (products, differences, quotients, and more complex functions) and other assumptions about the intervariable dependencies. There are convenient algorithms for computing these generalized convolutions under a variety of assumptions about the dependencies among the inputs.<ref name=Yager /><ref name=WilliamsonDowns /><ref name=Fersonetal03 /><ref name=Fersonetal04 />
===Mathematical details===
Let <math>\mathbb{D}</math> denote the space of distribution functions on the [[real number]]s <math>\R,</math> i.e.,
:<math> \mathbb{D} = \{ D | D: \R \to [0,1], D(x) \leq D(y) \text{ for all } x < y \}.</math>
A p-box is a quintuple
:<math>\left \{ \overline{F}, \underline{F}, m, v, \mathbf{F} \right \},</math>
where <math>\overline{F} </math> and <math>\underline{F} \in \mathbb{D}</math>, <math>m</math> and <math>v</math> are real intervals, and <math>\mathbf{F} \subset \mathbb{D}</math>. This quintuple denotes the set of distribution functions <math>F \in \mathbf{F} \subset \mathbb{D}</math> such that:
:<math>\begin{align}
\forall x \in \R: \qquad &\overline{F}(x) \leq F(x) \leq \underline{F}(x) \\[6pt]
&\int_\R x dF(x) \in m && \text{expectation condition} \\
&\int_\R x^2 dF(x) - \left ( \int_\R x dF(x) \right )^2 \in v && \text{variance condition}
\end{align}</math>
If a function satisfies all the conditions above it is said to be ''inside'' the p-box. In some cases, there may be no information about the moments or distribution family other than what is encoded in the two distribution functions that constitute the edges of the p-box. Then the quintuple representing the p-box <math>\{B_1, B_2, [-\infty, \infty], [0, \infty], \mathbb{D}\}</math> can be denoted more compactly as [''B''<sub>1</sub>, ''B''<sub>2</sub>]. This notation harkens to that of intervals on the real line, except that the endpoints are distributions rather than points.
The notation <math>X \sim F</math> denotes the fact that <math>X \in \R</math> is a random variable governed by the distribution function ''F'', that is,
:<math>\begin{cases} F: \R \to [0,1] \\ x \mapsto \Pr (X \leq x) \end{cases}</math>
Let us generalize the tilde notation for use with p-boxes. We will write ''X'' ~ ''B'' to mean that ''X'' is a random variable whose distribution function is unknown except that it is inside ''B''. Thus, ''X'' ~ ''F'' ∈ ''B'' can be contracted to X ~ B without mentioning the distribution function explicitly.
If ''X'' and ''Y'' are independent random variables with distributions ''F'' and ''G'' respectively, then ''X'' + ''Y'' = ''Z'' ~ ''H'' given by
:<math>H(z) = \int_{z=x+y} F(x) G(y) dz = \int_{\R} F(x) G(z-x) dx = F * G.</math>
This operation is called a [[convolution]] on ''F'' and ''G''. The analogous operation on p-boxes is straightforward for sums. Suppose
:<math>X \sim A = [A_1, A_2], \quad \text{and} \quad Y \sim B = [B_1, B_2].</math>
:<math> \left [A_1 * B_1, A_2 * B_2 \right ].</math>
Finding bounds on the distribution of sums ''Z'' = ''X'' + ''Y'' ''without making any assumption about the dependence'' between ''X'' and ''Y'' is actually easier than the problem assuming independence. Makarov<ref name=Makarov/><ref name=Franketal87/><ref name=WilliamsonDowns/> showed that
:<math>Z \sim \left [ \sup_{z=x+y} \max ( F(x) +G(y) -1, 0), \inf_{z=x+y} \min (F(x)+G(y), 1) \right ]</math>
These bounds are implied by the [[copula (probability theory)#Fr.C3.A9chet.E2.80.93Hoeffding copula bounds|Fréchet–Hoeffding]] [[copula (probability theory)|copula]] bounds. The problem can also be solved using the methods of [[mathematical programming]].<ref name=BerleantGoodmanStrauss />
The convolution under the intermediate assumption that ''X'' and ''Y'' have [[positive quadrant dependence|positive dependence]] is likewise easy to compute, as is the convolution under the extreme assumptions of [[Comonotonicity|perfect positive]] or [[countermonotonicity|perfect negative]] dependency between ''X'' and ''Y''.<ref name=Fersonetal04 />
Generalized convolutions for other operations such as subtraction, multiplication, division, etc., can be derived using transformations. For instance, p-box subtraction ''A'' − ''B'' can be defined as ''A'' + (−''B''), where the negative of a p-box ''B'' = [''B''<sub>1</sub>, ''B''<sub>2</sub>] is [''B''<sub>2</sub>(−''x''), ''B''<sub>1</sub>(−''x'')].
==Logical expressions==
Logical or [[
: P(A & B) = ''a''
:::: = [0.2, 0.25]
:::: = [0.2
:::: = [0.02, 0.075]
so long as A and B can be assumed to be independent events. If they are not independent, we can still bound the conjunction using the classical [[Frechet inequalities|
: P(A & B) = env(max(0, ''a''+''b''−1), min(''a'', ''b''))
:::: = env(max(0, [0.2, 0.25]+[0.1, 0.3]−1), min([0.2, 0.25], [0.1, 0.3]))
Line 125 ⟶ 80:
:::: = [0, 0.25]
where env([''x''<sub>1</sub>,''x''<sub>2</sub>], [''y''<sub>1</sub>,''y''<sub>2</sub>]) is [min(''x''<sub>1</sub>,''y''<sub>1</sub>), max(''x''<sub>2</sub>,''y''<sub>2</sub>)]. Likewise, the probability of the [[logical disjunction|disjunction]] is surely in the interval
: P(A v B) = ''a'' + ''b'' − ''a''
:::: = 1 − (1 − [0.2, 0.25])
:::: = 1 − [0.75, 0.8]
:::: = 1 − [0.525, 0.72]
:::: = [0.28, 0.475]
if A and B are independent events. If they are not independent, the
: P(A v B) = env(max(''a'', ''b''), min(1, ''a'' + ''b''))
:::: = env(max([0.2, 0.25], [0.1, 0.3]), min(1, [0.2, 0.25] + [0.1, 0.3]))
Line 136 ⟶ 91:
:::: = [0.2, 0.55].
It is also possible to compute interval bounds on the conjunction or disjunction under other assumptions about the dependence between A and B. For instance, one might assume they are positively dependent, in which case the resulting interval is not as tight as the answer assuming independence but tighter than the answer given by the
==Magnitude comparisons==
The probability that an uncertain number represented by a p-box ''D'' is less than zero is the interval Pr(''D'' < 0) = [<u>''F''</u>
:''A'' < ''B'' = Pr(''A'' − ''B'' < 0),
:''A'' > ''B'' = Pr(''B'' − ''A'' < 0),
:''A''
:''A''
Thus the probability that ''A'' is less than ''B'' is the same as the probability that their difference is less than zero, and this probability can be said to be the value of the expression ''A'' < ''B''.
Line 173 ⟶ 104:
==Sampling-based computation==
Some analysts<ref>Alvarez, D. A., 2006. On the calculation of the bounds of probability of events using infinite random sets. ''International Journal of Approximate Reasoning'' '''43''': 241–267.</ref><ref>Baraldi, P., Popescu, I. C., Zio, E., 2008. Predicting the time to failure of a randomly degrading component by a hybrid Monte Carlo and possibilistic method. ''IEEE Proc. International Conference on Prognostics and Health Management''.</ref><ref>Batarseh, O. G., Wang, Y., 2008. Reliable simulation with input uncertainties using an interval-based approach. ''IEEE Proc. Winter Simulation Conference''.</ref><ref>Roy, Christopher J., and Michael S. Balch (2012). A holistic approach to uncertainty quantification with application to supersonic nozzle thrust. ''International Journal for Uncertainty Quantification''
==Relationship to other uncertainty propagation approaches==
PBA belongs to a class of methods that use [[imprecise probability|imprecise probabilities]] to simultaneously represent [[
==Applications==
Line 213 ⟶ 144:
==References==
{{Reflist|30em}}
==Further references==
* {{cite book | last1 = Bernardini | first1 = Alberto | last2 = Tonon | first2 = Fulvio | title = Bounding Uncertainty in Civil Engineering: Theoretical Background | publisher = Springer | ___location = Berlin | year = 2010 | isbn = 978-3-642-11189-1 }}
* {{cite book | last = Ferson | first = Scott | title = RAMAS Risk Calc 4.0 Software : Risk Assessment with Uncertain Numbers | publisher = Lewis Publishers | ___location = Boca Raton, Florida | year = 2002 | isbn = 978-1-56670-576-9 }}
* {{cite journal |first=G. |last=Gerla |title=Inferences in Probability Logic |journal=Artificial Intelligence |volume=70 |issue=1–2 |pages=33–52 |year=1994 |doi=10.1016/0004-3702(94)90102-3 }}
* {{cite book | last1 = Oberkampf | first1 = William L. | last2 = Roy | first2 = Christopher J. | title = Verification and Validation in Scientific Computing | publisher = Cambridge University Press | ___location = New York | year = 2010 | isbn = 978-0-521-11360-1 }}<!-- In an email dated 28 March 2011, William Oberkampf stated "PBA is the only UQ method we discuss and apply in our examples in the book." -->
==External links==
* [http://www.ramas.com/pbawhite.pdf Probability bounds analysis in environmental risk assessments]
* [http://ualr.edu/jdberleant/intprob/ Intervals and probability distributions]
* [https://web.archive.org/web/20120210155925/http://www.sandia.gov/epistemic/ Epistemic uncertainty project]
* [http://www.sipta.org/ The Society for Imprecise Probability: Theories and Applications]
[[Category:Probability bounds analysis| ]]
[[Category:Mathematical analysis]]
|