Content deleted Content added
mNo edit summary |
Citation bot (talk | contribs) Altered url. URLs might have been anonymized. Added bibcode. | Use this bot. Report bugs. | Suggested by Abductive | Category:Monte Carlo methods | #UCB_Category 9/65 |
||
(27 intermediate revisions by 17 users not shown) | |||
Line 1:
{{short description|Distribution estimation technique}}
== Basic theory ==
<!-- '''E'''[''X;P''] / \mathbf{E}[X;P] represents the expectation of ''X''. -->
Let <math>X\colon \Omega\to \mathbb{R}</math> be a [[random variable]] in some [[probability space]] <math>(\Omega,\mathcal{F},\mathbb{P})</math>. We wish to estimate the [[expected value]] of
: <math>
\widehat{\
</math>
and the precision of this estimate depends on the variance of
: <math>
\operatorname{var}_\mathbb{P}\big[\widehat{\
</math>
The basic idea of importance sampling is to sample
This is accomplished by first choosing a random variable <math>
With the
: <math>
\
</math>
The variable
<math>\operatorname{var}\left[\frac{X}{L};P^{(L)}\right] < \operatorname{var}[X;P]</math>.▼
When ''X'' is of constant sign over Ω, the best variable ''L'' would clearly be <math>L^*=\frac{X}{\mathbf{E}[X;P]}\geq 0</math>, so that ''X''/''L''* is the searched constant '''E'''[''X;P''] and a single sample under ''P''<sup>(''L''*)</sup> suffices to give its value. Unfortunately we cannot take that choice, because '''E'''[''X;P''] is precisely the value we are looking for! However this theoretical best case ''L*'' gives us an insight into what importance sampling does:▼
: <math>
▲
\begin{align}\forall a\in\mathbb{R}, \; P^{(L^*)}(X\in[a;a+da]) &= \int_{\omega\in\{X\in[a;a+da]\}} \frac{X(\omega)}{E[X;P]} \, dP(\omega) \\[6pt] &= \frac{1}{E[X;P]}\; a\,P(X\in[a;a+da]) ▼
to the right, <math>a\,P(X\in[a;a+da])</math> is one of the infinitesimal elements that sum up to '''E'''[''X'';''P'']:▼
▲When
: <math>\begin{align}
▲
\end{align}
</math>
▲
: <math>\mathbb{E}_{\mathbb{P}}[X
therefore, a good probability change
Importance sampling is often used as a [[Monte Carlo integration|Monte Carlo integrator]].
When <math>\mathbb{P}</math> is the uniform distribution
== Application to probabilistic inference ==
Such methods are frequently used to estimate posterior densities or expectations in state and/or parameter estimation problems in probabilistic models that are too hard to treat analytically
== Application to simulation ==
'''Importance sampling''' is a [[variance reduction]] technique that can be used in the [[Monte Carlo method]]. The idea behind importance sampling is that certain values of the input [[random variables]] in a [[simulation]] have more impact on the parameter being estimated than others. If these "[[important]]" values are emphasized by sampling more frequently, then the [[estimator]] variance can be reduced. Hence, the basic methodology in importance sampling is to choose a distribution which "encourages" the important values. This use of "biased" distributions will result in a biased estimator if it is applied directly in the simulation. However, the simulation outputs are weighted to correct for the use of the biased distribution, and this ensures that the new importance sampling estimator is unbiased. The weight is given by the [[Likelihood-ratio test|likelihood ratio]], that is, the [[Radon–Nikodym derivative]] of the true underlying distribution with respect to the biased simulation distribution.
The fundamental issue in implementing importance sampling simulation is the choice of the biased distribution which encourages the important regions of the input variables. Choosing or designing a good biased distribution is the "art" of importance sampling. The rewards for a good distribution can be huge run-time savings; the penalty for a bad distribution can be longer run times than for a general Monte Carlo simulation without importance sampling.
Line 64 ⟶ 66:
=== Mathematical approach ===
Consider estimating by simulation the probability <math>p_t\,</math> of an event <math>X \ge t</math>, where <math>X</math> is a random variable with [[
:<math>P(k_t = k)={K\choose k}p_t^k(1-p_t)^{K-k},\,\quad \quad k=0,1,\dots,K.</math>
One can show that <math>\
:<math>
\begin{align}
p_t & = \mathbb{E} [
& = \int
& =
\end{align}
</math>
Line 84 ⟶ 86:
is a likelihood ratio and is referred to as the weighting function. The last equality in the above equation motivates the estimator
:<math>
This is the importance sampling estimator of <math>p_t\,</math> and is unbiased. That is, the estimation procedure is to generate i.i.d. samples from <math>f_*\,</math> and for each sample which exceeds <math>t\,</math>, the estimate is incremented by the weight <math>W\,</math> evaluated at the sample value. The results are averaged over <math>K\,</math> trials. The variance of the importance sampling estimator is easily shown to be
Line 90 ⟶ 92:
:<math>
\begin{align}
\operatorname{var}_*\widehat p_t & = \frac{1}{K}\operatorname{var}_* [
& = \frac{1}{K}\left\{\mathbb{
& = \frac{1}{K}\left\{\mathbb{E}[
\end{align}
</math>
Line 137 ⟶ 139:
=== Evaluation of importance sampling ===
In order to identify successful importance sampling techniques, it is useful to be able to quantify the run-time savings due to the use of the importance sampling approach. The performance measure commonly used is <math>\sigma^2_{MC} / \sigma^2_{IS} \,</math>, and this can be interpreted as the speed-up factor by which the importance sampling estimator achieves the same precision as the MC estimator. This has to be computed empirically since the estimator variances are not likely to be analytically possible when their mean is intractable. Other useful concepts in quantifying an importance sampling estimator are the variance bounds and the notion of asymptotic efficiency. One related measure is the so-called '''Effective Sample Size''' '''(ESS)'''.<ref>{{Cite journal|
=== Variance cost function ===
Line 146 ⟶ 148:
===Multiple and adaptive importance sampling ===
When different proposal distributions, <math>
== See also ==
Line 163 ⟶ 165:
== References ==
*{{cite journal |last=Arouna |first=Bouhari |title=Adaptative Monte Carlo Method, A Variance Reduction Technique |journal=Monte Carlo Methods and Their Applications |year=2004 |volume=10 |issue=1 |pages=1–24 |doi=10.1515/156939604323091180 |s2cid=21949573 }}
*{{cite book |title=Introduction to Rare Event Simulation |first=James Antonio |last=Bucklew |publisher=Springer-Verlag |___location=New York |year=2004 }}
*{{cite book |title=Sequential Monte Carlo Methods in Practice |
*{{cite book |
*{{cite journal |first=Oleg |last=Mazonka |title=Easy as Pi: The Importance Sampling Method |journal=Journal of Reference |volume=16 |year=2016 |url=
*{{cite book |first=Tommy |last=Oberg |title=Modulation, Detection, and Coding |publisher=John Wiley & Sons |___location=New York |year=2001 }}
*{{Cite book | last1=Press | first1=WH | last2=Teukolsky | first2=SA | last3=Vetterling | first3=WT | last4=Flannery | first4=BP | year=2007 | title=Numerical Recipes: The Art of Scientific Computing | edition=3rd | publisher=Cambridge University Press |
*{{cite book |first=B. D. |last=Ripley |title=Stochastic Simulation |year=1987 |publisher=Wiley & Sons }}
*{{cite journal |
*{{cite book |first=R. |last=Srinivasan |title=Importance sampling – Applications in communications and detection |publisher=Springer-Verlag |___location=Berlin |year=2002 }}
Line 177 ⟶ 179:
* [http://www-sigproc.eng.cam.ac.uk/smc/ Sequential Monte Carlo Methods (Particle Filtering)] homepage on University of Cambridge
* [http://www.iop.org/EJ/abstract/0143-0807/22/4/315 Introduction to importance sampling in rare-event simulations] European journal of Physics. PDF document.
* [http://portal.acm.org/citation.cfm?id=1030470 Adaptive
[[Category:Monte Carlo methods]]
|