Content deleted Content added
cdf is cumulative distribution function, not a density. Also linked a proper page on wiki to cdf. |
Citation bot (talk | contribs) Altered url. URLs might have been anonymized. Added bibcode. | Use this bot. Report bugs. | Suggested by Abductive | Category:Monte Carlo methods | #UCB_Category 9/65 |
||
(10 intermediate revisions by 8 users not shown) | |||
Line 1:
{{short description|Distribution estimation technique}}
'''Importance sampling''' is a [[Monte Carlo method]] for evaluating properties of a particular [[probability distribution|distribution]], while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally attributed to a paper by [[Teun Kloek]] and [[Herman K. van Dijk]] in 1978,<ref>{{cite journal |first1=T. |last1=Kloek |first2=H. K. |last2=van Dijk |title=Bayesian Estimates of Equation System Parameters: An Application of Integration by Monte Carlo |journal=[[Econometrica]] |volume=46 |issue=1 |year=1978 |pages=1–19 |doi=10.2307/1913641 |jstor=1913641 |url=https://ageconsearch.umn.edu/record/272139/files/erasmus076.pdf }}</ref> but its precursors can be found in [[Monte Carlo method in statistical physics|statistical physics]] as early as 1949.<ref>{{cite journal |first=G. |last=Goertzle |authorlink=Gerald Goertzel |title=Quota Sampling and Importance Functions in Stochastic Solution of Particle Problems |journal=Technical Report ORNL-434, Oak Ridge National Laboratory |series=Aecd
== Basic theory ==
<!-- '''E'''[''X;P''] / \mathbf{E}[X;P] represents the expectation of ''X''. -->
Let <math>X\colon \Omega\to \mathbb{R}</math> be a [[random variable]] in some [[probability space]] <math>(\Omega,\mathcal{F},\mathbb{P})</math>. We wish to estimate the [[expected value]] of
: <math>
\widehat{\
</math>
and the precision of this estimate depends on the variance of
: <math>
\operatorname{var}_\mathbb{P}\big[\widehat{\
</math>
The basic idea of importance sampling is to sample
This is accomplished by first choosing a random variable <math>
With the variable
: <math>
\
</math>
The variable
<math>\operatorname{var}\left[\frac{X}{L};P^{(L)}\right] < \operatorname{var}[X;P]</math>.▼
When ''X'' is of constant sign over Ω, the best variable ''L'' would clearly be <math>L^*=\frac{X}{\mathbf{E}[X;P]}\geq 0</math>, so that ''X''/''L''* is the searched constant '''E'''[''X;P''] and a single sample under ''P''<sup>(''L''*)</sup> suffices to give its value. Unfortunately we cannot take that choice, because '''E'''[''X;P''] is precisely the value we are looking for! However this theoretical best case ''L*'' gives us an insight into what importance sampling does:▼
: <math>
▲
\begin{align}\forall a\in\mathbb{R}, \; P^{(L^*)}(X\in[a;a+da]) &= \int_{\omega\in\{X\in[a;a+da]\}} \frac{X(\omega)}{E[X;P]} \, dP(\omega) \\[6pt] &= \frac{1}{E[X;P]}\; a\,P(X\in[a;a+da]) ▼
to the right, <math>a\,P(X\in[a;a+da])</math> is one of the infinitesimal elements that sum up to '''E'''[''X'';''P'']:▼
▲When
: <math>\begin{align}
▲
\end{align}
</math>
▲
: <math>\mathbb{E}_{\mathbb{P}}[X
therefore, a good probability change
Importance sampling is often used as a [[Monte Carlo integration|Monte Carlo integrator]].
When <math>\mathbb{P}</math> is the uniform distribution
== Application to probabilistic inference ==
Such methods are frequently used to estimate posterior densities or expectations in state and/or parameter estimation problems in probabilistic models that are too hard to treat analytically. Examples include [[Bayesian network]]s and importance weighted [[
== Application to simulation ==
Line 67 ⟶ 70:
:<math>P(k_t = k)={K\choose k}p_t^k(1-p_t)^{K-k},\,\quad \quad k=0,1,\dots,K.</math>
One can show that <math>\
:<math>
\begin{align}
p_t & = \mathbb{E} [
& = \int
& =
\end{align}
</math>
Line 83 ⟶ 86:
is a likelihood ratio and is referred to as the weighting function. The last equality in the above equation motivates the estimator
:<math>
This is the importance sampling estimator of <math>p_t\,</math> and is unbiased. That is, the estimation procedure is to generate i.i.d. samples from <math>f_*\,</math> and for each sample which exceeds <math>t\,</math>, the estimate is incremented by the weight <math>W\,</math> evaluated at the sample value. The results are averaged over <math>K\,</math> trials. The variance of the importance sampling estimator is easily shown to be
Line 89 ⟶ 92:
:<math>
\begin{align}
\operatorname{var}_*\widehat p_t & = \frac{1}{K}\operatorname{var}_* [
& = \frac{1}{K}\left\{\mathbb{
& = \frac{1}{K}\left\{\mathbb{E}[
\end{align}
</math>
Line 136 ⟶ 139:
=== Evaluation of importance sampling ===
In order to identify successful importance sampling techniques, it is useful to be able to quantify the run-time savings due to the use of the importance sampling approach. The performance measure commonly used is <math>\sigma^2_{MC} / \sigma^2_{IS} \,</math>, and this can be interpreted as the speed-up factor by which the importance sampling estimator achieves the same precision as the MC estimator. This has to be computed empirically since the estimator variances are not likely to be analytically possible when their mean is intractable. Other useful concepts in quantifying an importance sampling estimator are the variance bounds and the notion of asymptotic efficiency. One related measure is the so-called '''Effective Sample Size''' '''(ESS)'''.<ref>{{Cite journal|last1=Martino|first1=Luca|last2=Elvira|first2=Víctor|last3=Louzada|first3=Francisco|title=Effective sample size for importance sampling based on discrepancy measures|journal=Signal Processing|volume=131|pages=386–401|doi=10.1016/j.sigpro.2016.08.025|arxiv=1602.03572|year=2017|bibcode=2017SigPr.131..386M |s2cid=26317735 }}</ref>
=== Variance cost function ===
Line 145 ⟶ 148:
===Multiple and adaptive importance sampling ===
When different proposal distributions, <math>
== See also ==
Line 166 ⟶ 169:
*{{cite book |title=Sequential Monte Carlo Methods in Practice |first1=A. |last1=Doucet |first2=N. |last2=de Freitas |first3=N. |last3=Gordon |publisher=Springer |year=2001 |isbn=978-0-387-95146-1 }}
*{{cite book |first1=M. |last1=Ferrari |first2=S. |last2=Bellini |title=ICC 2001. IEEE International Conference on Communications. Conference Record (Cat. No.01CH37240) |chapter=Importance sampling simulation of turbo product codes |volume=9 |pages=2773–2777 |year=2001 |doi=10.1109/ICC.2001.936655 |isbn=978-0-7803-7097-5 |s2cid=5158473 }}
*{{cite journal |first=Oleg |last=Mazonka |title=Easy as Pi: The Importance Sampling Method |journal=Journal of Reference |volume=16 |year=2016 |url=
*{{cite book |first=Tommy |last=Oberg |title=Modulation, Detection, and Coding |publisher=John Wiley & Sons |___location=New York |year=2001 }}
*{{Cite book | last1=Press | first1=WH | last2=Teukolsky | first2=SA | last3=Vetterling | first3=WT | last4=Flannery | first4=BP | year=2007 | title=Numerical Recipes: The Art of Scientific Computing | edition=3rd | publisher=Cambridge University Press |
*{{cite book |first=B. D. |last=Ripley |title=Stochastic Simulation |year=1987 |publisher=Wiley & Sons }}
*{{cite journal |first1=P. J. |last1=Smith |first2=M. |last2=Shafi |first3=H. |last3=Gao |title=Quick simulation: A review of importance sampling techniques in communication systems |journal= IEEE Journal on Selected Areas in Communications |volume=15 |issue=4 |pages=597–613 |year=1997 |doi=10.1109/49.585771 }}
|