Content deleted Content added
→References: |doi-access=free |
ce |
||
Line 17:
Although Diggle and Gratton's approach had opened a new frontier, their method was not yet exactly identical to what is now known as ABC, as it aimed at approximating the likelihood rather than the posterior distribution. An article of [[Simon Tavaré]] and co-authors was first to propose an ABC algorithm for posterior inference.<ref name="Tavare" /> In their seminal work, inference about the genealogy of DNA sequence data was considered, and in particular the problem of deciding the posterior distribution of the time to the [[most recent common ancestor]] of the sampled individuals. Such inference is analytically intractable for many demographic models, but the authors presented ways of simulating coalescent trees under the putative models. A sample from the posterior of model parameters was obtained by accepting/rejecting proposals based on comparing the number of segregating sites in the synthetic and real data. This work was followed by an applied study on modeling the variation in human Y chromosome by [[Jonathan K. Pritchard]] and co-authors using the ABC method.<ref name="Pritchard1999" /> Finally, the term approximate Bayesian computation was established by Mark Beaumont and co-authors,<ref name="Beaumont2002" /> extending further the ABC methodology and discussing the suitability of the ABC-approach more specifically for problems in population genetics. Since then, ABC has spread to applications outside population genetics, such as systems biology, epidemiology, and [[phylogeography]].
Approximate Bayesian computation can be understood as a kind of Bayesian version of [[indirect inference]].<ref>
Several efficient Monte Carlo based approaches have been developed to perform sampling from the ABC posterior distribution for purposes of estimation and prediction problems. A popular choice is the SMC Samplers algorithim <ref>{{Cite journal |last1=Del Moral |first1=Pierre |last2=Doucet |first2=Arnaud |last3=Jasra |first3=Ajay |date=2006 |title=Sequential Monte Carlo Samplers |url=https://www.jstor.org/stable/3879283 |journal=Journal of the Royal Statistical Society. Series B (Statistical Methodology) |volume=68 |issue=3 |pages=411–436 |doi=10.1111/j.1467-9868.2006.00553.x |jstor=3879283 |issn=1369-7412|arxiv=cond-mat/0212648 }}</ref><ref>{{Cite journal |last1=Del Moral |first1=Pierre |last2=Doucet |first2=Arnaud |last3=Peters |first3=Gareth |date=2004 |title=Sequential Monte Carlo Samplers CUED Technical Report |url=https://www.ssrn.com/abstract=3841065 |journal=SSRN Electronic Journal |language=en |doi=10.2139/ssrn.3841065 |issn=1556-5068}}</ref><ref>{{Cite journal |last=Peters |first=Gareth |date=2005 |title=Topics in Sequential Monte Carlo Samplers |url=https://www.ssrn.com/abstract=3785582 |journal=SSRN Electronic Journal |language=en |doi=10.2139/ssrn.3785582 |issn=1556-5068}}</ref> adapted to the ABC context in the method (SMC-ABC).<ref>{{Cite journal |last1=Sisson |first1=S. A. |last2=Fan |first2=Y. |last3=Tanaka |first3=Mark M. |date=2007-02-06 |title=Sequential Monte Carlo without likelihoods |journal=Proceedings of the National Academy of Sciences |language=en |volume=104 |issue=6 |pages=1760–1765 |doi=10.1073/pnas.0607208104 |doi-access=free |issn=0027-8424 |pmc=1794282 |pmid=17264216|bibcode=2007PNAS..104.1760S }}</ref><ref>{{Cite journal |last=Peters |first=Gareth |date=2009 |title=Advances in Approximate Bayesian Computation and Trans-Dimensional Sampling Methodology |url=https://www.ssrn.com/abstract=3785580 |journal=SSRN Electronic Journal |language=en |doi=10.2139/ssrn.3785580 |issn=1556-5068}}</ref><ref>{{Cite journal |last1=Peters |first1=G. W. |last2=Sisson |first2=S. A. |last3=Fan |first3=Y. |date=2012-11-01 |title=Likelihood-free Bayesian inference for α-stable models |url=https://www.sciencedirect.com/science/article/pii/S0167947310003786 |journal=Computational Statistics & Data Analysis |series=1st issue of the Annals of Computational and Financial Econometrics |volume=56 |issue=11 |pages=3743–3756 |doi=10.1016/j.csda.2010.10.004 |issn=0167-9473}}</ref><ref>{{Cite journal |last1=Peters |first1=Gareth W. |last2=Wüthrich |first2=Mario V. |last3=Shevchenko |first3=Pavel V. |date=2010-08-01 |title=Chain ladder method: Bayesian bootstrap versus classical bootstrap |url=https://www.sciencedirect.com/science/article/pii/S0167668710000351 |journal=Insurance: Mathematics and Economics |volume=47 |issue=1 |pages=36–51 |doi=10.1016/j.insmatheco.2010.03.007 |arxiv=1004.2548 |issn=0167-6687}}</ref>
Line 189:
===Choice and sufficiency of summary statistics===
Summary statistics may be used to increase the acceptance rate of ABC for high-dimensional data. Low-dimensional sufficient statistics are optimal for this purpose, as they capture all relevant information present in the data in the simplest possible form.<ref name="Csillery" /><ref>{{Cite journal |last1=Peters |first1=Gareth William |last2=Wuthrich |first2=Mario V. |last3=Shevchenko |first3=Pavel V. |date=2009 |title=Chain Ladder Method: Bayesian Bootstrap Versus Classical Bootstrap |url=https://dx.doi.org/10.2139/ssrn.2980411 |journal=SSRN Electronic Journal |doi=10.2139/ssrn.2980411 |arxiv=1004.2548 |issn=1556-5068}}</ref><ref>{{
One approach to capture most of the information present in data would be to use many statistics, but the accuracy and stability of ABC appears to decrease rapidly with an increasing numbers of summary statistics.<ref name="Beaumont2010" /><ref name="Csillery" /> Instead, a better strategy is to focus on the relevant statistics only—relevancy depending on the whole inference problem, on the model used, and on the data at hand.<ref name="Nunes" />
Line 381:
<ref name="Gerstner">{{cite journal | last1 = Gerstner | first1 = T | last2 = Griebel | first2 = M | year = 2003 | title = Dimension-Adaptive Tensor-Product Quadrature | journal = Computing | volume = 71 | pages = 65–87 | doi=10.1007/s00607-003-0015-5| citeseerx = 10.1.1.16.2434 | s2cid = 16184111 }}</ref>
<ref name="Singer">{{cite journal | last1 = Singer | first1 = AB | last2 = Taylor | first2 = JW | last3 = Barton | first3 = PI | last4 = Green | first4 = WH | year = 2006 | title = Global dynamic optimization for parameter estimation in chemical kinetics | journal = J Phys Chem A | volume = 110 | issue = 3| pages = 971–976 | doi=10.1021/jp0548873| pmid = 16419997 | bibcode = 2006JPCA..110..971S }}</ref>
<ref name="Dean">{{arxiv|1103.5399}}</ref>
<ref name="Fearnhead">{{arxiv|1004.1112}}</ref>
<ref name="Wilkinson">{{arxiv|0811.3355}}</ref>
<ref name="Nunes">{{cite journal | last1 = Nunes | first1 = MA | last2 = Balding | first2 = DJ | year = 2010 | title = On optimal selection of summary statistics for approximate Bayesian computation | journal = Stat Appl Genet Mol Biol | volume = 9 | page = Article 34 | doi=10.2202/1544-6115.1576| pmid = 20887273 | s2cid = 207319754 }}</ref>
<ref name="Joyce">{{cite journal | last1 = Joyce | first1 = P | last2 = Marjoram | first2 = P | year = 2008 | title = Approximately sufficient statistics and bayesian computation | journal = Stat Appl Genet Mol Biol | volume = 7 | issue = 1| page = Article 26 | doi=10.2202/1544-6115.1389| pmid = 18764775 | s2cid = 38232110 }}</ref>
<ref name="Grelaud">{{cite journal | last1 = Grelaud | first1 = A | last2 = Marin | first2 = J-M | last3 = Robert | first3 = C | last4 = Rodolphe | first4 = F | last5 = Tally | first5 = F | year = 2009 | title = Likelihood-free methods for model choice in Gibbs random fields | journal = Bayesian Analysis | volume = 3 | pages = 427–442 }}</ref>
<ref name="Marin">{{arxiv|1110.4700}}</ref>
<ref name="Toni">{{cite journal | last1 = Toni | first1 = T | last2 = Welch | first2 = D | last3 = Strelkowa | first3 = N | last4 = Ipsen | first4 = A | last5 = Stumpf | first5 = M | year = 2007 | title = Approximate Bayesian computation scheme for parameter inference and model selection in dynamical systems | journal = J R Soc Interface | volume = 6 | issue = 31| pages = 187–202 | pmid = 19205079 | pmc = 2658655 | doi = 10.1098/rsif.2008.0172 }}</ref>
<ref name="Tavare">{{cite journal | last1 = Tavaré | first1 = S | last2 = Balding | first2 = DJ | last3 = Griffiths | first3 = RC | last4 = Donnelly | first4 = P | year = 1997 | title = Inferring Coalescence Times from DNA Sequence Data | journal = Genetics | volume = 145 | issue = 2 | pages = 505–518 | doi = 10.1093/genetics/145.2.505 | pmc = 1207814 | pmid=9071603}}</ref>
<ref name="Toni2010">{{doi|10.1093/bioinformatics/btp619}}</ref>
.<ref name="Pritchard1999">{{cite journal | last1 = Pritchard | first1 = JK | last2 = Seielstad | first2 = MT | last3 = Perez-Lezaun | first3 = A |display-authors=et al | year = 1999 | title = Population Growth of Human Y Chromosomes: A Study of Y Chromosome Microsatellites | journal = Molecular Biology and Evolution | volume = 16 | issue = 12| pages = 1791–1798 | doi=10.1093/oxfordjournals.molbev.a026091| pmid = 10605120 | doi-access = free }}</ref>
<ref name="Diggle">{{cite journal | last1 = Diggle | first1 = PJ | year = 1984 | title = Monte Carlo Methods of Inference for Implicit Statistical Models | journal = Journal of the Royal Statistical Society, Series B | volume = 46 | issue = 2 | pages = 193–227 | doi = 10.1111/j.2517-6161.1984.tb01290.x }}</ref>
Line 396:
<ref name="Lai">{{cite journal | last1 = Lai | first1 = K | last2 = Robertson | first2 = MJ | last3 = Schaffer | first3 = DV | year = 2004 | title = The sonic hedgehog signaling system as a bistable genetic switch | journal = Biophys. J. | volume = 86 | issue = 5| pages = 2748–2757 | doi=10.1016/s0006-3495(04)74328-3 | pmid = 15111393 | bibcode=2004BpJ....86.2748L | pmc=1304145}}</ref>
<ref name="Bartlett63">{{cite journal | last1 = Bartlett | first1 = MS | year = 1963 | title = The spectral analysis of point processes | journal = Journal of the Royal Statistical Society, Series B | volume = 25 | issue = 2 | pages = 264–296 | doi = 10.1111/j.2517-6161.1963.tb00508.x }}</ref>
<ref name="Blum12">
<ref name="Fearnhead12">{{cite journal | last1 = Fearnhead | first1 = P | last2 = Prangle | first2 = D | year = 2012 | title = Constructing summary statistics for approximate Bayesian computation: semi-automatic approximate Bayesian computation | journal = Journal of the Royal Statistical Society, Series B | volume = 74 | issue = 3| pages = 419–474 | doi=10.1111/j.1467-9868.2011.01010.x| citeseerx = 10.1.1.760.7753 | s2cid = 53861241 }}</ref>
<ref name="Blum10">Blum MGB (2010) Approximate Bayesian Computation: a nonparametric perspective, ''Journal of the American Statistical Association'' (105): 1178-1187</ref>
Line 413:
<ref name="Kangas16">{{cite journal |last1= Kangasrääsiö |first1= Antti |last2= Lintusaari |first2= Jarno |last3= Skytén |first3= Kusti |last4= Järvenpää |first4= Marko |last5= Vuollekoski |first5= Henri |last6= Gutmann |first6= Michael |last7= Vehtari |first7= Aki |last8= Corander |first8= Jukka |last9= Kaski |first9= Samuel|year= 2016 |title= ELFI: Engine for Likelihood-Free Inference |url=http://approximateinference.org/accepted/KangasraasioEtAl2016.pdf |journal= NIPS 2016 Workshop on Advances in Approximate Bayesian Inference|bibcode= 2017arXiv170800707L |arxiv= 1708.00707 }}</ref>
<ref name="Klinger2017">Klinger, E.; Rickert, D.; Hasenauer, J. (2017). pyABC: distributed, likelihood-free inference.</ref>
<ref name="Salvatier2016">
}}
|