Content deleted Content added
m The use of VAE mentioned was only partially correct - it can be used for compressing information from input space to latent space but its main goal is to map the input space to a latent distribution. This property of VAE differentiates it from autoencoders which performs only compression. |
m v2.04b - Bot T20 CW#61 - Fix errors for CW project (Reference before punctuation) |
||
Line 5:
In [[machine learning]], a '''variational autoencoder''',<ref name=":0">{{cite arXiv |last1=Kingma |first1=Diederik P. |last2=Welling |first2=Max |title=Auto-Encoding Variational Bayes |date=2014-05-01 |class=stat.ML |eprint=1312.6114}}</ref> also known as a '''VAE''', is the [[artificial neural network]] architecture introduced by [[Diederik P. Kingma]] and [[Max Welling]], belonging to the families of [[graphical model|probabilistic graphical models]] and [[variational Bayesian methods]].
It is often associated with the [[autoencoder]]<ref>{{cite journal |last1=Kramer |first1=Mark A. |title=Nonlinear principal component analysis using autoassociative neural networks |journal=AIChE Journal |date=1991 |volume=37 |issue=2 |pages=233–243 |doi=10.1002/aic.690370209 |url=https://aiche.onlinelibrary.wiley.com/doi/abs/10.1002/aic.690370209 |language=en}}</ref><ref>{{cite journal |last1=Hinton |first1=G. E. |last2=Salakhutdinov |first2=R. R. |title=Reducing the Dimensionality of Data with Neural Networks |journal=Science |date=2006-07-28 |volume=313 |issue=5786 |pages=504–507 |doi=10.1126/science.1127647 |pmid=16873662 |bibcode=2006Sci...313..504H |s2cid=1658773 |url=https://science.sciencemag.org/content/313/5786/504.abstract?casa_token=ZLsQ9vPfFA4AAAAA:3iBJRtRFr9RzkbbGpAJQtghIAndmRGEPVxW-yixDgfiXqWuuaQs8WjDMf-fkzTIe8RKn_J9o1aFozD4 |language=en}}</ref> model because of its architectural affinity, but there are significant differences both in the goal and in the mathematical formulation. Variational autoencoders allows us to re-write statistical inference problems (i.e infer the value of one random variable from another random variable) as statistical optimization problems (i.e find the parameter values that minimize some objective function).<ref>{{cite web |title=A Beginner's Guide to Variational Methods: Mean-Field Approximation |url=https://blog.evjang.com/2016/08/variational-bayes.html |website=Eric Jang |language=en |date=2016-07-08}}</ref>
== Architecture ==
|