Content deleted Content added
MishchenkoA (talk | contribs) mNo edit summary |
MishchenkoA (talk | contribs) →Formulation: Tighten up wording, make notation and terminology more consistent with the variational Bayesian methods page. |
||
Line 12:
== Formulation ==
[[File:VAE Basic.png|thumb|425x425px|The basic scheme of a variational autoencoder. The model receives <math>\mathbf{x}</math> as input. The encoder compresses it into the latent space. The decoder receives as input the information sampled from the latent space and produces <math>\mathbf{x'}</math> as similar as possible to <math>\mathbf{x}</math>.]]
From a formal perspective, given an input dataset <math>\mathbf{x}</math> characterized by an unknown probability
: <math>p_\theta(\mathbf{x}) = \int_{\mathbf{z}}p_\theta(\mathbf{x,z}) \, d\mathbf{z}, </math>▼
where <math>p_\theta(\mathbf{x,z})</math> represents the [[joint distribution]] under <math>p_\theta</math> of the observable data <math>\mathbf x </math> and its latent representation or encoding <math>\mathbf z </math>. According to the [[Chain rule (probability)|chain rule]], the equation can be rewritten as
▲: <math>p_\theta(\mathbf{x}) = \int_{\mathbf{z}}p_\theta(\mathbf{x,z}) \, d\mathbf{z} </math>
: <math>p_\theta(\mathbf{x}) = \int_{\mathbf{z}}p_\theta(\mathbf{x\mid z})p_\theta(\mathbf{z}) \, d\mathbf{z}</math>
In the vanilla variational autoencoder, we
It is now possible to define the set of the relationships between the input data and its latent representation as
|