# Variational Autoencoders

Like autoencoders, but penalize latent variables for deviating from a Gaussian distribution.

$p(x|z)$ can be calculated using our decoder neural network, which generates images by taking samples from the latent space $x$.
$p(z)$ is our prior over the latent space $z$ (usually a unit Gaussian).
$p(x) = \int {p(z) p(x) dz}$ is difficult to calculate because it requires calculating $p(x)$ for all possible values of $z$.
$q(z|x)$ is the variational posterior, parameterized by a second set of parameters $\lambda$, which approximates the true posterior distribution $p(z|x)$.