Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Variational Autoencoder (VAE)

Variational Autoencoder (VAE)

A generative model that learns a structured latent space via probabilistic encoding and decoding.

Year: 2013Generality: 720
Back to Vocab

A Variational Autoencoder (VAE) is a class of generative model that combines deep neural networks with principles from Bayesian inference to learn compact, structured representations of data. Unlike standard autoencoders, which map inputs to fixed points in a latent space, VAEs map inputs to probability distributions — typically Gaussians — over that space. During training, the encoder network outputs the parameters of these distributions (mean and variance), a sample is drawn from them, and the decoder network attempts to reconstruct the original input from that sample. This stochastic bottleneck forces the model to learn smooth, continuous latent representations rather than memorizing individual data points.

The training objective of a VAE is the Evidence Lower Bound (ELBO), which balances two competing terms. The first is a reconstruction loss that penalizes the model when its output diverges from the input. The second is a Kullback-Leibler (KL) divergence term that regularizes the learned distributions to remain close to a standard normal prior. This regularization is what gives the latent space its generative utility: because the space is structured and continuous, interpolating between points or sampling randomly tends to produce coherent, meaningful outputs rather than noise or artifacts.

VAEs matter because they offer a principled, mathematically grounded framework for generative modeling and representation learning. They enable tasks such as image synthesis, data augmentation, anomaly detection, and disentangled representation learning — where individual latent dimensions correspond to interpretable factors of variation in the data. Compared to Generative Adversarial Networks (GANs), VAEs are generally more stable to train and provide an explicit likelihood estimate, though they can produce blurrier outputs due to the averaging effect of the reconstruction loss.

Introduced by Diederik Kingma and Max Welling in their 2013 paper "Auto-Encoding Variational Bayes," VAEs quickly became a foundational tool in deep generative modeling. They have since been extended in numerous directions — including conditional VAEs, hierarchical VAEs, and vector-quantized variants — and remain central to research in representation learning, multimodal generation, and latent diffusion models.

Related

Related

Autoencoder
Autoencoder

A neural network that compresses data into a compact representation, then reconstructs it.

Generality: 795
Denoising Autoencoder
Denoising Autoencoder

A neural network that learns robust representations by reconstructing clean data from corrupted inputs.

Generality: 694
Spatial Autoencoder
Spatial Autoencoder

An autoencoder variant that learns compact representations by preserving spatial structure in data.

Generality: 391
Variational Free Energy
Variational Free Energy

A bound on model evidence used to approximate intractable posterior distributions efficiently.

Generality: 650
Generative Model
Generative Model

A model that learns data distributions to synthesize realistic new samples.

Generality: 896
Latent Space
Latent Space

A compressed, learned representation where similar data points cluster geometrically.

Generality: 794