A model that learns data distributions to synthesize realistic new samples.
A generative model is a class of machine learning model designed to learn the underlying probability distribution of a training dataset and use that knowledge to produce new data samples that resemble the original. Unlike discriminative models, which learn to map inputs to labels by estimating conditional probabilities, generative models capture the joint distribution of inputs and outputs — or inputs alone — enabling them to synthesize entirely new examples. This distinction makes them foundational to tasks where the goal is creation rather than classification.
Generative models operate through a variety of mechanisms depending on their architecture. Variational Autoencoders (VAEs) encode data into a compressed latent space and learn to decode samples from that space back into realistic outputs. Generative Adversarial Networks (GANs) pit two neural networks — a generator and a discriminator — against each other in a minimax game, driving the generator to produce increasingly convincing outputs. Diffusion models, which have risen to prominence more recently, learn to reverse a gradual noising process, reconstructing coherent data from pure noise. Each approach involves different trade-offs between training stability, output diversity, and computational cost.
The practical applications of generative models span an enormous range of domains. In computer vision, they power image synthesis, super-resolution, inpainting, and style transfer. In natural language processing, large autoregressive language models generate coherent text, code, and dialogue. In science and medicine, generative models assist with drug molecule design, protein structure prediction, and synthetic data generation for privacy-preserving research. Their ability to augment datasets also makes them valuable when labeled data is scarce.
Generative models have become one of the most active and consequential areas of modern AI research. The introduction of GANs in 2014 catalyzed enormous interest, and subsequent advances — including transformer-based language models and score-based diffusion models — have dramatically expanded what generative systems can produce. As these models grow more capable, they raise important questions around authenticity, intellectual property, and the potential for misuse, making them as socially significant as they are technically impressive.