Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Autoregressive Sequence Generator

Autoregressive Sequence Generator

A model that predicts each next output using its own previous outputs as inputs.

Year: 2014Generality: 752
Back to Vocab

An autoregressive sequence generator is a predictive model that produces outputs sequentially, feeding each generated element back as input for the next prediction. Rather than processing all outputs simultaneously, the model conditions each new prediction on the history of what it has already produced — or, in supervised settings, on the true prior values during training. This self-referential structure makes autoregressive models naturally suited to any domain where order and context matter: time-series forecasting, language modeling, audio synthesis, and image generation.

The mechanics vary by architecture, but the core idea is consistent. At each step, the model estimates a probability distribution over possible next values given all previous ones, then samples or selects from that distribution. In classical statistical models like AR(p) or ARIMA, this relationship is expressed as a weighted linear combination of the last p observations. In modern deep learning, the same principle is implemented through recurrent neural networks, transformers, or masked convolutional networks — all of which can capture far richer, nonlinear dependencies across long sequences.

Autoregressive models became central to deep learning after researchers demonstrated that neural language models could generate coherent, fluent text by predicting one token at a time. The transformer-based GPT family exemplifies this approach at scale: each token is predicted from all preceding tokens using self-attention, enabling the model to draw on long-range context efficiently. Similar autoregressive designs power WaveNet for audio and PixelCNN for images, showing the paradigm's versatility across modalities.

The primary trade-off of autoregressive generation is speed: because each output depends on the previous one, generation is inherently sequential and cannot be trivially parallelized at inference time. This has motivated research into speculative decoding, distillation into non-autoregressive models, and other acceleration strategies. Despite this limitation, autoregressive generators remain among the most powerful and widely deployed generative architectures in modern AI, largely because their training objective — predicting the next element — is simple, scalable, and produces models with strong generalization.

Related

Related

Autoregressive Generation
Autoregressive Generation

Generating sequences by predicting each element conditioned on all previous outputs.

Generality: 794
Autoregressive Prediction
Autoregressive Prediction

A modeling approach that predicts each sequence element from its preceding values.

Generality: 792
Autoregressive
Autoregressive

A model that predicts future sequence values from weighted combinations of past values.

Generality: 794
Sequence Prediction
Sequence Prediction

Forecasting the next item(s) in a sequence by learning patterns from prior observations.

Generality: 794
Sequence Model
Sequence Model

A model that learns patterns and dependencies within ordered data sequences.

Generality: 840
Sequential Models
Sequential Models

AI models that process ordered data by capturing dependencies across time or position.

Generality: 795