Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Autoregressive

Autoregressive

A model that predicts future sequence values from weighted combinations of past values.

Year: 1970Generality: 794
Back to Vocab

An autoregressive (AR) model is a statistical framework that predicts the next value in a sequence as a linear combination of its previous values. Formally, an AR model of order p expresses the current observation as a weighted sum of the p preceding observations plus a noise term, where the weights — called autoregressive coefficients — are estimated from training data. The order p determines how far back in time the model looks, and selecting it appropriately is critical to balancing model complexity against predictive accuracy.

In machine learning, autoregressive thinking has expanded well beyond classical time series into a broad modeling paradigm. Modern autoregressive models — including PixelCNN for images, WaveNet for audio, and large language models like GPT — treat sequence generation as a chain of conditional probability estimates: each output token or value is sampled from a distribution conditioned on everything generated so far. This decomposition follows directly from the probability chain rule and allows arbitrarily complex distributions to be modeled without assuming independence between elements.

Training autoregressive models is relatively straightforward because ground-truth context is available at every step during training — a technique called teacher forcing. At inference time, however, the model must consume its own previous outputs, which can cause errors to compound over long sequences. This exposure bias is an active area of research, with approaches like scheduled sampling and reinforcement learning fine-tuning proposed as mitigations.

Autoregressive models matter because they offer a principled, tractable way to assign exact likelihoods to sequences, unlike generative adversarial networks or certain latent variable models. This property makes them valuable for density estimation, anomaly detection, and controlled generation. Their dominance in natural language processing — where the autoregressive transformer architecture underpins most state-of-the-art systems — has made understanding this concept essential for anyone working in modern AI.

Related

Related

Autoregressive Prediction
Autoregressive Prediction

A modeling approach that predicts each sequence element from its preceding values.

Generality: 792
Autoregressive Sequence Generator
Autoregressive Sequence Generator

A model that predicts each next output using its own previous outputs as inputs.

Generality: 752
Autoregressive Generation
Autoregressive Generation

Generating sequences by predicting each element conditioned on all previous outputs.

Generality: 794
Sequence Model
Sequence Model

A model that learns patterns and dependencies within ordered data sequences.

Generality: 840
Sequential Models
Sequential Models

AI models that process ordered data by capturing dependencies across time or position.

Generality: 795
Time Series Analysis
Time Series Analysis

Statistical and computational methods for analyzing chronologically ordered data to reveal patterns.

Generality: 720