Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Sequence Model

Sequence Model

A model that learns patterns and dependencies within ordered data sequences.

Year: 1997Generality: 840
Back to Vocab

A sequence model is a class of machine learning model designed to process, analyze, and generate data where order matters — such as text, audio, time series, or genomic data. Unlike models that treat inputs as independent and interchangeable, sequence models explicitly account for the temporal or positional relationships between elements, recognizing that the meaning or value of any single element often depends on what came before or after it. This makes them essential for tasks like language modeling, speech recognition, machine translation, and protein structure prediction.

The core challenge sequence models address is learning dependencies across varying distances within a sequence. Early approaches used Recurrent Neural Networks (RNNs), which process inputs one step at a time while maintaining a hidden state that carries information forward. This allows the model to accumulate context, but vanilla RNNs struggle with long-range dependencies due to the vanishing gradient problem — gradients shrink exponentially as they propagate back through many time steps. Long Short-Term Memory networks (LSTMs) and Gated Recurrent Units (GRUs) addressed this with gating mechanisms that selectively retain or discard information, enabling the model to learn dependencies spanning hundreds of steps.

The introduction of the Transformer architecture in 2017 marked a turning point for sequence modeling. Rather than processing tokens sequentially, Transformers apply self-attention mechanisms that directly compute relationships between all pairs of positions in a sequence simultaneously. This parallelism dramatically accelerates training and allows the model to capture long-range dependencies without the compounding errors of recurrent processing. Transformers became the foundation for large language models like BERT and GPT, which have set state-of-the-art benchmarks across virtually every sequence-based NLP task.

Sequence models matter because so much real-world data is inherently sequential. Language, music, financial markets, sensor readings, and biological signals all unfold over time or position, and capturing that structure is critical to making accurate predictions or generating coherent outputs. As architectures have evolved from RNNs to attention-based models, the range and quality of tasks sequence models can handle has expanded dramatically, making them one of the most impactful paradigms in modern machine learning.

Related

Related

Sequential Models
Sequential Models

AI models that process ordered data by capturing dependencies across time or position.

Generality: 795
Sequence Prediction
Sequence Prediction

Forecasting the next item(s) in a sequence by learning patterns from prior observations.

Generality: 794
Seq2Seq (Sequence-to-Sequence)
Seq2Seq (Sequence-to-Sequence)

A neural architecture that maps variable-length input sequences to variable-length output sequences.

Generality: 794
Autoregressive Sequence Generator
Autoregressive Sequence Generator

A model that predicts each next output using its own previous outputs as inputs.

Generality: 752
RNN (Recurrent Neural Network)
RNN (Recurrent Neural Network)

Neural networks with feedback connections that process sequential data using internal memory.

Generality: 838
Encoder-Decoder Models
Encoder-Decoder Models

Deep learning architectures that compress input into a representation and generate output.

Generality: 792