Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Base Model

Base Model

A pre-trained model used as a starting point for task-specific adaptation.

Year: 2018Generality: 794
Back to Vocab

A base model is a large neural network trained on broad, diverse data that serves as a reusable foundation for downstream tasks. Rather than learning from random initialization, practitioners begin with a base model that has already internalized rich representations of language, images, or other modalities. This pre-trained knowledge dramatically reduces the data and compute required to build capable, specialized systems.

The mechanics of using a base model typically involve one of two approaches: fine-tuning, where the model's weights are further updated on task-specific data, or prompting and in-context learning, where the model is queried directly without any weight modification. In both cases, the base model's learned representations act as a powerful prior. Architectures like transformers have proven especially effective as base models because their attention mechanisms capture long-range dependencies and generalize well across domains.

Base models became central to modern AI practice with the release of large-scale pretrained systems such as BERT and GPT-2 around 2018–2019, which demonstrated that a single model trained on internet-scale text could be adapted to dozens of benchmarks with minimal additional effort. The paradigm shifted research and industry workflows alike: instead of training task-specific models from scratch, teams now routinely start from a shared base. This has democratized access to high-performance AI, since organizations without massive compute budgets can still build competitive systems by fine-tuning publicly released base models.

The importance of base models extends beyond convenience. They encode a form of transferable knowledge that often generalizes to tasks and distributions not seen during pretraining, a property known as transfer learning. As base models have grown in scale—from hundreds of millions to hundreds of billions of parameters—their emergent capabilities have expanded, enabling few-shot and zero-shot performance that was previously unattainable. Choosing, evaluating, and responsibly deploying base models has consequently become one of the most consequential decisions in applied machine learning.

Related

Related

Foundation Model
Foundation Model

A large pre-trained model adaptable to many tasks without retraining from scratch.

Generality: 838
Pretrained Model
Pretrained Model

A model trained on large data, reused or fine-tuned for new tasks.

Generality: 838
Baseline
Baseline

A reference model used to benchmark whether new AI approaches actually improve performance.

Generality: 795
Self-Supervised Pretraining
Self-Supervised Pretraining

A technique where models learn rich representations from unlabeled data before fine-tuning on specific tasks.

Generality: 794
Fine-Tuning
Fine-Tuning

Adapting a pre-trained model to a specific task by continuing training on new data.

Generality: 796
Transfer Learning
Transfer Learning

Reusing a model trained on one task to accelerate learning on another.

Generality: 820