Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Baseline

Baseline

A reference model used to benchmark whether new AI approaches actually improve performance.

Year: 2012Generality: 795
Back to Vocab

In machine learning, a baseline is a reference point against which the performance of new models, algorithms, or techniques is measured. Rather than evaluating a model in isolation, researchers compare it to a baseline to determine whether their innovations produce genuine improvements. Baselines range from trivially simple heuristics — such as always predicting the majority class in a classification task — to strong prior methods that represent the current state of the art. Choosing an appropriate baseline is as important as designing the model itself: a baseline that is too weak makes modest gains look impressive, while one that is too strong can obscure real progress.

Baselines function by establishing a performance floor using a well-understood, reproducible method on the same dataset and evaluation metric as the proposed approach. Common choices include random predictors, rule-based systems, linear models, or the best previously published result on a given benchmark. In reinforcement learning, a baseline often refers specifically to a value function used to reduce variance in policy gradient estimates — a distinct but related usage of the term. In both cases, the baseline provides a stable reference that makes comparisons interpretable and scientifically meaningful.

The importance of baselines grew substantially with the proliferation of public benchmarks and ML competitions in the 2010s. Datasets like ImageNet, GLUE, and SQuAD formalized the practice of reporting results relative to established reference points, making it easier to track genuine progress across the research community. Without strong baselines, it becomes difficult to distinguish meaningful advances from results that merely exploit quirks in evaluation protocols or benefit from additional compute and data.

Beyond academic research, baselines are equally critical in applied ML settings. When deploying a new recommendation system or fraud detection model, practitioners compare against the existing production system as a baseline to justify the cost and risk of switching. A well-chosen baseline enforces intellectual honesty, guards against overfitting to benchmarks, and ensures that reported improvements translate to real-world value.

Related

Related

Benchmark
Benchmark

A standardized test used to measure and compare AI model performance.

Generality: 796
Base Model
Base Model

A pre-trained model used as a starting point for task-specific adaptation.

Generality: 794
Foundation Model
Foundation Model

A large pre-trained model adaptable to many tasks without retraining from scratch.

Generality: 838
Biomarkers
Biomarkers

Measurable biological indicators used by AI models to assess health and disease.

Generality: 679
Bias
Bias

Systematic errors in data or algorithms that produce unfair or skewed outcomes.

Generality: 854
Comparative Advantage
Comparative Advantage

The relative edge one AI model or approach holds over others for specific tasks.

Generality: 384