Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Meta-Regressor

Meta-Regressor

An ensemble model that learns from base regressors' predictions to produce a final output.

Year: 1996Generality: 451
Back to Vocab

A meta-regressor is a second-level model in a stacked ensemble architecture that takes the predictions of multiple base regression models as its input features and learns to combine them into a single, more accurate final prediction. Rather than relying on any one model's output, the meta-regressor treats those outputs as a new feature space, discovering how to weight and blend them in ways that reduce overall prediction error. This technique is a specific application of stacking — or stacked generalization — applied to regression tasks.

In practice, training a meta-regressor requires care to avoid data leakage. Base models are typically trained using cross-validation, and their out-of-fold predictions on the training set are used to train the meta-regressor. This ensures the second-level model learns from predictions that the base models made on data they had not seen, producing a more honest and generalizable combination. At inference time, all base models generate predictions on new data, and the meta-regressor combines those predictions into the final output.

The power of a meta-regressor lies in its ability to exploit the complementary strengths of diverse base models. A linear model may capture global trends efficiently, while a gradient-boosted tree handles local nonlinearities, and a neural network models complex interactions. The meta-regressor learns which model to trust under which conditions, often outperforming any individual constituent. This makes the approach especially valuable in domains like finance, healthcare, and climate modeling, where prediction accuracy carries significant real-world consequences.

Meta-regressors became a practical tool in machine learning following David Wolpert's formalization of stacked generalization in 1992, but gained widespread adoption through the 1990s and 2000s as ensemble methods proved their value in competitions and applied research. They remain a staple technique in modern predictive modeling pipelines, frequently appearing in winning solutions to regression benchmarks and production forecasting systems alike.

Related

Related

Meta-Classifier
Meta-Classifier

An algorithm that combines multiple ML models to improve prediction accuracy.

Generality: 660
Stacking
Stacking

An ensemble method that trains a meta-model on the outputs of multiple base models.

Generality: 650
Regression
Regression

A supervised learning approach that predicts continuous numerical outcomes from input variables.

Generality: 909
Ensemble Algorithm
Ensemble Algorithm

Combines multiple models to boost predictive accuracy, robustness, and generalization.

Generality: 796
Ensemble Methods
Ensemble Methods

Combining multiple trained models to produce predictions stronger than any single model.

Generality: 771
Meta-Learning
Meta-Learning

A paradigm enabling models to learn how to learn across tasks efficiently.

Generality: 756