Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Ensemble Methods

Ensemble Methods

Combining multiple trained models to produce predictions stronger than any single model.

Year: 1996Generality: 771
Back to Vocab

Ensemble methods are machine learning techniques that aggregate the outputs of multiple models—called base learners or weak learners—to produce a single, more accurate and robust prediction. The core intuition is that diverse models tend to make different errors, and by combining them intelligently, those errors partially cancel out. This principle, sometimes called the "wisdom of crowds," allows ensembles to outperform any individual constituent model, even when each base learner is only marginally better than random guessing.

The three dominant ensemble strategies are bagging, boosting, and stacking. Bagging (Bootstrap Aggregating) trains multiple models independently on random subsamples of the training data and averages their predictions, reducing variance—Random Forests are the canonical example. Boosting trains models sequentially, with each new model focusing on the examples the previous ones got wrong, reducing bias—Gradient Boosted Trees and AdaBoost are well-known implementations. Stacking takes a different approach: it trains a meta-learner to optimally combine the predictions of several heterogeneous base models, learning how to weight each contributor based on their strengths.

Ensemble methods matter because they consistently rank among the highest-performing approaches across a wide range of supervised learning tasks. Gradient boosting libraries like XGBoost, LightGBM, and CatBoost have dominated structured-data competitions for years, while bagging-based forests remain a reliable baseline in production systems. The trade-off is interpretability and computational cost—ensembles are harder to explain than single decision trees and require training and storing multiple models. Nevertheless, their empirical reliability makes them a cornerstone of practical machine learning, and understanding them is essential for anyone working on predictive modeling.

Related

Related

Ensemble Learning
Ensemble Learning

Combining multiple models to produce predictions more accurate than any single model.

Generality: 836
Ensemble Algorithm
Ensemble Algorithm

Combines multiple models to boost predictive accuracy, robustness, and generalization.

Generality: 796
Bagging
Bagging

Ensemble method that trains multiple models on random data subsets and aggregates predictions.

Generality: 694
Boosting
Boosting

An ensemble method that combines weak learners sequentially into a strong predictor.

Generality: 796
Stacking
Stacking

An ensemble method that trains a meta-model on the outputs of multiple base models.

Generality: 650
Meta-Classifier
Meta-Classifier

An algorithm that combines multiple ML models to improve prediction accuracy.

Generality: 660