Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Ensemble Algorithm

Ensemble Algorithm

Combines multiple models to boost predictive accuracy, robustness, and generalization.

Year: 1994Generality: 796
Back to Vocab

An ensemble algorithm is a machine learning approach that combines the predictions of multiple individual models—called base learners or weak learners—to produce a single, more accurate and reliable output. Rather than relying on any one model's judgment, ensemble methods exploit the diversity among models to cancel out individual errors, reduce variance, lower bias, or smooth out noise. This collective decision-making consistently outperforms single-model approaches across a wide range of tasks, from classification and regression to ranking and anomaly detection.

The two dominant families of ensemble methods are bagging and boosting. Bagging (bootstrap aggregating) trains multiple models independently on different random subsets of the training data and averages their outputs—Random Forest is the canonical example, applying this strategy to decision trees. Boosting, by contrast, trains models sequentially, with each new model focusing on the examples that previous models got wrong; AdaBoost and Gradient Boosting Machines (GBM) are the most influential implementations. A third approach, stacking, trains a meta-model to learn how best to combine the predictions of several diverse base models, often yielding further gains in accuracy.

Ensemble methods became a cornerstone of practical machine learning in the 1990s and 2000s, with Leo Breiman's introduction of bagging in 1994 and the development of AdaBoost by Freund and Schapire shortly after. Their dominance was cemented in competitive machine learning: gradient boosting variants such as XGBoost, LightGBM, and CatBoost have won a remarkable share of structured-data competitions on platforms like Kaggle, while Random Forests remain a go-to baseline for practitioners.

The power of ensemble methods stems from a well-understood bias-variance tradeoff. Individual models are often either too simple (high bias) or too sensitive to training data (high variance); combining many such models statistically reduces these errors provided the base learners are sufficiently diverse. This diversity can be introduced through different data subsets, different feature subsets, different model architectures, or different random initializations. As a result, ensemble algorithms are among the most reliable and widely deployed techniques in applied machine learning.

Related

Related

Ensemble Methods
Ensemble Methods

Combining multiple trained models to produce predictions stronger than any single model.

Generality: 771
Ensemble Learning
Ensemble Learning

Combining multiple models to produce predictions more accurate than any single model.

Generality: 836
Bagging
Bagging

Ensemble method that trains multiple models on random data subsets and aggregates predictions.

Generality: 694
Boosting
Boosting

An ensemble method that combines weak learners sequentially into a strong predictor.

Generality: 796
Meta-Classifier
Meta-Classifier

An algorithm that combines multiple ML models to improve prediction accuracy.

Generality: 660
Stacking
Stacking

An ensemble method that trains a meta-model on the outputs of multiple base models.

Generality: 650