Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Ensemble Learning

Ensemble Learning

Combining multiple models to produce predictions more accurate than any single model.

Year: 1996Generality: 836
Back to Vocab

Ensemble learning is a machine learning paradigm in which multiple models are trained independently and their outputs are aggregated to produce a final prediction. The core intuition is that diverse models tend to make different errors, and combining them can cancel out individual weaknesses. This approach consistently outperforms single-model solutions across a wide range of tasks, which is why ensemble methods dominate machine learning competitions and production systems alike.

The three dominant ensemble strategies are bagging, boosting, and stacking. Bagging (bootstrap aggregating) trains multiple instances of the same model on different random subsamples of the training data and averages their predictions, reducing variance without significantly increasing bias — Random Forests are the canonical example. Boosting takes a sequential approach, training each new model to correct the errors made by its predecessors, with AdaBoost and Gradient Boosting Machines being the most influential implementations. Stacking goes further by training a meta-learner to optimally combine the outputs of several heterogeneous base models, allowing the system to learn which models to trust under which conditions.

Ensemble methods are particularly effective because they address the bias-variance tradeoff from multiple angles simultaneously. Bagging primarily reduces variance, making it well-suited for high-variance models like deep decision trees. Boosting primarily reduces bias, making it powerful for weak learners that individually underfit the data. In practice, gradient boosted trees — implemented in libraries like XGBoost, LightGBM, and CatBoost — have become the default choice for structured tabular data, routinely outperforming deep learning on such tasks.

The practical impact of ensemble learning is difficult to overstate. It underpins many real-world systems in finance, healthcare, and recommendation engines, and has been the winning strategy in a large proportion of Kaggle competitions. Understanding ensemble methods is essential for any practitioner, as they represent one of the most reliable and well-understood tools for improving predictive performance with relatively modest computational overhead.

Related

Related

Ensemble Methods
Ensemble Methods

Combining multiple trained models to produce predictions stronger than any single model.

Generality: 771
Ensemble Algorithm
Ensemble Algorithm

Combines multiple models to boost predictive accuracy, robustness, and generalization.

Generality: 796
Bagging
Bagging

Ensemble method that trains multiple models on random data subsets and aggregates predictions.

Generality: 694
Boosting
Boosting

An ensemble method that combines weak learners sequentially into a strong predictor.

Generality: 796
Stacking
Stacking

An ensemble method that trains a meta-model on the outputs of multiple base models.

Generality: 650
Meta-Classifier
Meta-Classifier

An algorithm that combines multiple ML models to improve prediction accuracy.

Generality: 660