Combines multiple models to boost predictive accuracy, robustness, and generalization.
An ensemble algorithm is a machine learning approach that combines the predictions of multiple individual models—called base learners or weak learners—to produce a single, more accurate and reliable output. Rather than relying on any one model's judgment, ensemble methods exploit the diversity among models to cancel out individual errors, reduce variance, lower bias, or smooth out noise. This collective decision-making consistently outperforms single-model approaches across a wide range of tasks, from classification and regression to ranking and anomaly detection.
The two dominant families of ensemble methods are bagging and boosting. Bagging (bootstrap aggregating) trains multiple models independently on different random subsets of the training data and averages their outputs—Random Forest is the canonical example, applying this strategy to decision trees. Boosting, by contrast, trains models sequentially, with each new model focusing on the examples that previous models got wrong; AdaBoost and Gradient Boosting Machines (GBM) are the most influential implementations. A third approach, stacking, trains a meta-model to learn how best to combine the predictions of several diverse base models, often yielding further gains in accuracy.
Ensemble methods became a cornerstone of practical machine learning in the 1990s and 2000s, with Leo Breiman's introduction of bagging in 1994 and the development of AdaBoost by Freund and Schapire shortly after. Their dominance was cemented in competitive machine learning: gradient boosting variants such as XGBoost, LightGBM, and CatBoost have won a remarkable share of structured-data competitions on platforms like Kaggle, while Random Forests remain a go-to baseline for practitioners.
The power of ensemble methods stems from a well-understood bias-variance tradeoff. Individual models are often either too simple (high bias) or too sensitive to training data (high variance); combining many such models statistically reduces these errors provided the base learners are sufficiently diverse. This diversity can be introduced through different data subsets, different feature subsets, different model architectures, or different random initializations. As a result, ensemble algorithms are among the most reliable and widely deployed techniques in applied machine learning.