Combining multiple trained models to produce predictions stronger than any single model.
Ensemble methods are machine learning techniques that aggregate the outputs of multiple models—called base learners or weak learners—to produce a single, more accurate and robust prediction. The core intuition is that diverse models tend to make different errors, and by combining them intelligently, those errors partially cancel out. This principle, sometimes called the "wisdom of crowds," allows ensembles to outperform any individual constituent model, even when each base learner is only marginally better than random guessing.
The three dominant ensemble strategies are bagging, boosting, and stacking. Bagging (Bootstrap Aggregating) trains multiple models independently on random subsamples of the training data and averages their predictions, reducing variance—Random Forests are the canonical example. Boosting trains models sequentially, with each new model focusing on the examples the previous ones got wrong, reducing bias—Gradient Boosted Trees and AdaBoost are well-known implementations. Stacking takes a different approach: it trains a meta-learner to optimally combine the predictions of several heterogeneous base models, learning how to weight each contributor based on their strengths.
Ensemble methods matter because they consistently rank among the highest-performing approaches across a wide range of supervised learning tasks. Gradient boosting libraries like XGBoost, LightGBM, and CatBoost have dominated structured-data competitions for years, while bagging-based forests remain a reliable baseline in production systems. The trade-off is interpretability and computational cost—ensembles are harder to explain than single decision trees and require training and storing multiple models. Nevertheless, their empirical reliability makes them a cornerstone of practical machine learning, and understanding them is essential for anyone working on predictive modeling.