An algorithm that combines multiple ML models to improve prediction accuracy.
A meta-classifier is a higher-order learning system that aggregates the outputs of multiple base models — often called weak learners — to produce predictions that are more accurate and robust than any single model could achieve alone. Rather than relying on one algorithm's perspective on the data, a meta-classifier synthesizes diverse viewpoints, exploiting the complementary strengths of its constituent models while dampening their individual weaknesses.
The mechanics of meta-classification vary by strategy. In voting or averaging schemes, base model outputs are combined directly — majority vote for classification, mean for regression. In stacking (stacked generalization), a dedicated meta-learner is trained on the predictions of the base models as its input features, learning an optimal weighting or nonlinear combination. In boosting, models are trained sequentially, with each new learner focusing on the examples the previous ones got wrong. In bagging, models are trained in parallel on bootstrapped subsets of the data, and their outputs are averaged to reduce variance. Each approach targets a different source of model error — bias, variance, or noise.
Meta-classifiers matter because real-world prediction problems are rarely solved optimally by a single algorithm. Ensemble approaches consistently rank among the top performers in machine learning competitions and benchmarks. Random Forests, Gradient Boosted Trees (including XGBoost and LightGBM), and stacked ensembles have become standard tools in applied ML precisely because they generalize better to unseen data, are less sensitive to hyperparameter choices, and degrade more gracefully when assumptions are violated. The diversity among base learners is key — if all models make the same errors, combining them offers no benefit.
The concept became practically significant in the mid-1990s with Leo Breiman's introduction of bagging (1996) and the subsequent development of AdaBoost by Freund and Schapire (1997). These foundational methods demonstrated empirically that ensemble strategies could dramatically outperform individual classifiers, spurring decades of research into how, when, and why combining models works — a question that remains active in modern deep learning through techniques like model ensembling and mixture-of-experts architectures.