Increasing the influence of selected data points or features during model training.
Upweighting is a training technique in machine learning that assigns greater importance to specific data points, samples, or features by increasing their associated weights during the optimization process. Rather than treating all examples equally, upweighting instructs the model to pay more attention to certain instances — effectively making errors on those examples more costly and pushing the learned parameters to better capture their patterns. This is typically implemented by multiplying sample losses by a scalar weight greater than one before computing gradients, or by oversampling targeted examples so they appear more frequently in each training batch.
The most common motivation for upweighting is class imbalance. When a dataset contains far fewer examples of one class than another — such as rare disease cases in a medical dataset — a model trained naively will often ignore the minority class entirely. Upweighting minority-class samples forces the model to treat misclassifications there as disproportionately important, improving recall and overall fairness. A related application is domain adaptation and fine-tuning: when a practitioner wants a pretrained model to specialize on a particular distribution or task, upweighting examples from that target domain steers the model accordingly without discarding the broader training corpus.
Upweighting is also central to several foundational ML algorithms. Boosting methods like AdaBoost operate by iteratively upweighting the examples that previous weak learners misclassified, forcing each successive learner to focus on the hardest cases. This principled use of adaptive sample weights was instrumental in demonstrating that ensembles of weak classifiers could achieve strong generalization. In modern deep learning, similar ideas appear in focal loss, which dynamically downweights easy examples and upweights hard or rare ones, and in reinforcement learning from human feedback (RLHF), where certain preference examples may be weighted more heavily to align model behavior with specific values.
Choosing appropriate upweighting factors requires care. Excessive upweighting of noisy or mislabeled examples can degrade overall model quality, and the optimal weights are often dataset- and task-dependent. Practitioners typically tune weights through cross-validation or use principled heuristics based on class frequency. Despite these challenges, upweighting remains one of the most practical and widely used tools for improving model fairness, robustness, and targeted performance.