Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Upweighting

Upweighting

Increasing the influence of selected data points or features during model training.

Year: 1997Generality: 620
Back to Vocab

Upweighting is a training technique in machine learning that assigns greater importance to specific data points, samples, or features by increasing their associated weights during the optimization process. Rather than treating all examples equally, upweighting instructs the model to pay more attention to certain instances — effectively making errors on those examples more costly and pushing the learned parameters to better capture their patterns. This is typically implemented by multiplying sample losses by a scalar weight greater than one before computing gradients, or by oversampling targeted examples so they appear more frequently in each training batch.

The most common motivation for upweighting is class imbalance. When a dataset contains far fewer examples of one class than another — such as rare disease cases in a medical dataset — a model trained naively will often ignore the minority class entirely. Upweighting minority-class samples forces the model to treat misclassifications there as disproportionately important, improving recall and overall fairness. A related application is domain adaptation and fine-tuning: when a practitioner wants a pretrained model to specialize on a particular distribution or task, upweighting examples from that target domain steers the model accordingly without discarding the broader training corpus.

Upweighting is also central to several foundational ML algorithms. Boosting methods like AdaBoost operate by iteratively upweighting the examples that previous weak learners misclassified, forcing each successive learner to focus on the hardest cases. This principled use of adaptive sample weights was instrumental in demonstrating that ensembles of weak classifiers could achieve strong generalization. In modern deep learning, similar ideas appear in focal loss, which dynamically downweights easy examples and upweights hard or rare ones, and in reinforcement learning from human feedback (RLHF), where certain preference examples may be weighted more heavily to align model behavior with specific values.

Choosing appropriate upweighting factors requires care. Excessive upweighting of noisy or mislabeled examples can degrade overall model quality, and the optimal weights are often dataset- and task-dependent. Practitioners typically tune weights through cross-validation or use principled heuristics based on class frequency. Despite these challenges, upweighting remains one of the most practical and widely used tools for improving model fairness, robustness, and targeted performance.

Related

Related

Weight
Weight

A learnable parameter that scales the influence of inputs within a model.

Generality: 850
Boosting
Boosting

An ensemble method that combines weak learners sequentially into a strong predictor.

Generality: 796
Participation Bias
Participation Bias

A dataset imbalance where certain groups are over- or underrepresented, skewing model outcomes.

Generality: 524
Sampling Bias
Sampling Bias

A data flaw where training samples misrepresent the true population, distorting model behavior.

Generality: 794
Fine-Tuning
Fine-Tuning

Adapting a pre-trained model to a specific task by continuing training on new data.

Generality: 796
Coverage Bias
Coverage Bias

A dataset imbalance where underrepresented groups cause skewed model performance.

Generality: 520