Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Non-Response Bias

Non-Response Bias

Skew introduced when survey non-respondents differ systematically from respondents.

Year: 2000Generality: 383
Back to Vocab

Non-response bias is a form of selection bias that arises when individuals who do not participate in a survey or data collection process differ in meaningful ways from those who do. In machine learning, this matters because models are only as representative as the data they are trained on. If a training dataset is built from survey responses, user feedback, or opt-in interactions, the people who choose not to respond may hold systematically different characteristics, opinions, or behaviors — and their absence quietly distorts the dataset's picture of reality.

The mechanism is straightforward but insidious. Suppose a health app collects symptom data from users who voluntarily complete weekly check-ins. Users who feel well may be less motivated to respond, while those experiencing symptoms engage more consistently. A model trained on this data would overestimate disease prevalence and underweight healthy baselines. Similarly, customer satisfaction surveys tend to attract strongly positive or strongly negative respondents, leaving moderate opinions underrepresented. In each case, the model learns from a skewed slice of the population and generalizes poorly to the full distribution it is meant to serve.

Mitigating non-response bias typically involves a combination of data collection design and post-hoc statistical correction. On the collection side, strategies include follow-up outreach, incentive structures, and reducing survey burden to improve response rates across diverse groups. After data collection, analysts may apply inverse probability weighting — upweighting responses from underrepresented groups based on known population characteristics — or use imputation methods to estimate missing values for non-respondents. Multiple imputation and propensity score adjustments are common tools borrowed from survey methodology and adapted for ML pipelines.

The stakes are particularly high when models inform decisions in sensitive domains such as healthcare, criminal justice, or credit scoring. A model trained on non-representative feedback can systematically disadvantage the very groups least likely to have participated in its training data, compounding existing inequities. Recognizing non-response bias as a data quality issue — not merely a statistical nuisance — is essential for building models that are both accurate and fair.

Related

Related

Participation Bias
Participation Bias

A dataset imbalance where certain groups are over- or underrepresented, skewing model outcomes.

Generality: 524
Sampling Bias
Sampling Bias

A data flaw where training samples misrepresent the true population, distorting model behavior.

Generality: 794
Reporting Bias
Reporting Bias

A systematic distortion in training data caused by selective omission of outcomes or observations.

Generality: 694
Coverage Bias
Coverage Bias

A dataset imbalance where underrepresented groups cause skewed model performance.

Generality: 520
Bias
Bias

Systematic errors in data or algorithms that produce unfair or skewed outcomes.

Generality: 854
De-Biasing
De-Biasing

Techniques that reduce unfair bias in machine learning models and their outputs.

Generality: 694