Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. In-Group Bias

In-Group Bias

AI systems unfairly favoring certain demographic groups due to biased training data.

Year: 2016Generality: 520
Back to Vocab

In-group bias in machine learning refers to the tendency of AI systems to produce outputs that systematically favor one demographic, social, or cultural group over others. This phenomenon typically emerges when training data reflects existing societal inequalities — if historical records, text corpora, or labeled datasets over-represent certain groups or encode preferential treatment toward them, models trained on that data will learn and perpetuate those patterns. The result is a system that performs better, assigns higher scores, or makes more favorable decisions for members of the dominant group, while disadvantaging those in out-groups.

The mechanism behind in-group bias is closely tied to how models learn statistical associations. A language model trained predominantly on text authored by or about a particular demographic may develop stronger, more accurate representations for that group. Similarly, a facial recognition system trained mostly on lighter-skinned faces will generalize poorly to darker-skinned individuals. Because these disparities are embedded in the learned weights rather than explicit rules, they can be difficult to detect without targeted auditing. Bias can also be amplified through feedback loops, where biased model outputs influence future data collection, reinforcing the original skew.

Addressing in-group bias has become a central concern in the field of algorithmic fairness. Mitigation strategies operate at multiple stages of the ML pipeline: at the data level through resampling, reweighting, or curating more representative datasets; at the model level through fairness-aware training objectives and regularization; and at the output level through post-processing techniques that calibrate predictions across groups. Evaluation requires disaggregated metrics — measuring performance separately across demographic subgroups rather than relying on aggregate accuracy, which can mask significant disparities.

The practical stakes are high. AI systems exhibiting in-group bias have been documented in consequential domains including hiring, credit scoring, criminal risk assessment, and medical diagnosis, where biased outputs can cause real harm to already marginalized populations. This has driven regulatory interest and spurred the development of formal fairness criteria — such as demographic parity, equalized odds, and individual fairness — that provide rigorous frameworks for defining and measuring what it means for a model to treat groups equitably.

Related

Related

Out-group Homogeneity Bias
Out-group Homogeneity Bias

The tendency to perceive out-group members as more similar to each other than in-group members.

Generality: 380
Algorithmic Bias
Algorithmic Bias

Systematic unfairness embedded in algorithmic outputs due to biased data or design.

Generality: 792
Bias
Bias

Systematic errors in data or algorithms that produce unfair or skewed outcomes.

Generality: 854
Historical Bias
Historical Bias

Bias in AI systems inherited from prejudiced or unrepresentative historical training data.

Generality: 626
Participation Bias
Participation Bias

A dataset imbalance where certain groups are over- or underrepresented, skewing model outcomes.

Generality: 524
Fairness-Aware Machine Learning
Fairness-Aware Machine Learning

Building ML algorithms that produce equitable outcomes across demographic groups.

Generality: 694