Systematic unfairness embedded in algorithmic outputs due to biased data or design.
Algorithmic bias occurs when machine learning models or automated decision systems produce outputs that systematically disadvantage certain individuals or groups based on characteristics such as race, gender, age, or socioeconomic status. This bias can originate from multiple sources: training data that reflects historical inequalities, design choices that inadvertently encode human prejudices, or feedback loops that amplify small initial disparities over time. Because modern ML systems learn statistical patterns from data, they readily absorb and reproduce whatever biases exist in that data — sometimes in subtle, hard-to-detect ways.
Detecting algorithmic bias typically involves auditing model outputs across demographic subgroups and measuring disparities in error rates, approval rates, or other outcomes. Common fairness metrics include demographic parity (equal outcome rates across groups), equalized odds (equal true and false positive rates), and individual fairness (similar treatment for similar individuals). A persistent challenge is that these metrics are often mathematically incompatible with one another, meaning that optimizing for one definition of fairness can worsen another — a tension that has no purely technical resolution and requires normative judgment about what fairness means in a given context.
Mitigation strategies operate at multiple stages of the ML pipeline. Pre-processing techniques rebalance or re-weight training data to reduce representational skew. In-processing methods incorporate fairness constraints directly into the model's objective function during training. Post-processing approaches adjust model outputs after the fact to equalize outcomes across groups. None of these methods eliminates bias entirely, and each involves trade-offs with predictive accuracy or other performance goals.
The stakes are highest in domains where algorithmic decisions carry serious real-world consequences — criminal risk scoring, loan approvals, medical diagnosis, and hiring. In these settings, biased models can entrench and amplify existing social inequalities at scale, affecting millions of people with limited opportunity for recourse. This has driven growing regulatory attention, with frameworks such as the EU AI Act requiring bias assessments for high-risk systems, and has made algorithmic fairness a central concern in responsible AI development.