Get AI summaries of any video or article — Sign up free

Discrimination, artificial intelligence, and algorithmic decision-making

Frederik Zuiderveen Borgesius
ArXiv.org·2025·Decision Sciences·156 citations
9 min read

Read the full paper at DOI or on arxiv

TL;DR

The report maps AI discrimination risks to six machine-learning mechanisms: target/class label choices, biased training data, biased data collection and feedback loops, feature selection, proxy effects/redundant encodings, and intentional discrimination.

Briefing

This report by Frederik Zuiderveen Borgesius (Council of Europe, Anti-discrimination department) addresses a central question for decision sciences and human-rights governance: how and where algorithmic decision-making and other AI systems can produce discriminatory effects, what legal safeguards currently exist to mitigate those risks, and what recommendations should be made to organizations deploying AI, equality bodies, and human-rights monitoring institutions.

The question matters because AI is increasingly used to make or support decisions with far-reaching consequences for individuals in both the public and private sectors. The report emphasizes that AI can appear neutral or rational, but can still generate unfair or illegal discrimination. It also highlights a compounding mechanism: AI systems can learn from biased human decisions, and many small discriminatory or exclusionary decisions—such as targeted advertising—can aggregate into large-scale social inequality.

The report’s significance lies in its cross-disciplinary framing. Rather than treating discrimination as only a legal concept, it connects discrimination pathways in machine learning to existing European non-discrimination and data protection regimes. It also argues that current law may not fully cover “new types of unfair differentiation” created by AI, especially when differentiation is based on invented categories that do not map neatly onto legally protected characteristics (e.g., race or gender). Consequently, the report positions AI discrimination as both a compliance problem (enforcing existing norms) and a regulatory design problem (potentially requiring sector-specific rules and new enforcement mechanisms).

Methodologically, the report is not an empirical study with a dataset or statistical analysis. It is a literature review and “quick scan” intended to map relevant issues and safeguards. The author synthesizes prior academic work (notably Barocas and Selbst’s framework for disparate impact in machine learning) and uses illustrative real-world cases (e.g., COMPAS in criminal justice; biased AI hiring systems; discriminatory ad targeting; online price discrimination; facial recognition errors; and gender bias in translation tools). The report’s “analysis technique” is therefore conceptual and legal-analytical: it organizes discrimination risks into mechanisms, then evaluates how legal regimes respond to those mechanisms.

Key findings are presented primarily as structured claims and concrete case-based evidence rather than as new quantitative results. The report identifies at least six pathways through which AI can lead to discrimination:

1) Defining the target variable and class labels in ways that embed social disadvantage (e.g., defining “good employee” using outcomes correlated with protected groups). 2) Biased training data, including when historical human decisions were discriminatory. 3) Biased data collection processes, including feedback loops (e.g., policing attention that increases recorded crime in certain neighborhoods). 4) Biased feature selection, where proxy variables for protected characteristics are chosen because they are predictive. 5) Proxy effects and “redundant encodings,” where protected-class membership is encoded in other variables; the report notes that one approach to avoid systematic disadvantage can reduce overall accuracy. 6) Intentional discrimination, where proxies are selected with discriminatory intent.

The report then provides field examples to show how these mechanisms manifest:

  • Criminal justice and risk assessment: The COMPAS example is used to illustrate disparate error patterns. The report cites ProPublica’s findings that COMPAS “correctly predicts recidivism 61 percent of the time,” while Black defendants were “almost twice as likely as whites to be labelled a higher risk but not actually reoffend,” and that Black defendants were “twice as likely” to be misclassified as higher risk of violent recidivism. It also discusses the calibration and fairness trade-offs, noting that equalizing error rates can be mathematically impossible when base rates differ.

  • Employment and education: The report describes a UK medical school admissions example where an AI system trained on historical admissions decisions reproduced bias against women and people with an immigrant background. It also references Amazon’s decision to stop using an AI recruiting tool after it was found to be biased against women.

  • Advertising: The report cites evidence that online ad systems can encode discrimination through targeting and exclusion. Examples include:

  • Google ad delivery differences by race-associated names (Sweeney, 2013).
  • Simulated-user experiments suggesting gendered ad delivery for high-paying jobs (Datta, Tschantz, Datta, 2015).
  • Facebook ad targeting/exclusion by race and other protected or sensitive “ethnic affinities” (ProPublica reporting), and the Dutch Data Protection Authority’s findings that Facebook enabled advertisers to target based on sensitive characteristics.

  • Online price discrimination: The report describes Princeton Review’s differentiated pricing across US areas, citing that customers in areas with high density of Asian residents were “1.8 times as likely to be offered higher prices, regardless of income” (Angwin et al., 2015).

  • Image search and recognition: The report cites examples where image search results reflect societal stereotypes (e.g., “three black teenagers” vs “three white kids”) and where face recognition systems show higher error rates for darker-skinned groups. It specifically cites Buolamwini and Gebru’s finding that “darker-skinned females are the most misclassified group (with error rates of up to 34.7%),” while “the maximum error rate for lighter-skinned males is 0.8%.”

  • Translation tools: The report cites research showing gender bias in machine translation, including that gender-neutral source sentences can produce gendered defaults when translated back into English, and that male defaults are “strong” and “exaggerated” for stereotype-associated fields such as STEM.

Limitations are acknowledged in the report’s framing: it is a “quick scan” constrained by length, based on literature review rather than original empirical measurement. It also explicitly limits scope: it focuses on discrimination risks and does not deeply cover privacy-related questions, automated weapons, unemployment, filter bubbles, or data monopolies. Additionally, the report notes that it is “too early” to assess the practical effect of newer legal provisions (e.g., GDPR automated decision rules) and that enforcement and compliance deficits remain.

Practical implications follow from the report’s legal and governance analysis. The author argues that the most relevant legal tools are non-discrimination law and data protection law, and that effective enforcement could mitigate many AI discrimination risks. However, the report stresses that enforcement is difficult due to (i) AI opacity (“black box” effects), (ii) evidentiary burdens for proving indirect discrimination, (iii) limited resources and enforcement powers of authorities, and (iv) gaps where AI differentiates using invented categories not covered by protected-characteristic statutes.

Accordingly, the report recommends concrete organizational steps: education to reduce automation bias and awareness of discrimination risks; risk assessment and mitigation involving multidisciplinary teams; documenting and monitoring AI projects; and considering impact assessments inspired by GDPR DPIAs. It also recommends transparency and auditability in the public sector, potentially controlled access for researchers, and “sunset clauses” for public AI systems.

For equality bodies and human-rights monitoring bodies, the report emphasizes building AI technical capacity (including hiring or consulting computer scientists), conducting public awareness campaigns, engaging in early consultation for public procurement, cooperating with data protection authorities, and possibly commissioning or supporting research and strategic litigation.

Finally, the report’s regulatory recommendations argue against a single one-size-fits-all AI law. Instead, it suggests sector-specific regulation and a multi-level approach combining broad principles with faster-updating guidelines. It also calls for improved enforcement of existing norms via transparency requirements, auditing capabilities, and adequate investigative powers.

Who should care? The report is aimed at policymakers and regulators (especially equality bodies and data protection authorities), organizations deploying AI in high-stakes domains (public agencies, banks, employers, insurers, and platforms), and researchers and civil society groups that investigate discrimination. For decision sciences practitioners, the report provides a structured map from machine-learning design choices to legal risk, reinforcing that fairness is not only a technical property but also a governance and accountability requirement.

Overall, the core contribution is a synthesized, mechanism-based account of AI discrimination risks tied to the practical reach and limits of non-discrimination and data protection law, culminating in recommendations for enforcement, organizational governance, and sector-specific regulatory evolution.

Cornell Notes

The report synthesizes how AI systems can produce discriminatory effects through specific machine-learning mechanisms (target labels, training data, feature selection, proxies, and intentional misuse). It then evaluates how European non-discrimination law and data protection law—especially GDPR rules on automated decisions and DPIAs—can mitigate these risks, while arguing that gaps remain and sector-specific regulation and stronger enforcement are needed.

What is the report’s main research question and why does it matter?

It asks where algorithmic decision-making and AI create discriminatory effects, what safeguards currently exist, and what recommendations should follow for organizations and equality/human-rights bodies. It matters because AI is increasingly used for high-stakes decisions in both public and private sectors, and discrimination can be unintentionally learned from biased data or amplified through opaque systems.

What study design or methodology does the report use?

It is a literature review and conceptual legal analysis (a “quick scan”), synthesizing prior scholarship and illustrative cases rather than conducting new experiments or collecting original datasets.

How does the report explain the mechanisms by which AI can lead to discrimination?

Following Barocas and Selbst, it identifies six pathways: (1) defining target variables/class labels, (2) biased training data, (3) biased data collection and feedback loops, (4) biased feature selection, (5) proxy effects/redundant encodings, and (6) intentional discrimination using proxies.

What is the report’s key example from criminal justice, and what quantitative claims does it cite?

It discusses COMPAS, citing ProPublica’s summary that COMPAS correctly predicts recidivism 61% of the time, while Black defendants were almost twice as likely as whites to be labelled higher risk but not actually reoffend, and twice as likely to be misclassified as higher risk of violent recidivism.

What does the report say about fairness trade-offs in risk scoring?

It notes that calibration and error-rate equality can conflict when base rates differ, implying that some fairness criteria may be mathematically incompatible (so discrimination prevention must choose among standards).

How does the report connect AI discrimination to advertising and pricing?

It argues that targeted ad delivery and online price differentiation can exclude or overcharge groups even without explicit protected-attribute inputs, citing examples such as race-associated ad delivery differences and Princeton Review’s finding that customers in high Asian-density areas were 1.8 times as likely to be offered higher prices.

Which legal regimes does the report identify as most relevant safeguards?

Non-discrimination law and data protection law. It emphasizes that indirect discrimination analysis focuses on effects rather than intent, while data protection law adds transparency, purpose limitation, accountability, DPIAs, and GDPR rules on automated decisions.

What are the main limitations of relying on current law?

Enforcement and compliance deficits, evidentiary difficulty due to AI opacity, gaps where AI differentiates using invented categories not tied to protected characteristics, and partial coverage limits in data protection law (e.g., models not linked to identifiable persons).

What recommendations does the report make for organizations and regulators?

Organizations should educate staff, conduct multidisciplinary risk assessments, document and monitor systems, and use impact assessments; public bodies should prioritize transparency/auditability and consider sunset clauses. Equality bodies should build AI technical expertise, consult early in procurement, cooperate with data protection authorities, and support research and strategic enforcement.

Review Questions

  1. Which of the six discrimination pathways is most directly implicated by biased feature selection, and how would you detect it in practice?

  2. Why does the report argue that indirect discrimination law can be hard to apply to AI systems, even when discrimination is present?

  3. How do GDPR automated decision rules (Article 22) and DPIAs function as safeguards, and what limitations does the report highlight?

  4. What does the report mean by “new types of unfair differentiation” escaping protected-characteristic statutes, and what kind of regulatory response does it suggest?

  5. Compare the report’s treatment of transparency/auditability in the public sector versus the private sector—what governance mechanisms does it recommend?

Key Points

  1. 1

    The report maps AI discrimination risks to six machine-learning mechanisms: target/class label choices, biased training data, biased data collection and feedback loops, feature selection, proxy effects/redundant encodings, and intentional discrimination.

  2. 2

    It argues that AI discrimination often arises unintentionally because systems learn from biased human decisions and because AI outputs are opaque, making discrimination harder to detect and contest.

  3. 3

    Non-discrimination law (direct and indirect discrimination) can address many AI harms, but enforcement is difficult due to open-ended legal standards, evidentiary burdens, and the “black box” nature of AI decisions.

  4. 4

    Data protection law—especially GDPR transparency duties, DPIAs, and Article 22’s limits on certain fully automated decisions—can mitigate discrimination risks, but has coverage gaps (e.g., when models are not tied to identifiable persons) and faces enforcement/compliance deficits.

  5. 5

    The report highlights concrete examples across domains: COMPAS risk assessment (with cited 61% overall accuracy and disparate error patterns), biased AI hiring, discriminatory ad targeting, online price discrimination (1.8× higher pricing likelihood in high Asian-density areas), face recognition error disparities (up to 34.7% for darker-skinned females), and gender bias in translation tools.

  6. 6

    A central regulatory claim is that current law may not cover AI-driven differentiation based on invented categories that do not correlate with protected characteristics, even if outcomes reinforce inequality.

  7. 7

    The report recommends sector-specific regulation and stronger enforcement, plus organizational governance practices such as multidisciplinary risk assessment, monitoring, documentation, and impact assessments inspired by GDPR DPIAs.

Highlights

“By exposing so-called ‘machine learning’ algorithms to examples of the cases of interest… the algorithm ‘learns’ which related attributes… can serve as potential proxies for those qualities or outcomes of interest.” (Barocas and Selbst, as quoted in the report)
COMPAS example: COMPAS “correctly predicts recidivism 61 percent of the time,” but “blacks are almost twice as likely as whites to be labelled a higher risk but not actually reoffend.”
Face recognition disparity: “darker-skinned females are the most misclassified group (with error rates of up to 34.7%). The maximum error rate for lighter-skinned males is 0.8%.”
Price discrimination: “Customers in areas with a high density of Asian residents were 1.8 times as likely to be offered higher prices, regardless of income.”
Core governance recommendation: “Risk assessment and risk mitigation” should involve “multiple disciplines… define the risks… record… monitor… and often report outward in some way.”

Topics

  • Algorithmic fairness
  • Discrimination law
  • Indirect discrimination / disparate impact
  • Machine learning bias
  • Proxy discrimination and redundant encodings
  • Automated decision-making governance
  • Data protection and privacy law
  • GDPR automated decision rules
  • Impact assessments (DPIA)
  • Regulatory enforcement and auditing
  • Sector-specific AI regulation
  • Human rights and AI

Mentioned

  • GDPR (General Data Protection Regulation)
  • COMPAS
  • Facebook
  • Google
  • Amazon
  • Princeton Review
  • Google Translate
  • Street Bump
  • FATML (Fairness, Accountability, and Transparency in Machine Learning)
  • IEEE (Global Initiative on Ethics of Autonomous and Intelligent Systems)
  • Partnership on AI to Benefit People and Society
  • Frederik Zuiderveen Borgesius
  • Solon Barocas
  • Solon Selbst
  • Angwin (ProPublica)
  • Michael Veale
  • Joanna Bryson
  • Cynthia Dwork
  • C. Caliskan
  • J. Buolamwini
  • T. Gebru
  • Latanya Sweeney
  • A. Datta
  • M. Tschantz
  • A. Narayanan
  • Kroll (et al.)
  • Lum and Isaac
  • Chouldechova
  • Wachter
  • Mittelstadt
  • Russell
  • Parasuraman
  • Manzey
  • Dieterich
  • Mendoza
  • Brennan
  • Dastin (Reuters)
  • Valentino-Devries (et al.)
  • AI - Artificial Intelligence
  • GDPR - General Data Protection Regulation
  • DPIA - Data Protection Impact Assessment
  • COE - Council of Europe
  • ECHR - European Convention on Human Rights
  • ECtHR - European Court of Human Rights
  • TETs - Transparency-Enhancing Technologies
  • FATML - Fairness, Accountability, and Transparency in Machine Learning
  • IP - Intellectual Property
  • COMPAS - Correctional Offender Management Profiling for Alternative Sanctions
  • RCT - Randomized Controlled Trial (not used in this report)