The COMPAS Algorithm Scandal: When AI Decides Who Goes to Jail ⚖️

As AI enters courts and welfare systems worldwide, the COMPAS debate reveals a critical lesson: fairness depends on context, and exporting models without reform risks scaling inequality.

The COMPAS Algorithm Scandal: When AI Decides Who Goes to Jail ⚖️

🎥 Explained: The COMPAS Algorithm Scandal: When AI Decides Who Goes to Jail ⚖️**


🧭 A Deep Dive into the COMPAS Fairness Dilemma

⚖️ The Algorithm’s Forked Tongue

🤖 When algorithms speak with two voices, justice becomes a matter of mathematics.


🧩 1. What Happened: A Tale of Two Truths

For decades, the U.S. criminal justice system has moved toward 📊 evidence-based sentencing.

The goal: reduce inconsistent judicial discretion.
The tool: actuarial risk models.
The promise: ⚙️ objectivity and scientific rigor.

That promise fractured in 2016 with controversy surrounding COMPAS — the Correctional Offender Management Profiling for Alternative Sanctions system.

🥊 The Core Conflict

Two institutions. Two truths.

  • 📰 ProPublica → investigative nonprofit
  • 🏢 Northpointe (now equivant) → COMPAS developer

ProPublica analyzed ~10,000 defendants in Broward County, Florida.
Later benchmarks using AIF360 standardized the dataset to 6,167 rows (5,723 for Black and White defendants).


📌 Controversy Cheat Sheet

🚨 The Accusation (ProPublica)

The algorithm is racially biased.

  • Black defendants who did not reoffend were nearly 2× more likely to be labeled High Risk.
  • White defendants who did reoffend were more likely to be labeled Low Risk.

🛡️ The Defense (Northpointe)

The algorithm is fair.

  • It maintains predictive parity.
  • A risk score has the same statistical meaning across races.

🧮 The Mathematical Deadlock

At the heart lies the Fairness Impossibility Theorem:

❗ No model can satisfy both equal error rates and predictive parity unless base rates are equal or accuracy is perfect.

Because arrest and recidivism rates differ across groups—driven by systemic factors like over-policing—both sides are technically correct.

⚠️ The result: a mathematical conflict with human consequences.


🔗 2. Impact: When Statistics Become Sentences

Choosing a fairness metric is not academic.
It defines who loses freedom.

When COMPAS labels someone High Risk, it influences:

  • 💰 Higher bond
  • 🚫 Denied parole
  • ⛓️ Longer incarceration

📉 The Human Cost of Misclassification

MetricBlack DefendantsWhite Defendants
❌ False Positive (High Risk, no reoffense)45%23%
⚠️ False Negative (Low Risk, reoffended)28%47%

🧠 Why This Matters

  • 45% false positives → nearly half of law-abiding Black defendants lost liberty.
  • 47% false negatives → the system extended more benefit of the doubt to White defendants.

👉 To understand why, we must examine the data feeding the machine.


☠️ 3. Lifecycle Failure: The Poisoned Well

COMPAS did not fail at prediction.
It failed at data design.

If training data reflects historical oppression, the algorithm becomes a ⚡ high-speed engine for reinforcing it.

🧬 The Proxy Variable Problem

COMPAS avoids race directly but uses proxies:

  • 📍 Zip Codes → segregated, heavily policed neighborhoods
  • 👥 Family/Friend Victim History → systemic inequality exposure
  • 🚨 Gang Affiliation → subjective urban profiling
  • 🏚️ Socialization History → housing instability, family disruption

These encode structural inequality while appearing neutral.


⚖️ Due Process vs Trade Secrets

In State v. Loomis, the Wisconsin Supreme Court allowed COMPAS use with disclaimers.

🧾 The problem:
Defendants cannot challenge a proprietary algorithm used to sentence them.

Trade secrets > due process.


🔧 Attempts at Reform

equivant’s COMPAS-R Core introduced:

  • 🧼 Neutral language revisions
  • ❌ Removal of ambiguous responses
  • 🧪 Experimental bias-testing questions

These improve optics, not structural fairness.


⚔️ 4. Bias Types: A Clash of Definitions

Fairness is not one thing.

🎯 Three Competing Definitions

1️⃣ Equalized Odds
Equal error rates across groups.

2️⃣ Predictive Parity
Risk scores mean the same across groups.

3️⃣ Historical Bias
Ground truth reflects systemic inequality.

⚖️ The Trade-off

  • 🧑‍🤝‍🧑 Group fairness → correct systemic disadvantage
  • 👤 Individual fairness → treat similar individuals consistently

Improving one often worsens the other.

👉 Justice becomes a statistical choice.


🌍 5. Global South Lens: Exporting Inequality

COMPAS is not just a U.S. story.

Western models are exported as “modernization tools” to the Global South.

But variables like:

  • 🏠 Residential stability
  • 💼 Vocational history

are not neutral where:

  • Informal economies dominate
  • Conflict causes displacement
  • Colonial policies shaped land access

⚠️ Exporting such models risks technological solutionism — fixing justice with code while ignoring poisoned data.


🧠 6. Bigger Picture: Can We Code Our Way to Justice?

Technical fixes exist. None solve the root problem.

🛠️ Mitigation Strategies

StrategyEffectivenessLimitation
⚖️ ReweighingImproves fairness metricsCannot fix biased data
🧮 Prejudice RemoverEnforces independence from race~38% accuracy
🔄 Calibrated Equalized OddsBalances errorsMisses real recidivism

🧭 Radical Proposal: Affirmative Algorithms

Use race as a corrective factor to counter systemic bias.

This acknowledges reality — and raises democratic questions.


❓ Curiosity Provocations

  • 🎲 If the model is barely better than a coin flip, why use it for prison time?
  • 🔒 Should trade secrets override a defendant’s right to examine evidence?
  • 🗳️ If fairness definitions conflict, who decides which one governs justice?

🧵 Final Reflection: The Human System Behind the Machine

Algorithms reflect the systems that produce their data:

  • 🚔 Over-policing
  • 📉 Structural inequality
  • 🏛️ Historical discrimination

Technical fixes treat symptoms.

👤 Human accountability must remain the final check.

⚠️ A black box cannot be the final word on human freedom.

COMPAS recidivism fairness incident


📚 References

🔎 Investigative Journalism & Public Analyses

  • Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016).
    Machine Bias: There’s Software Used Across the Country to Predict Future Criminals — And It’s Biased Against Blacks.
    ProPublica.

  • Chawla, M. (2022).
    COMPAS Case Study: Investigating Algorithmic Fairness of Predictive Policing.
    Medium.

  • Deb, E. (2023).
    COMPAS — an AI Tool Sending or Keeping People in Jail.
    Medium.

  • Murray, K. (2025).
    What the Legal Drama For the People Teaches Us About AI and Legal Ethics.

  • Wikipedia. (2025).
    COMPAS (software).


  • State v. Loomis. (2016). 881 N.W.2d 749 (Wis. 2016).

  • U.S. Congress. (2025).
    S.2164 — Algorithmic Accountability Act of 2025. 119th Congress.

  • New Jersey Judiciary. (2025).
    Criminal Justice Reform — Myth v. Fact.

  • Tait, E. J., Linas, J. M., Bergstrom, R., Kukkonen, C. A., & Schulman, E. R. (2019).
    Proposed Algorithmic Accountability Act Targets Bias in Artificial Intelligence.
    Jones Day.

  • Stevenson, M. T., & Doleac, J. L. (2018).
    Algorithmic Risk Assessment in the Hands of Humans.
    American Constitution Society.

  • Stensrud, A. (2025).
    The COMPAS Case’s Impact on the EU’s AI Act.
    Lov&Data.


🧮 Fairness Theory & Technical Research

  • Acharya, A., Caravela, D., Kim, E., Kornberg, E., & Nesmith, E. (2022).
    Does the COMPAS Needle Always Point Towards Equity? Finding Fairness in the COMPAS Risk Assessment Algorithm: A Case Study.
    CAUSEweb.

  • Ejike, U. (2026).
    The Fairness–Accuracy Frontier: Impossibility Theorems and Optimal Tradeoffs in Algorithmic Decision-Making.
    WorldQuant University.

  • Hellman, D. (2020).
    Measuring Algorithmic Fairness.
    Virginia Law Review, 106(4).

  • Hsu, B., Mazumder, R., Nandy, P., & Basu, K. (2022).
    Pushing the Limits of Fairness Impossibility: Who’s the Fairest of Them All?
    NeurIPS.

  • Schmid, F. (2022).
    Understanding the Importance of Algorithmic Fairness.
    Gen Re.

  • Kartha, N., & Young, W. D. (n.d.).
    An Overview of Algorithmic Bias in Artificial Intelligence.
    University of Texas at Austin.


🧠 Ethics, Due Process & Sociotechnical Perspectives

  • Humerick, J. (2020).
    Reprogramming Fairness: Affirmative Action in Algorithmic Criminal Sentencing.
    Columbia Human Rights Law Review.

  • Israni, E. (2017).
    Algorithmic Due Process: Mistaken Accountability and Attribution in State v. Loomis.
    JOLT Digest.

  • Rev. (n.d.).
    AI Sentencing Ethics: Balancing Justice and Innovation.

  • Vaccaro, M. A. (2019).
    Algorithms in Human Decision-Making: A Case Study With the COMPAS Risk Assessment Software.
    Harvard University DASH Repository.


🏢 Vendor & System Documentation

  • equivant Supervision. (2024).
    Debunking Misconceptions About the COMPAS Core Instrument: What You Need to Know.

  • equivant Supervision. (2023).
    How Do the Scales in the COMPAS-R Core Differ From Those in the Standard COMPAS Core?

  • equivant Supervision. (2023).
    Why Was the COMPAS-R Core Created and How Does It Differ From the Standard COMPAS Core?


🧾 Critical Responses & Methodological Debates

  • Flores, A. W., Lowenkamp, C. T., & Bechtel, K. (2016).
    False Positives, False Negatives, and False Analyses: A Rejoinder to “Machine Bias.”
    Federal Probation Journal, 80(2).

  • Fossett, J. (2020).
    Response to “How We Analyzed the COMPAS Recidivism Algorithm” (Larson et al.).


📥 AI Fairness 101 — Real-World Incidents: The COMPAS Algorithm Scandal Case Deck (PDF)

👉 Download the The COMPAS Algorithm Scandal Case Deck (PDF)

🔎 Explore the AI Fairness 101 Series

This post is part of the AI Fairness 101 — Real-World Incidents learning track.

Stay tuned — new posts every week!

💬 Join the Conversation

Have thoughts, experiences, or questions about AI fairness? Share your comments, discuss with global experts, and connect with the community:

👉 Reach out via the Contact page
📧 Write to us: [email protected]


🌍 Follow GlobalSouth.AI

Stay connected and join the conversation on AI governance, fairness, safety, and sustainability.

Subscribe to stay updated on new case studies, frameworks, and Global South perspectives on responsible AI.

Related Posts