The Ethical Case for Algorithmic Adjudication

Sebastian Hietsch

AI Ethics | The Ethical Case for Algorithmic Adjudication

Literature

The Normative Question: Huq (2020) argues there is no absolute or inherent "right to a human decision" in the justice system.

The "Human-AI Fairness Gap": Chen et al. (2022) empirically demonstrate that the public initially perceives human judges as procedurally fairer than algorithms.

Achieving Procedural Justice: Lee et al. (2019) show that algorithmic fairness is accepted by the public when systems implement "algorithmic transparency" and "outcome control".

Huq, Aziz Z. "A right to a human decision." Va. L. Rev. 106 (2020): 611. | Chen, B.M., Stremitzer, A. and Tobia, K., 2022. Having your day in robot court. Harv. JL & Tech., 36, p.127. | Lee, M.K., Jain, A., Cha, H.J., Ojha, S. and Kusbit, D., 2019. Procedural justice in algorithmic fairness: Leveraging transparency and outcome control for fair algorithmic mediation. Proceedings of the ACM on human-computer interaction, 3(CSCW), pp.1-26.
AI Ethics | The Ethical Case for Algorithmic Adjudication

Ethical Thesis

If an AI can be statistically proven to be fairer, more consistent, and less biased than human judges, utilizing algorithmic adjudication is not only permissible but ethically obligatory.

AI Ethics | The Ethical Case for Algorithmic Adjudication

Eliminating Unwarranted Disparity

  • Human decision-making leads to unwanted variability in judgments that should be identical
  • Sentencing susceptible to extralegal factors and physiological inconsistencies

Research

  • Study: compare retired judges and LLMs on criminal cases
  • Human judges: large sentencing disparity and variance
  • LLMs: highly consistent sentences.
  • LLMs deviated less from the human mean than the individual human judges did.
Gazal Ayal, O., Elyoseph, Z. and Solomon, A., 2026. Evaluating Large Language Models as Judicial Decision-Makers. Justice Quarterly, pp.1-36.
AI Ethics | The Ethical Case for Algorithmic Adjudication

The Human "Black Box"

  • Human biases: fundamentally opaque and unauditable
  • Controlled experiments difficult

Algorithmic Auditability

  • Algorithms operate on explicitly written code and data
  • Algorithms can be rigorously tested, evaluated, and deliberately reprogrammed to ignore specific variables
Kleinberg, J., Ludwig, J., Mullainathan, S. and Sunstein, C.R., 2018. Discrimination in the Age of Algorithms. Journal of legal analysis, 10, pp.113-174.
AI Ethics | The Ethical Case for Algorithmic Adjudication

Algorithmic Offsetting

  • Human-AI Fairness Gap
    • Public preference for human judges driven by "hard factors"
  • Algorithmic Offsetting: eliminate fairness penalty by embedding procedural safeguards

Procedural Safeguards

  • Voice/Hearing: Opportunity to present their case
  • Interpretability: Transparent, easily understandable reasoning
  • Outcome Control: Ability to appeal, contest, or modify decisions
Chen, B.M., Stremitzer, A. and Tobia, K., 2022. Having your day in robot court. Harv. JL & Tech., 36, p.127.
AI Ethics | The Ethical Case for Algorithmic Adjudication

Democratizing Access to Justice

  • Traditional court proceedings: expensive, complex, and slow
  • AI Solution
    • reduce administrative burdens
    • lower costs
    • accelerate case resolution.
    • real-time translation
    • High-Volume, Low-Value Disputes: justice for disputes where litigation costs outweigh claim's value
Dolidze, T., 2026. Artificial Intelligence in Judicial Decision-Making: Can a Robot Replace a Judge?. Law and world, 12(37), pp.6-24.
AI Ethics | The Ethical Case for Algorithmic Adjudication

Conclusion

  • More consistent sentencing
  • Bias better understood and controlled
  • Fairness gap resolvable
  • Better access to justice
AI Ethics | The Ethical Case for Algorithmic Adjudication

Discussion Questions

  • If statistical audits prove that a "robot judge" is significantly more accurate and less biased than the average human judge, does demanding a human decision-maker become an unethical demand?

  • A classical trial satisfies the psychological need for an opportunity to speak and be heard. Does having a hearing before an AI satisfy this need for procedural dignity, or is human empathy strictly required for justice to feel fair?

Today, we are going to explore the intersection of artificial intelligence, ethics, and the law. In recent years, generative AI and Large Language Models have advanced significantly, moving from basic text processing to executing complex legal reasoning. We are now seeing the rapid development of Legal Judgment Prediction systems, which are designed to predict court outcomes, charges, and penalties directly from the factual descriptions of cases. This technological leap forces us to confront a profound ethical dilemma: the prospect of the 'robot judge'. Legal systems traditionally assume that justice requires a human arbiter. Indeed, many argue there is a fundamental 'right to a human decision'. However, our presentation will challenge this assumption using a consequentialist, specifically utilitarian, framework.

I will start with a brief summary of the most relevant literature. First, legal scholar Aziz Huq challenges traditional assumptions by arguing that there is actually no normative justification for an absolute 'right to a human decision' in the justice system. However, we must address the reality of public perception. Chen et al.'s experimental research identifies a 'Human-AI Fairness Gap,' showing that ordinary people initially view human decision-makers as procedurally fairer than algorithmic ones. But how do we close that gap? Research by Lee et al. provides the solution: they demonstrate that we can achieve procedural justice in algorithmic systems by integrating strict algorithmic transparency and 'outcome control' - giving litigants the ability to understand and appeal decisions.

I will be examining this thesis through a lens of consequentialism.

To begin our defense of algorithmic adjudication, we must first confront a harsh reality about human judges: they are notoriously inconsistent. While we frequently worry about algorithms being biased, human adjudication is severely plagued by unwanted and arbitrary variability in judgments that should theoretically be identical. In the criminal justice system, this results in the 'lawlessness' of sentencing, where similarly situated defendants receive vastly different punishments based on which judge they happen to draw. Furthermore, human performance in decision-making is affected by mundane, extralegal factors such as the time of day, fatigue, or the notorious 'hungry judge' effect, where leniency can vary simply based on when the judge last ate. To prove that AI can ethically outperform humans in this regard, we can look to a recent empirical pilot study conducted by researchers at the University of Haifa in Israel. They presented 123 retired judges with the exact same fictional criminal cases and asked them to determine a sentence. The results were staggering but unsurprising: the human judges exhibited massive sentencing disparity and variance. The researchers then ran the exact same cases through Large Language Models, including GPT, Gemini, and Claude. The LLMs exhibited significantly lower sentence disparity than the human judges, producing highly consistent outcomes. But consistency alone isn't enough; the decisions must also be accurate. Because there is no objective 'correct' sentence in a fictional case, the researchers used the collective average sentence of the 123 human judges as a conservative benchmark for the most 'accurate' or 'fair' outcome. Remarkably, the study found that all of the LLMs deviated less from the judges' own collective mean than the individual human judges did. What does this mean for our ethical framework? A core principle of justice is treating similar cases equally. If statistical audits prove that a machine is not only more consistent but actually clusters closer to the fair human consensus than individual humans do, then demanding a flawed, human decision-maker over an accurate machine becomes morally unjustifiable.

A major objection you will inevitably hear when proposing 'robot judges' is the fear of algorithmic bias. People worry that an AI will act as a 'black box' that perpetuates systemic discrimination. However, to defend our consequentialist thesis, we must turn this argument around: the human mind is the ultimate, impenetrable black box. Human biases are largely hidden, and we lack the tools to definitively root them out. You cannot run a controlled experiment on a human judge. You cannot ask a judge to decide the exact same case today with a white defendant, and tomorrow with a Black defendant, and reliably measure their prejudice without them catching on. AI systems, by contrast, are fundamentally auditable. While the implicit biases of human decision-makers are incredibly difficult to find and root out, we can actually 'peer into the brain of an algorithm'. Because AI operates on data and code, we can subject it to rigorous sensitivity testing and explicitly reprogram the system to ignore protected variables. If our ethical goal is to maximize justice and minimize harm, we must choose the system that can be fixed. Algorithms offer us 'far greater' visibility into the motivations of decisions, and therefore a far greater opportunity to remove discrimination. Therefore, rejecting an auditable, fixable AI in favor of an un-auditable, inherently biased human judge is not a defense of justice - it is a tolerance of hidden, systemic inequality.

Even if we prove that AI is more accurate and less biased, critics will inevitably raise the objection that defendants have a fundamental right to their 'day in court' - a process that affords them human empathy and dignity. However, we must critically examine whether 'empathy' is sometimes just a polite word for arbitrary inconsistency, as human responsiveness varies wildly from judge to judge. It is true that ordinary citizens initially exhibit a 'human-AI fairness gap,' perceiving human-led proceedings as fairer than AI-led ones. But a study conducted at Harvard University reveals a fascinating truth about this gap. The preference for humans is driven primarily by what researchers call 'hard factors' - concerns over the accuracy of the outcome and the thoroughness of the machine's consideration. It is driven much less by 'soft factors,' such as whether the decision-maker possesses human empathy or understands the litigant's feelings. Because the fairness gap is based on performance rather than an irreducible need for human emotion, it can be mitigated, or even eliminated, through a concept called 'algorithmic offsetting'. We can offset the AI's perceived fairness penalty by intentionally designing the system with robust procedural justice mechanisms. First, we must give litigants a 'voice.' Research shows that simply affording a litigant a hearing to present their case to an AI judge significantly increases their perception of procedural fairness. Second, the AI's decisions must feature high 'interpretability,' meaning the algorithm provides transparent and easily understandable reasoning for its outcome. Finally, we must implement 'outcome control' - the right to appeal or modify an algorithmic decision if it is wrong. Outcome control universally improves perceived fairness because it offers a safety net against algorithmic errors. Ultimately, if a 'robot court' offers an interpretable decision, a chance to be heard, and the right to appeal, the public might view it as being just as fair as a human judge. Therefore, a human is not a strict prerequisite for public acceptance of court decisions.

For our final argument, we must zoom out from the individual courtroom and look at the justice system as a whole. We have established that AI can be consistent, auditable, and procedurally fair. But none of these benefits matter if a citizen cannot afford to get into the courtroom in the first place. The harsh reality is that our traditional justice system is frequently too expensive, complex, and slow for ordinary people. Individuals facing language barriers, geographic distance, or a lack of financial resources are routinely priced out of legal representation and traditional litigation. This is where algorithmic adjudication becomes a moral imperative. AI tools can drastically reduce administrative burdens and lower the costs of dispute resolution. For instance, AI-powered systems can provide real-time translation for litigants and offer accessible guidance for self-represented individuals, serving as an alternative for low-income populations who cannot afford human legal services. Crucially, AI allows us to automate justice for high-volume, low-value disputes - such as small claims or consumer conflicts - where traditional court processes are simply too cumbersome relative to the amount at stake. If we apply our consequentialist ethical framework here, the conclusion is clear. Demanding a human judge for every single case, no matter how small, is not a defense of justice; it is a barrier to it. If providing a highly accurate, automated AI judge gives marginalized individuals a pathway to resolve disputes that they would otherwise abandon due to cost, then deploying that technology is not just permissible - it is an ethical obligation to democratize access to justice.

- "Would you rather have a biased human who looks you in the eye, or a perfectly fair machine?" - Does speaking to a machine rob us of our dignity, even if the machine gives a fair verdict?