“If it didn’t happen, why would I change my decision?”: How Judges Respond to Counterfactual Explanations for the Public Safety Assessment

Type: Article

Publication Date: 2022-10-14

Citations: 3

DOI: https://doi.org/10.1609/hcomp.v10i1.22001

Abstract

Many researchers and policymakers have expressed excitement about algorithmic explanations enabling more fair and responsible decision-making. However, recent experimental studies have found that explanations do not always improve human use of algorithmic advice. In this study, we shed light on how people interpret and respond to counterfactual explanations (CFEs)---explanations that show how a model's output would change with marginal changes to its input(s)---in the context of pretrial risk assessment instruments (PRAIs). We ran think-aloud trials with eight sitting U.S. state court judges, providing them with recommendations from a PRAI that includes CFEs. We found that the CFEs did not alter the judges' decisions. At first, judges misinterpreted the counterfactuals as real---rather than hypothetical---changes to defendants. Once judges understood what the counterfactuals meant, they ignored them, stating their role is only to make decisions regarding the actual defendant in question. The judges also expressed a mix of reasons for ignoring or following the advice of the PRAI without CFEs. These results add to the literature detailing the unexpected ways in which people respond to algorithms and explanations. They also highlight new challenges associated with improving human-algorithm collaborations through explanations.

Locations

  • arXiv (Cornell University) - View - PDF
  • Proceedings of the AAAI Conference on Human Computation and Crowdsourcing - View - PDF

Similar Works

Action Title Year Authors
+ "If it didn't happen, why would I change my decision?": How Judges Respond to Counterfactual Explanations for the Public Safety Assessment 2022 Yaniv Yacoby
Ben Green
Christopher L. Griffin
Finale Doshi Velez
+ PDF Chat “If it Didn't Happen, Why Would I Change My Decision?”: How Judges Respond to Counterfactual Explanations for the Public Safety Assessment 2022 Ben Green
Yaniv Yacoby
Christopher L. Griffin
Finale Doshi‐Velez
+ PDF Chat Experimental evaluation of algorithm-assisted human decision-making: application to pretrial public safety assessment* 2023 Kosuke Imai
Zhichao Jiang
D. James Greiner
Ryan Halen
Sooahn Shin
+ Counterfactual risk assessments, evaluation, and fairness 2020 Amanda Coston
Alan Mishler
Edward H. Kennedy
Alexandra Chouldechova
+ Can counterfactual explanations of AI systems’ predictions skew lay users’ causal intuitions about the world? If so, can we correct for that? 2022 Marko Tešić
Ulrike Hahn
+ PDF Chat Understanding the Effect of Counterfactual Explanations on Trust and Reliance on AI for Human-AI Collaborative Clinical Decision Making 2023 Min Hun Lee
Chong Jun Chew
+ On the Impact of Explanations on Understanding of Algorithmic Decision-Making 2023 Timothée Schmude
Laura Koesten
Torsten Möller
Sebastian Tschiatschek
+ Experimental Evaluation of Algorithm-Assisted Human Decision-Making: Application to Pretrial Public Safety Assessment 2020 Kosuke Imai
Zhichao Jiang
James Greiner
Ryan Halen
Sooahn Shin
+ PDF Chat On the Impact of Explanations on Understanding of Algorithmic Decision-Making 2023 Timothée Schmude
Laura Koesten
Torsten Möller
Sebastian Tschiatschek
+ PDF Chat Counterfactual Explanations and Algorithmic Recourses for Machine Learning: A Review 2024 Sahil Verma
Varich Boonsanong
Minh Hoang
Keegan Hines
John P. Dickerson
Chirag Shah
+ Counterfactual Explanations and Algorithmic Recourses for Machine Learning: A Review 2020 Sahil Verma
Varich Boonsanong
Minh Hoang
Keegan Hines
John P. Dickerson
Chirag Shah
+ PDF Chat “There Is Not Enough Information”: On the Effects of Explanations on Perceptions of Informational Fairness and Trustworthiness in Automated Decision-Making 2022 Jakob Schoeffer
Niklas Kuehl
Yvette Machowski
+ For Better or Worse: The Impact of Counterfactual Explanations' Directionality on User Behavior in xAI 2023 Ulrike Kuhl
André Artelt
Barbara Hammer
+ Counterfactual Explanations for Machine Learning: A Review. 2020 Sahil Verma
John P. Dickerson
Keegan Hines
+ Can counterfactual explanations of AI systems' predictions skew lay users' causal intuitions about the world? If so, can we correct for that? 2022 Marko Tešić
Ulrike Hahn
+ Information That Matters: Exploring Information Needs of People Affected by Algorithmic Decisions 2024 Timothée Schmude
Laura Koesten
Torsten Möller
Sebastian Tschiatschek
+ PDF Chat Explainable and Human-Grounded AI for Decision Support Systems: The Theory of Epistemic Quasi-Partnerships 2024 John Dorsch
Maximilian Moll
+ PDF Chat Don't be Fooled: The Misinformation Effect of Explanations in Human-AI Collaboration 2024 Philipp Spitzer
Joshua Holstein
Katelyn Morrison
Kenneth Holstein
Gerhard Satzger
Niklas Kühl
+ Disagreement amongst counterfactual explanations: How transparency can be deceptive 2023 Dieter Brughmans
Lissa Melis
David Martens
+ PDF Chat Understanding the Role of Human Intuition on Reliance in Human-AI Decision-Making with Explanations 2023 Valerie Chen
Q. Vera Liao
Jennifer Wortman Vaughan
Gagan Bansal

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (15)

Action Title Year Authors
+ “Why Should I Trust You?”: Explaining the Predictions of Any Classifier 2016 Marco Ribeiro
Sameer Singh
Carlos Guestrin
+ Explainable AI: Beware of Inmates Running the Asylum Or: How I Learnt to Stop Worrying and Love the Social and Behavioural Sciences 2017 Tim Miller
Piers D. L. Howe
Liz Sonenberg
+ Explanation in Human-AI Systems: A Literature Meta-Review, Synopsis of Key Ideas and Publications, and Bibliography for Explainable AI 2019 Shane T. Mueller
Robert R. Hoffman
William J. Clancey
Abagail K. Emery
Gary Klein
+ PDF Chat A Survey of Methods for Explaining Black Box Models 2018 Riccardo Guidotti
Anna Monreale
Salvatore Ruggieri
Franco Turini
Fosca Giannotti
Dino Pedreschi
+ A Unified Approach to Interpreting Model Predictions 2017 Scott Lundberg
Su‐In Lee
+ PDF Chat Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR 2017 Sandra Wachter
Brent Mittelstadt
Chris Russell
+ Counterfactual Visual Explanations 2019 Yash Goyal
Ziyan Wu
Jan Ernst
Dhruv Batra
Devi Parikh
Stefan Lee
+ PDF Chat Auditing black-box models for indirect influence 2017 Philip Adler
Casey Falk
Sorelle A. Friedler
Tionney Nix
Gabriel Rybeck
Carlos Scheidegger
Brandon Smith
Suresh Venkatasubramanian
+ Counterfactual Explanations for Machine Learning: A Review. 2020 Sahil Verma
John P. Dickerson
Keegan Hines
+ Explaining models: an empirical study of how explanations impact fairness judgment 2019 Jonathan Dodge
Q. Vera Liao
Yunfeng Zhang
Rachel Bellamy
Casey Dugan