TY - GEN
T1 - Understanding the Affordances and Constraints of Explainable AI in Safety-Critical Contexts
T2 - 17th IFIP WG 8.5 International Conference on Electronic Participation, ePart 2025
AU - Buszydlik, Aleksander
AU - Altmeyer, Patrick
AU - Dobbe, Roel
AU - Liem, Cynthia C.S.
N1 - Green Open Access added to TU Delft Institutional Repository as part of the Taverne amendment. More information about this copyright law amendment can be found at https://www.openaccess.nl. Otherwise as indicated in the copyright section: the publisher is the copyright holder of this work and the author uses the Dutch legislation to make this work public.
PY - 2026
Y1 - 2026
N2 - We focus on explainability as a desideratum for automated decision-making systems, rather than only models. Although the explainable artificial intelligence (XAI) paradigm offers an impressive variety of solutions to increase the transparency of automated decisions, XAI contributions rarely account for the complete systems—social and institutional environments—where models operate. Our work focuses on one such system in the domain of social welfare, which increasingly turns to automated decision-making to carry out targeted digital surveillance. Specifically, we present a case study of a black-box machine learning model previously used in a major Dutch city to support its officials in the task of detecting fraud. Employing analyses established in the field of system safety, we identify five types of hazards that could have occurred after the introduction of the model. For each of them, we reason about the potential value of XAI interventions as hazard mitigation strategies. The case study illustrates how the deployment of models may impact processes that exist far upstream and downstream from their decision logic, making explainability and/or interpretability insufficient to guarantee the systems’ safe operation. In many cases, XAI techniques may only be able to reasonably address a small fraction of hazards related to the use of algorithms; several major hazards that we identify would have still posed risks if the system had relied on an interpretable model. Thus, we empirically demonstrate that the values, which lie at the heart of XAI research, such as responsibility, safety, or transparency, ultimately necessitate a broader outlook on automated decision-making systems.
AB - We focus on explainability as a desideratum for automated decision-making systems, rather than only models. Although the explainable artificial intelligence (XAI) paradigm offers an impressive variety of solutions to increase the transparency of automated decisions, XAI contributions rarely account for the complete systems—social and institutional environments—where models operate. Our work focuses on one such system in the domain of social welfare, which increasingly turns to automated decision-making to carry out targeted digital surveillance. Specifically, we present a case study of a black-box machine learning model previously used in a major Dutch city to support its officials in the task of detecting fraud. Employing analyses established in the field of system safety, we identify five types of hazards that could have occurred after the introduction of the model. For each of them, we reason about the potential value of XAI interventions as hazard mitigation strategies. The case study illustrates how the deployment of models may impact processes that exist far upstream and downstream from their decision logic, making explainability and/or interpretability insufficient to guarantee the systems’ safe operation. In many cases, XAI techniques may only be able to reasonably address a small fraction of hazards related to the use of algorithms; several major hazards that we identify would have still posed risks if the system had relied on an interpretable model. Thus, we empirically demonstrate that the values, which lie at the heart of XAI research, such as responsibility, safety, or transparency, ultimately necessitate a broader outlook on automated decision-making systems.
KW - Automated decision-making
KW - Explainable artificial intelligence
KW - Social welfare
KW - System safety
KW - Technology audits
UR - http://www.scopus.com/inward/record.url?scp=105014495080&partnerID=8YFLogxK
U2 - 10.1007/978-3-032-02515-9_8
DO - 10.1007/978-3-032-02515-9_8
M3 - Conference contribution
AN - SCOPUS:105014495080
SN - 9783032025142
T3 - Lecture Notes in Computer Science
SP - 118
EP - 136
BT - Electronic Participation - 17th IFIP WG 8.5 International Conference, ePart 2025, Proceedings
A2 - Hofmann, Sara
A2 - Danneels, Lieselot
A2 - Dobbe, Roel
A2 - Ubacht, Jolien
A2 - Novak, Anna-Sophie
A2 - Parycek, Peter
A2 - Schwabe, Gerhard
A2 - Spitzer, Vera
PB - Springer
Y2 - 31 August 2025 through 4 September 2025
ER -