TY - JOUR
T1 - Evaluating XAI
T2 - A comparison of rule-based and example-based explanations
AU - van der Waa, Jasper
AU - Nieuwburg, Elisabeth
AU - Cremers, Anita
AU - Neerincx, Mark
PY - 2021
Y1 - 2021
N2 - Current developments in Artificial Intelligence (AI) led to a resurgence of Explainable AI (XAI). New methods are being researched to obtain information from AI systems in order to generate explanations for their output. However, there is an overall lack of valid and reliable evaluations of the effects on users' experience of, and behavior in response to explanations. New XAI methods are often based on an intuitive notion what an effective explanation should be. Rule- and example-based contrastive explanations are two exemplary explanation styles. In this study we evaluate the effects of these two explanation styles on system understanding, persuasive power and task performance in the context of decision support in diabetes self-management. Furthermore, we provide three sets of recommendations based on our experience designing this evaluation to help improve future evaluations. Our results show that rule-based explanations have a small positive effect on system understanding, whereas both rule- and example-based explanations seem to persuade users in following the advice even when incorrect. Neither explanation improves task performance compared to no explanation. This can be explained by the fact that both explanation styles only provide details relevant for a single decision, not the underlying rational or causality. These results show the importance of user evaluations in assessing the current assumptions and intuitions on effective explanations.
AB - Current developments in Artificial Intelligence (AI) led to a resurgence of Explainable AI (XAI). New methods are being researched to obtain information from AI systems in order to generate explanations for their output. However, there is an overall lack of valid and reliable evaluations of the effects on users' experience of, and behavior in response to explanations. New XAI methods are often based on an intuitive notion what an effective explanation should be. Rule- and example-based contrastive explanations are two exemplary explanation styles. In this study we evaluate the effects of these two explanation styles on system understanding, persuasive power and task performance in the context of decision support in diabetes self-management. Furthermore, we provide three sets of recommendations based on our experience designing this evaluation to help improve future evaluations. Our results show that rule-based explanations have a small positive effect on system understanding, whereas both rule- and example-based explanations seem to persuade users in following the advice even when incorrect. Neither explanation improves task performance compared to no explanation. This can be explained by the fact that both explanation styles only provide details relevant for a single decision, not the underlying rational or causality. These results show the importance of user evaluations in assessing the current assumptions and intuitions on effective explanations.
KW - Artificial Intelligence (AI)
KW - Contrastive explanations
KW - Decision support systems
KW - Explainable Artificial Intelligence (XAI)
KW - Machine learning
KW - User evaluations
UR - http://www.scopus.com/inward/record.url?scp=85097186283&partnerID=8YFLogxK
U2 - 10.1016/j.artint.2020.103404
DO - 10.1016/j.artint.2020.103404
M3 - Article
AN - SCOPUS:85097186283
VL - 291
JO - Artificial Intelligence
JF - Artificial Intelligence
SN - 0004-3702
M1 - 103404
ER -