Evaluating XAI: A comparison of rule-based and example-based explanations
作者:
摘要
Current developments in Artificial Intelligence (AI) led to a resurgence of Explainable AI (XAI). New methods are being researched to obtain information from AI systems in order to generate explanations for their output. However, there is an overall lack of valid and reliable evaluations of the effects on users' experience of, and behavior in response to explanations. New XAI methods are often based on an intuitive notion what an effective explanation should be. Rule- and example-based contrastive explanations are two exemplary explanation styles. In this study we evaluate the effects of these two explanation styles on system understanding, persuasive power and task performance in the context of decision support in diabetes self-management. Furthermore, we provide three sets of recommendations based on our experience designing this evaluation to help improve future evaluations. Our results show that rule-based explanations have a small positive effect on system understanding, whereas both rule- and example-based explanations seem to persuade users in following the advice even when incorrect. Neither explanation improves task performance compared to no explanation. This can be explained by the fact that both explanation styles only provide details relevant for a single decision, not the underlying rational or causality. These results show the importance of user evaluations in assessing the current assumptions and intuitions on effective explanations.
论文关键词:Explainable Artificial Intelligence (XAI),User evaluations,Contrastive explanations,Artificial Intelligence (AI),Machine learning,Decision support systems
论文评审过程:Received 21 February 2020, Revised 20 August 2020, Accepted 26 October 2020, Available online 28 October 2020, Version of Record 3 December 2020.
论文官网地址:https://doi.org/10.1016/j.artint.2020.103404