Abstract
One aspect of a responsible application of Artificial Intelligence (AI) is ensuring that the operation and outputs of an AI system are understandable for non-technical users, who need to consider its recommendations in their decision making. The importance of explainable AI (XAI) is widely acknowledged; however, its practical implementation is not straightforward. In particular, it is still unclear what the requirements are of non-technical users from explanations, i.e. what makes an explanation meaningful. In this paper, we synthesize insights on meaningful explanations from a literature study and two use cases in the financial sector. We identified 30 components of meaningfulness in XAI literature. In addition, we report three themes associated with explanation needs that were central to the users in our use cases, but are not prominently described in literature: actionability, coherent narratives and context. Our results highlight the importance of narrowing the gap between theoretical and applied responsible AI.
Original language | English |
---|---|
Title of host publication | Proceedings of the Workshops at the Third International Conference on Hybrid Human-Artificial Intelligence |
Subtitle of host publication | co-located with (HHAI 2024) |
Editors | Petter Ericson, Nina Khairova, Marina De Vos |
Place of Publication | Malmö |
Pages | 221-227 |
Number of pages | 6 |
Volume | 3825 |
Publication status | Published - 10 Jun 2024 |
Event | HHAI-WS 2024: Workshops at the Third International Conference on Hybrid Human-Artificial Intelligence (HHAI) - Malmö, Sweden Duration: 10 Jun 2024 → 14 Jun 2024 |
Conference
Conference | HHAI-WS 2024: Workshops at the Third International Conference on Hybrid Human-Artificial Intelligence (HHAI) |
---|---|
Country/Territory | Sweden |
City | Malmö |
Period | 10/06/24 → 14/06/24 |