Information-seeking dialogue for explainable artificial intelligence: Modelling and analytics

Argument and Computation 15 (1):49-107 (2024)
  Copy   BIBTEX

Abstract

Explainable artificial intelligence has become a vitally important research field aiming, among other tasks, to justify predictions made by intelligent classifiers automatically learned from data. Importantly, efficiency of automated explanations may be undermined if the end user does not have sufficient domain knowledge or lacks information about the data used for training. To address the issue of effective explanation communication, we propose a novel information-seeking explanatory dialogue game following the most recent requirements to automatically generated explanations. Further, we generalise our dialogue model in form of an explanatory dialogue grammar which makes it applicable to interpretable rule-based classifiers that are enhanced with the capability to provide textual explanations. Finally, we carry out an exploratory user study to validate the corresponding dialogue protocol and analyse the experimental results using insights from process mining and argument analytics. A high number of requests for alternative explanations testifies the need for ensuring diversity in the context of automated explanations.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 92,682

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

The virtues of interpretable medical artificial intelligence.Joshua Hatherley, Robert Sparrow & Mark Howard - forthcoming - Cambridge Quarterly of Healthcare Ethics:1-10.

Analytics

Added to PP
2023-05-25

Downloads
18 (#851,175)

6 months
12 (#239,586)

Historical graph of downloads
How can I increase my downloads?

Author Profiles

Jose Moral
Centro de Investigación Científica y de Educación Superior de Ensenada