Natural Language Interaction with Explainable AI Models Arjun R Akula1, Sinisa Todorovic2, Joyce Y Chai3 and Song-Chun Zhu4 1,4University of California, Los Angeles 2Oregon State University 3Michigan State University
[email protected] [email protected] [email protected] [email protected] Abstract This paper presents an explainable AI (XAI) system that provides explanations for its predictions. The system consists of two key components – namely, the predic- tion And-Or graph (AOG) model for rec- ognizing and localizing concepts of inter- est in input data, and the XAI model for providing explanations to the user about the AOG’s predictions. In this work, we focus on the XAI model specified to in- Figure 1: Two frames (scenes) of a video: (a) teract with the user in natural language, top-left image (scene1) shows two persons sitting whereas the AOG’s predictions are consid- at the reception and others entering the audito- ered given and represented by the corre- rium and (b) top-right (scene2) image people run- sponding parse graphs (pg’s) of the AOG. ning out of an auditorium. Bottom-left shows the Our XAI model takes pg’s as input and AOG parse graph (pg) for the top-left image and provides answers to the user’s questions Bottom-right shows the pg for the top-right image using the following types of reasoning: direct evidence (e.g., detection scores), medical diagnosis domains (Imran et al., 2018; part-based inference (e.g., detected parts Hatamizadeh et al., 2019)). provide evidence for the concept asked), Consider for example, two frames (scenes) of and other evidences from spatiotemporal a video shown in Figure1.