<<

Int J Soc manuscript No. (will be inserted by the editor)

Knowledge-Grounded Dialogue Flow Management for Social and Conversational Agents

Lucrezia Grassi · Carmine Tommaso Recchiuto · Antonio Sgorbissa

Received: date / Accepted: date

Abstract The article proposes a system for knowledge- 1 Introduction based conversation designed for Social Robots and other conversational agents. The proposed system relies on an Social is a research field aimed at providing Ontology for the description of all concepts that may robots with a brand new set of skills, specifically re- be relevant conversation topics, as well as their mutual lated to social behaviour and natural interaction with relationships. The article focuses on the algorithm for humans. Dialogue Management that selects the most appropri- Social robots can be used in many contexts such ate conversation topic depending on the user’s input. as education [1], welcoming guests in hotels [2], cruise Moreover, it discusses strategies to ensure a conversa- ships [3], malls [4], and elderly care [5]. A noteworthy tion flow that captures, as more coherently as possible, application of Socially Assistive Robots (SARs) is in the user’s intention to drive the conversation in spe- the healthcare field: it has been argued that robots can cific directions while avoiding purely reactive responses be used to make people feel less lonely and help human to what the user says. To measure the quality of the caregivers taking care of elders in care homes [6][7]. In conversation, the article reports the tests performed particular, robots may help to cope with caregiver bur- with 100 recruited participants, comparing five conver- den, i.e., the stress perceived by formal and informal sational agents: (i) an agent addressing dialogue flow caregivers, a relevant problem both in care homes and management based only on the detection of keywords domestic environments. This subjective burden is one of in the speech, (ii) an agent based both on the detec- the most important predictors for adverse outcomes of tion of keywords and the Content Classification feature the care situation for the caregivers themselves, as well of Google Cloud Natural Language, (iii) an agent that as for the one who requires care. Clinicians frequently picks conversation topics randomly, (iv) a human pre- overlook the caregiver burden problem [8]. tending to be a chatbot, and (v) one of the most famous Recently, social robots have been recognized as a chatbots worldwide: Replika. The subjective perception very important resource by one of the most recent ed- of the participants is measured both with the SASSI

arXiv:2108.02174v1 [cs.RO] 4 Aug 2021 itorials of Science Robotics [9], which analyzes the po- (Subjective Assessment of Speech System Interfaces) tential of robots during the COVID-19 pandemic and tool, as well as with a custom survey for measuring the underlines how Social Robotics is a very challenging subjective perception of coherence. area: social interactions require the capability of au- tonomously handling people’s knowledge, beliefs, and Keywords Social Robotics · Conversational Agents · emotions. In the last year, COVID-19 threatened the Knowledge-Grounded Conversation life of people with a higher risk for severe illness, i.e., older adults or people with certain underlying medi- L. Grassi (B) · C.T. Recchiuto · A. Sgoribissa cal conditions. To slow down the spread of the virus University of , DIBRIS, via all’Opera Pia 13 Tel.: +39 010 3532801 [10], there is the need to reduce social contacts and, as Fax: +39 010 3532154 a consequence, many older adults have been left even E-mail: [email protected] more socially isolated than before. However, in many cases, loneliness can spring adverse psychological effects 2 Lucrezia Grassi et al. such as anxiety, psychiatric disorders, depression, and conversation to provide coherent replies, even when it decline of cognitive functions [11]. In this scenario, So- is not able to fully understand what the user is talking cial Robotics and Artificial Intelligence, in general, may about (which is likely to happen very often if the user have a crucial role: conversational robots and virtual is free to raise any topic). agents can provide social interactions, without spread- To increase the impact of the conversational sys- ing the virus. tems of Social Robots, [19] argues that they shall be In many Social Robotics applications, the main fo- designed to overcome a number of limitations. Some cus is on the conversation. Based on its purpose, the desirable features that are not present in most systems conversation can be subdivided into: are: (1) Breaking the “simple commands only” bar- – task-oriented: it is used to give commands to per- rier; (2) Multiple speech acts; (3) Mixed initiative dia- form a task or retrieve information. Examples of logue;(4) Situated language and the symbol grounding task-oriented conversations can be found both dur- problem; (5) Affective interaction; (6) Motor correlates ing the interaction with a robot, i.e., “Go to the and Non-Verbal Communication; (7) Purposeful speech kitchen”, and during the interaction with a smart and planning; (8) Multi-level learning; (9) Utilization of assistant, i.e., “Turn off the light”. In this case, the online resources and services. problem is to understand the semantics of what the The main contribution of the article is to investigate user is saying, to provide the required service. possible solutions to issue 7, that we aim to achieve by – chit-chatting: it has the purpose of gratifying the properly managing the conversation flow towards the user in the medium to long term, keeping him/her execution of tasks or the exploration of relevant top- engaged and interested. For this kind of conversa- ics, thus ultimately leading to a more engaging interac- tion, the most relevant problem, rather than per- tion with the user. During the conversation, it typically fectly grasping the semantics of what the person happens that a robot may provide answers that have says, is to be able to manage the conversation, show- nothing to do with what the user says: this has a very ing the knowledge and competence on a huge num- negative impact on the “suspension of disbelief” that is ber of different topics coherently and entertainingly required to give the user the impression of genuine in- [12]. telligence, and ultimately it generates frustration. Here, a key element is the capability to know when to further In this scenario, the CARESSES project1 is the first explore the current topic or choose the next conversa- project having the goal of designing SARs that are cul- tion topic, coherently with the user’s sentence: if the turally competent, i.e., able to understand the culture, algorithm does not pick a topic coherently, the agent’s customs, and etiquette of the person they are assisting, reply will not be appropriate, even if the system had while autonomously reconfiguring their way of acting the knowledge to interact consistently. and speaking [13]. CARESSES exploited the humanoid Specifically, we proceed as follows. robot Pepper2 as the main robotic platform for inter- acting with people [6], in particular with older people First, we propose a novel system for knowledge- in care homes, to make them feel less isolated and re- based conversation and Dialogue Management that re- duce the caregiver burden. Under these conditions, the lies on an Ontology for the description of all relevant capability of engaging the user to chit-chat with the sys- concepts that may play a key role in the conversation: tem becomes of the utmost importance, as it has been the Ontology is designed to take into account the pos- shown [7] that this may have a positive impact on qual- sible cultural differences between different users in a ity of life [16], negative attitude towards robots [17], and no-stereotyped way, and it stores chunks of sentences loneliness [18]. However, for a proper conversation to be that can be composed in run-time, therefore, enabling possible, the system shall be able to talk about a huge the system to talk about the aforementioned concepts number of topics that may be more or less relevant for in a culture-aware and engaging way [14][20]. different cultures, by properly managing the flow of the Second, we test the developed solutions by compar- ing five Artificial Conversational Agents during a con- 1 http://caressesrobot.org/ versation with 100 recruited participants: (i) an agent 2 CARESSES architecture has been recently revised to al- addressing dialogue flow management based only on the low for a Cloud-based implementation [14], thus enabling vir- tually any device equipped with a network interface to per- detection of keywords, (ii) an agent addressing dialogue form long-term, culture-aware conversation with the user. flow management based both on the detection of key- Currently, the robots Pepper and NAO, the pill-dispenser words and the Content Classification feature of Google Pillo, Prof. Einstein, and a virtual character implemented as Cloud Natural Language, (iii) an agent that picks ran- an application have been provided with the onboard functionalities to implement the instructions received by the dom topics among those present in the Ontology, (iv) a CARESSES Cloud [15]. human pretending to be a chatbot, and (v) one of the Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 3 most famous chatbots worldwide: Replika. The subjec- as “I see, very interesting” as it has no previous knowl- tive perception of the participants is measured both edge about Pizza. On the other side, the latter may be with the SASSI (Subjective Assessment of Speech Sys- more specific, i.e., “Oh, pizza is a very delicious Italian tem Interfaces) tool [21][22] and a custom survey for food”, if it had the chance to recall internal or exter- measuring the subjective perception of coherence in the nal knowledge about pizza, e.g., using resources on the conversation. web. The remainder of this article is organized as follows. Building a knowledge-grounded conversational sys- Section 2 presents an overview of previous works related tem raises many challenges. For instance, the internal to Knowledge-Grounded Conversation, the most popu- knowledge may be static or “expandable”, i.e., updated lar validated tools to evaluate the User Experience, and in run-time with new external information retrieved introduces the concept of “coherence” in Dialogue Man- from the user utterances, which will be considered as agement. Eventually, it presents a typical classification internal knowledge in the next interactions. The exter- for Artificial Conversational Agents and describes the nal knowledge in most cases will come from websites: up-to-date most famous chatbots. An overview of the however, even if the system can find some documents CARESSES knowledge-based conversational system is associated with the current conversation topic by using given in Section 3. Section 4 describes the experiment a state-of-the-art information retrieval system, it may carried on to evaluate the user satisfaction when inter- be difficult to extract knowledge from the search results acting with the agents, along with the statistical tools because this would typically require complex Natural used to analyse the collected data. Section 5 presents Language Processing techniques to select the appropri- the results obtained, discussed in detail in Section 6. ate knowledge. Eventually, Section 7 draws the conclusions. This is done in [23] using a Generative Transformer Memory Network and in [24] using an SVM classifier, to name a few. Another complex task is to generate 2 State of the Art appropriate responses that reflect the acquired knowl- edge as a consequence of the conversation history. The This section addresses Knowledge-Grounded Conversa- problems of knowledge extraction and response gener- tion, presenting a brief analysis of the related Litera- ation, among the others, are addressed in [25], where a ture. Moreover, it gives an overview of the most used knowledge-grounded multi-turn chatbot model is pro- tools to evaluate the User Experience during the in- posed (see also 2.3 for popular chatbots based on this teraction with conversational agents and introduces a principle). new metric exploited in our experiment. Eventually, A data-driven and knowledge-grounded conversa- it presents a classification for Artificial Conversational tion model, where both conversation history and rel- Agents and describes the up-to-date most famous chat- evant facts are fed into a neural architecture that fea- bots. tures distinct encoders for the two entries, is proposed in [26]. Such model aims to produce more appropriate responses: the article shows that the outputs generated 2.1 Knowledge-Grounded Conversation by a model trained with sentences and facts related to the conversation history were evaluated by human A knowledge-grounded conversational system is a dia- judges as remarkably more informative (i.e., knowledge- logue system that can communicate by recalling inter- able, helpful, specific) with respect to those of a model nal and external knowledge, similarly to how humans trained with no facts. do, typically to increase the engagement of the user dur- In [27], a large corpus of status-response pairs found ing chit-chatting. The internal knowledge is composed on Twitter has been employed to develop a system of things that the system already knows, while exter- based on phrase-based Statistical Machine Translation, nal knowledge refers to the knowledge acquired in run- able to respond to Twitter status posts. As stated in the time. To this end, knowledge-grounded systems must article, data-driven response generation will provide an not only understand what the user says but also store important breakthrough in the conversational capabil- external knowledge during the conversations, and their ities of a system. Authors claim that when this kind of responses should be based on the stored knowledge. approach is used inside a broad dialogue system and is There are many differences between the replies of a combined with the dialogue state, it can generate lo- normal conversational system and those of a knowledge- cally coherent, purposeful, and more natural dialogue. grounded system. For example, if the user says “I love Finally, the problem of knowledge representation pizza”, the former might provide a general answer such and knowledge-based chit-chatting, keeping into account 4 Lucrezia Grassi et al. the cultural identity of the person, is addressed in [20] utterance, and then act appropriately. However, accord- and [14] using an Ontology designed by cultural experts ing to the aforementioned definition, what the System coupled with a Bayesian network, to avoid rigid rep- Response Accuracy scale aims to measure is not equiv- resentations of cultures that may lead to stereotypes. alent to the broader concept of coherence that has been Such an approach allows the system to have more con- given: it measures the quality of the system’s reply to trol over what the robot says during the conversation: the user’s input in a purely reactive fashion, without re- the idea of having a knowledge base designed by experts lying on the concept of “current topic of conversation”. is particularly suitable for sensitive situations, i.e. when For this reason, a coherence measure in the spirit of dealing with the more fragile population such as older [31] and [32] has been used to supplement the SASSI adults or children. Approaches that automatically ac- questionnaire, which requires users to rate individual quire knowledge from the Internet may not be optimal sentences pronounced by the robot at the light of the in these scenarios. context (details are given in Section 4.2.1). Notice that [32] and [33] propose also methods to automatically evaluate the coherence, as this is considered an impor- 2.2 User Experience tant metrics to evaluate multi-turn conversation in open domains: however, in this work we are only interested In the Literature, tools exist that can be used to eval- in evaluating the subjective perception of users, which uate the User Experience (UX) and the overall qual- makes a simple rating mechanism perfectly fitting our ity of the conversation. The work carried on in [28] re- purposes. views the six main questionnaires for evaluating conver- sational systems: AttrakDiff, SASSI, SUISQ, MOS-X, PARADISE, and SUS. Moreover, it assesses the poten- 2.3 Artificial Conversational Agents tial suitability of these questionnaires to measure vari- ous UX dimensions. Depending on the emphasis they put on a task-oriented As a measure of the quality of the flow of conversa- conversation or chit-chatting, a typical classification for tion, we aim to evaluate how “coherent” the system is: artificial conversational agents is based on their scope: this shall be done not only taking into account the last user’s utterance but also the current conversation topic – Question answering bots: knowledge-based conver- (i.e., what is referred to as context in [29][30]). To clar- sational systems that answer to users’ queries by ify what this means, suppose a conversational system analysing the underlying information collected from that replies very accurately to everything the user says: various sources like Wikipedia, DailyMail, Allen AI if the user talks about football, the system will reply science and Quiz Bowl [34]; talking about football, if the user talks about apples, – Task-oriented bots: conversational systems that as- the system will reply talking about apples. But what if sist in achieving a particular task or attempt to solve the person and the system are talking about football, a specific problem such as a flight booking or hotel and the person says something that is not recognized as reservation [35]; related to anything specific, such as “I will think about – Social bots: conversational systems that communi- that”, “I think so”, “It’s so nice to talk with you”, and cate with users as companions, and possibly en- so on? A coherent Dialogue Management system shall tertain or give recommendations to them [36]: re- be able to understand when it is more appropriate to cent notable examples are Microsoft Xiaoice [37] and further explore the current conversation topic (in this Replika3. example, by taking the initiative to ask the person a question about his/her preferred team or players) or In the following, we discuss more in detail the agents lead the dialogue to another topic, coherently with what belonging to the third class pointing out differences and the user said. similarities with out solution, without making a dis- Unfortunately, the aforementioned UX tools are not tinction between robots and chatbots, unless strictly meant to evaluate a medium-term mixed-initiative dia- required. logue, as in our case. The most similar measure to what For many decades, the development of social bots, we want to evaluate is the System Response Accuracy or intelligent dialogue systems that can engage in em- scale of the SASSI questionnaire: such a questionnaire, pathetic conversations with humans, is one of the main described more in detail in Section 4.2.2, has been used goals of Artificial Intelligence. As stated in [38], early during our experiments. The System Response Accuracy conversational systems such as Eliza [39], Parry [40], is defined as the system’s ability to correctly recognise 3 https://help.replika.ai/hc/en-us/articles/ the speech input, correctly interpret the meaning of the 115001070951-What-is-Replika- Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 5 and Alice [41] were designed to mimic human behaviour conversation. Unfortunately, we could not try XiaoIce in a text-based conversation, hence to pass the Tur- as it is not available in . ing Test [42] within a controlled scope. Despite their Replika is presented as a messaging app where users impressive successes, these systems, which were pre- answer questions to build a digital library of informa- cursors to today’s social chatbots, worked well only in tion about themselves. Its creator, a San Francisco- constrained environments. In more recent times, among based startup called Luka, sees see a whole bunch of the most successful conversational agents for general possible uses for it: a digital twin to serve as a compan- use that can act as digital friends and entertainers, Xi- ion for the lonely, a living memorial of the dead, created aoice, Replika, Mitsuku, and Insomnobot-3000 are the for those left behind, or even, one day, a version of our- most frequently mentioned [43]. selves that can carry out all the mundane tasks that we XiaoIce (“Little Ice” in Chinese) is one of the most humans have to do, but never want to. popular chatbots in the world. It is available in 5 coun- To the best of the authors’ knowledge, there is no tries (i.e., China, Japan, US, India, and Indonesia) un- detailed explanation of the implementation of Replika, der different names (e.g., Rinna in Japan) on more hence we can only make assumptions on how issues re- than 40 platforms, including WeChat, QQ, Weibo, and lated to Dialogue Management are faced. After having Meipai in China, Facebook Messenger in the United intensively tried Replika, it seems that the conversation States and India, and LINE in Japan and Indonesia. is partitioned into “sessions” in which it has specific Its primary goal is to be an AI companion with which competencies, somehow playing a similar role as top- users form long-term emotional connections: this dis- ics in XiaoIce and our system. Concerning issue 7 in tinguishes XiaoIce not only from early chatbots but Section 1, the chatbot uses a Neural Network to hold also from other recently developed conversational AI an ongoing, one-on-one conversation with its users, and personal assistants such as Apple Siri, Amazon Alexa, over time, learn how to talk back [44]. The agent is Google Assistant, and Microsoft Cortana. trained on texts from more than 8 million web pages, As stated in [37], the topic database of XiaoIce is from Twitter posts to Reddit forums, and it can re- periodically updated by collecting popular topics and spond in a thoughtful and human-like way. From time related comments and discussions from high-quality In- to time, it will push the user to have a “session” to- ternet forums, such as Instagram in the US and the gether: in this session, it will ask questions regarding website douban.com in China. To generate responses, what the user did during the day, what was the best XiaoIce has a paired database that consists of query- part of the day, what is the person looking forward to response pairs collected from two data sources: human tomorrow, and eventually, the user has to rate his/her conversational data from the Internet, (e.g., social net- mood on a scale of 1 to 10. During this session, Rep- works, public forums, news comments, etc.), and human- lika takes control of the conversation and insists that machine conversations generated by XiaoIce and her the user answers its questions about a specific matter, users. Even if the data collected from the Internet are something that may be annoying and frustrating, and subjected to quality control to remove personally iden- we avoid by letting the user free to easily switch to an- tifiable information (PII), messy code, inappropriate other topic at any time. Replika’s responses are backed 4 content, spelling mistakes, etc., the knowledge acqui- by Open AI’s GPT-2 text-generating AI system [45]. 5 sition process is automated and it is not supervised Mitsuku , or Kuki as her close friends call her, is a 6 by humans: differently from our system, this solution chatbot created by Pandorabots : an open-source chat- may not be suitable in sensitive situations, e.g., when bot framework that allows people to build and publish dealing with the more fragile population such as older AI-powered chatbots on the web, mobile applications, people or children. Regarding issue 7 in Section 1 (i.e., and messaging apps like LINE, Slack, WhatsApp, and the lack of flow in the conversation), the implementa- Telegram. tion of XiaoIce addressed this problem by using a Topic The Pandorabots chatbot framework is based on the Manager that mimics human behavior of changing top- Artificial Intelligence Markup Language (AIML) script- ics during a conversation. It consists of a classifier for ing language, which developers can use to create con- deciding at each dialogue turn whether or not to switch versational bots. The downside is that Pandorabots do topics and a topic recommendation engine for suggest- not include machine learning tools that are common ing a new topic. Topic switching is triggered if XiaoIce on other chatbot building platforms. A specific AIML does not have sufficient knowledge about the topic to file allows users to teach Mitsuku new facts: the user engage in a meaningful conversation, an approach that 4 https://en.wikipedia.org/wiki/GPT-2 our system adopts as well by relying on an Ontology 5 https://www.pandorabots.com/mitsuku/ for representing relationships among different topics of 6 https://home.pandorabots.com/home.html 6 Lucrezia Grassi et al. should say “Learn” followed by the fact (i.e., “Learn time, more specific to the context. Authors report also the sun is hot”). The taught information is emailed to that issues related to inappropriate and biased language its creator Steve Worswick, which will personally su- have still to be solved: to the best of our knowledge, a pervise the learning process. To address issue 7 in Sec- publicly available version of this chatbot has not been tion 1, specific areas of competence of the chatbot are released yet. managed by different AIML files, which are responsible BlenderBot [30] is a very recent open-domain chat- for maintaining coherence in the responses: AIML files, bot developed at Facebook AI Research. As like as we therefore, play a similar role as topics/concepts in the do, the authors put a stronger emphasis on aspects that Ontology that are the core elements of our system, but are not only related to sentence generation: they argue without the benefit of having a hierarchical structure that good conversation requires a number of skills to as a key element for Dialogue Management. Mitsuku is be blended in a seamless way, including providing en- a five-time Loebner Prize winner (in 2013, 2016, 2017, gaging topics of conversation as well as listening and 2018, 2019), and it is available in a web version7, Face- showing interest to what the user says, among the oth- book Messenger, Kik Messenger, and Telegram. ers. Current generative models, they argue, tend to pro- Insomnobot-3000 8 is the world’s first bot that is duce dull and repetitive responses (and, we would add, only available to chat, exclusively via SMS, between they do not take cultural appropriateness into account), 11pm and 5am regardless of the time zone. The bot was whereas retrieval models may produce human written built by the mattress company Casper and, according utterances that tend to include more vibrant (and cul- to its creators, it was programmed to sound like a real turally appropriate) language. To overcome the limita- person and talk about almost anything. Insomnobot- tions of purely generative models, BlenderBot includes 3000 cannot learn new things and expand its knowl- a retrieval step before generation according to a retrieve edge base: however, it can generate over 2,000 different and refine model [46], which retrieves initial dialogue responses, depending on which category and emotion utterances and/or knowledge from a large knowledge the keyword falls under. Concerning issue 7 in Section base. As like Meena, the system is trained to choose 1, when the bot receives a text, it chooses an appro- the next dialogue utterance given the recent dialogue priate response by identifying keywords: however, dif- history, referred to as context: training is performed us- ferently from the aforementioned systems and our solu- ing a number of datasets that are key to exhibit differ- tion, it lacks a more sophisticated mechanism for Dia- ent skills (i.e., engaging personality, emotional talking, logue Management to switch in-between different areas knowledge-grounded conversation), as well as a Blended of competence or topics to provide contextual replies. Skill Talk dataset that merges utterances taken from In addition to the aforementioned chatbot avail- the three. Authors claim that their model outperforms able to the large public, recent research on data-driven Meena in human evaluations, however, they acknowl- knowledge-grounded conversation based on sophisticated edge that the system still suffers from a lack of in- generative models is worth being mentioned. depth knowledge if sufficiently interrogated, a tendency Meena [29] is a generative chatbot model trained to stick to simpler language, and a tendency to repeat end-to-end on 40B words mined from public domain often used phrases. social media conversations, that addresses the problem DialoGPT [47] is a neural conversational response of multi-turn conversation in open domains. Accord- generation model trained on 147M conversation-like ex- ing to its authors, this chatbot can conduct conversa- changes extracted from Reddit comment chains over a tions that are more sensible and specific than existing period spanning from 2005 through 2017 and it extends state-of-the-art chatbots. Such improvements are mea- GPT-2. Authors claim that conversational systems that sured by a new human evaluation metric, called Sensi- leverage DialoGPT generate more relevant, contentful, bleness and Specificity Average (SSA), which captures and context-consistent responses than strong baseline important attributes for human conversation. Meena is systems. As it is fully open-sourced and easy to deploy, based on the concept of multi-turn contexts to evaluate users can extend the pre-trained conversational system the quality of a response: not only the last sentences to bootstrap training using various datasets. Authors but the recent conversation history in terms of (con- claim that the detection and control of toxic output text, response) pairs is used to train and evaluate the will be a major focus of future investigation. network. Experiments show significant improvements in All generative models based on human-human data SSA with respect to competitors, providing replies that offers many advantages in producing human-like utter- are coherent with what the user said and, at the same ances fitting the context, but have the major drawback 7 https://chat.kuki.ai that they can learn undesirable features leading to toxic 8 https://insomnobot3000.com or biased language. Classifiers to filter out inappropri- Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 7 ate language exist [48], but they still have limitations: whichever the cultural identity of the user is, to avoid this issue is particularly important when dealing with stereotypes (an example related to different kind of bev- the more frail populations, especially by considering erages is shown in Figure 2). The system will initially that classifying language as inappropriate or offensive guess the user’s beliefs, values, habits, customs, pref- typically depends on cultural factors, and the devel- erences based on the culture he/she declared to self- opment of classifiers working properly in multiple cul- identify with, but it will then be open to considering tures may be very challenging. Since we claim that a choices that may be less likely in a given culture, as the proper flow of conversation deserves higher attention user explicitly declares his/her attitude towards them. than the automatic generation of sentences, we address According to this principle, the system may initially in- these problems through an Ontology of conversation fer that an English person may be more interested to topics and topic-related chunks of sentences generated talk about Tea rather than Coffee, and the opposite by cultural experts, that are then composed in run-time may be initially inferred for an Italian user. However, using the hierarchical structure of the knowledge-base during the conversation, initial assumptions may be re- to produce a huge variety of sentences. Finally, even if vised, thus finally leading to a fully personalized repre- this is not the main objective of this article, it is worth sentation of the user’s attitude towards all concepts in reminding that our system may take cultural aspects the TBox, to be used for conversation. into account in Dialogue Management, an element that is completely ignored by all state-of-the-art systems.

3 System Architecture

Even if the focus of this article is on knowledge-based chit-chatting and dialogue flow management, and not on cultural adaptation, it is necessary to briefly in- troduce the structure of the cultural knowledge base used by the CARESSES conversational system since it strongly influences the algorithms we implemented.

3.1 Knowledge Representation

In CARESSES, the ability of the companion robot to naturally converse with the user has been achieved by Fig. 1 Knowledge representation architecture for a cultur- ally competent robot. creating a framework for cultural knowledge represen- tation that relies on an Ontology [49] implemented in OWL2 [50]. According to the Description Logics formal- To implement this mechanism, the Culture-Specific ism, concepts (i.e., topics of conversation the system is ABox layer comprises instances of concepts (with prefix capable of talking about) and their mutual relations are EN for “English” in Figure 2) encoding culturally ap- stored in the terminological box (TBox) of the Ontol- propriate chunks of sentences to be automatically com- ogy. Instead, instances of concepts and their associated posed (Data Property hasSentence) and the probabil- data (e.g., chunks of sentences automatically composed ity that the user would have a positive attitude toward to enable the system to talk about the corresponding that concept, given that he/she belongs to that cultural topics) are stored in the assertional box (ABox). group (Data Property hasLikeliness). To deal with representations of the world that may Eventually, the Person-Specific ABox comprises in- vary across different cultures [51], the Ontology is or- stances of concepts (with prefix DS for a user called ganized into three layers, as shown in Figure 1. The “Dorothy Smith” in Figure 2) encoding the actual user’s TBox (layer I) encodes concepts at a generic, culture- attitude towards a concept (Mrs Dorothy Smith may be agnostic level, which can be inherited from existing more familiar with having tea than the average English upper and domain-specific Ontologies (grey boxes) or person, hasLikeliness=Very High), sentences to talk explicitly defined to enable culture-aware conversation about a topic explicitly taught by the user to the sys- with the user (white boxes). An important point to be tem (hasSentence=“You can never get a cup of tea highlighted is that the TBox should include concepts large enough or a book long enough to suit me”) or that are typical of the union all cultures considered, other knowledge explicitly added during setup (e.g., the 8 Lucrezia Grassi et al.

user’s name and the town of residence). At the first en- counter between the robot and a user, many instances of the Ontology will not contain Person-Specific knowl- edge: the robot will acquire this awareness at run-time either from its perceptual system or during the interac- tion with the user, e.g., asking questions. Figure 1 also shows that some instances of existing Ontologies that are not culture- or person-dependent (dark circles) may not change between the two ABox layers: this may re- fer, for instance, to technical data such as the serial number of the robot or physical quantities, if they need to be encoded in the Ontology. For a detailed descrip- tion of the terminological box (TBox) and assertional box (ABox) of the Ontology, as well as the algorithms for cultural adaptation, see [14][20]. The Dialogue Tree (DT) (Figure 2), used by the chit-chatting system (Section 3.2), is built starting from the Ontology structure: each concept of the TBox and the corresponding instances of the ABox are mapped into a conversation topic, i.e., a node of the tree. The re- lation between topics is borrowed from the structure of the Ontology: specifically, the Object Property hasTopic and the hierarchical relationships among concepts and instances are analyzed to define the branches of the DT. In the example of Figure 2, the instance of Tea for the English culture is connected in the DT to its child node GreenTea (which is a subclass of Tea in the Ontology), and its sibling MilkTea (since EN MILK is a filler of EN TEA for the Object Property hasTopic). As mentioned, each conversation topic has chunks of culturally appropriate sentences associated with it that are automatically composed and used during the conversation. Such sentences can be of different types (i.e., positive assertions, negative assertions, different kinds of questions, or proposals for activities) and are selected in subsequent iterations, also depending on in- put received from the user. Typically, when exploring a topic, the system may first ask a question to understand if the user is familiar with that topic: if so, this may be followed by general considerations, proposals for activ- ities, or open questions, until the system moves to the next topic if the conditions hold. Multiple sentences of the same type are present, randomly chosen to reduce the chance of repetitions. At any time, the user is free to take the initiative to express his/her considerations or lead the conversation somewhere else. Sentences may contain variables that are instantiated when creating the DT. For instance, a hypothetical sentence “Do you like $hasName?” encoded in the concept Coffee might Fig. 2 The three layers of the Ontology: TBox, CS-ABox (for be used to automatically produce both “Do you like the English culture), PS-ABox (for the user Dorothy Smith), and the Dialogue Tree generated from the Ontology structure. Coffee?” and “Do you like Espresso?” (being Espresso a subclass of Coffee). As further example, the sentence “I love $hasName with $hasTopic*hasName” encoded Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 9 in the concept Movie can be used to produce both “I BreakfastHabits) to more specific (e.g., the topic love movies with great actors” and “You know... I love HavingTeaForBreakfast) or related ones (e.g., the Bollywood movies with Amitabh Bachchan”, depend- topic TakingCareOfOneself). This is achieved by ing on the taxonomy of the knowledge base and ad- selecting the branches to maximise the probability ditional rules for sentence composition. In the current that the user will be interested in the next topic version used for testing, exploiting the taxonomy of the (also based on cultural factors). When in a topic, Ontology allowed us to easily produce a DT with 2,470 the system asks questions, reply with positive or topics of conversation and 24,033 sentences, with ran- negative assertions, propose activities or simply en- dom variations made in run-time. courage the user to freely express his/her view, by A few additional words are worth spending to clar- composing chunks of sentences encoded in the on- ify this concept. Currently, as in the aforementioned ex- tology, until the topic has been sufficiently explored. ample, one sentence may contain two variables whose Probabilities associated with topics have been ini- value depends on (i) the specific concept that inherited tially assigned with the help of Transcultural Nurs- that Data Property from superconcepts in the Ontol- ing experts as well as looking for information on the ogy, and (ii) related concepts connected to it through an Internet about habits, foods, sports, religions, etc. in Object Property. Then, when manually adding a sen- different cultures. Please, notice however that these tence as a Data Property of a given concept, a num- are only initial guesses, as the system will update ber of sentences will be inherited and added automat- probabilities during the conversation depending on ically depending on the number of (i) its subconcepts the user’s preferences and attitudes to avoid stereo- and (ii) their role-fillers. By considering a DT gener- typed representations. ated from the Ontology and having a branching factor 2. The second mechanism enables the Dialogue Man- B = Bs + Bp, where Bs is the branching factor due agement algorithm to jump to a different topic of to subconcepts and Bp is the branching factor due to the DT depending on what the user says (e.g., when role-fillers, adding a sentence in a concept at a height talking about BreakfastHabits the person may start H from DT leaves produces a number of sentence in talking about a Restaurant he/she loves). After do- the order of O(HB); by considering that all sentences ing so, the system reverts to the first mechanism are randomly associated in run-time with a prefix (e.g., but starting from the new topic: it asks questions, “You know...”, “I heard that...”) and may be appended makes positive and negative comments, proposes ac- to each other, this increases the number of variations tivities, or allows the person to freely express his/her even more. For instance, in the current configuration view about the new topic, and then re-start follow- with 2,470 concepts, adding a sentence at the root of ing branches as previously. the DT will produce a maximum of 2,470 inherited sen- With reference to Figure 2, a possible dialogue pat- tences in the whole Ontology (which however, may lack tern between the robot and Mrs Dorothy Smith, using specificity, as it is unlikely that the same pattern may the aforementioned mechanisms, would be the follow- be easily adapted to talk about everything in the ontol- ing (labels 1 and 2 identify the mechanism currently ogy: i.e., “I love talking about hasTopic with you” may $ adopted by the system to navigate the DT): work for all subconcepts, but “I like eating hasTopic” $ R. (1): Do you usually drink something with your will not). meals? Mrs S.: Always! At least I try... R. (1): Drinking water is very important for health! 3.2 Chit-Chatting Please, tell me more about your drinking preferences. When the user pronounces a sentence that triggers the Mrs S.: Well, I mostly drink water during lunch. start of a conversation, the system switches to the chit- R. (1): I know that many English persons love tea. chatting state. In this case, the Dialogue Management Do you? algorithm operates to keep the conversation as engaging Mrs S.: Yes, of course! But mostly in the afternoon. and natural as possible by implementing two different, R. (1): You can never get a cup of tea large enough intertwined methods to navigate the DT based on what or a book long enough to suit me. Do you usually have the user says: milk with your tea? Mrs S.: Yes, and scones, too! But they are not very 1. The first mechanism implemented by the Dialogue healthy. Management algorithm simply chooses the topic based R. (2): I love scones with jam and whipped cream! on the DT, following the branches from more general Figure 3 depicts a detailed flow-chart that clarifies conversation topics close to the root (e.g., the topic what happens during the interaction with the user. 10 Lucrezia Grassi et al.

The diagram shows that, as soon as the person says something, the Dialogue Management algorithm first checks if the user’s sentence contains some command that is aimed to trigger an activity (box Perform the required activity). This task-oriented part of the di- alogue has been implemented in our experiments with caring robots during the CARESSES project, but not in the tests described in this article, and it may rely on a third-party system for NLP9 or a pre-existing com- mand interface that is already implemented onboard the robot (e.g., verbal commands to clean the room in case of a robotic vacuum cleaner, or to deliver medicines in case of a robotic pill dispenser). If a command is not found, the user’s input is pro- cessed (box Process user sentence with Dialogue Management mechanism 1 or 2) by using the two mech- anisms described so far: as a first step the user’s utter- ance is analysed to check if something relevant is de- tected, and possibly used to jump to another topic us- ing mechanism 2; if no information is found to jump to another topic, mechanism 1 checks if the current topic shall be further explored or it is time to follow the branches of the DT from a parent node to a children node that is semantically related in the Ontology. Even- tually, it should be mentioned that activities (e.g., vac- uum cleaning or pill delivery) may even be proposed by the Dialogue Management System itself when related Fig. 3 Flow chart of the basic chit-chatting framework. to the current topic of discussion (i.e., if the person is talking about cleaning and the robot has this capabil- ity). The interaction continues in this way until the user at evaluating the quality of the conversation: the anal- explicitly asks to stop it. ysis performed was rather aimed to measure improve- This approach to Dialogue Management, despite its ment in health-related quality of life (SF-36 [16]), neg- simplicity, reveals to be very effective for chit-chatting. ative attitude towards robots (NARS [17]), loneliness It should be reminded that here Dialogue Management (ULS-8 [18]). Indeed, additional analysis with a wider is not meant to understand all the nuances of human population is needed to evaluate the quality of the con- language when giving a specific command to be ex- versation produced by our system in terms of issue 7 in ecuted (i.e., “Put the red ball on the green table”), Section 1, and motivates the present study. When nav- but rather it aims at engaging the user in rewarding igating in the knowledge base following the branches conversations: an objective that we try to achieve by of the DT (aforementioned mechanism 1), coherence in enabling the system to coherently talk about a huge the flow of conversation is straightforwardly preserved number of different topics while properly managing the by the fact that two nodes of the DT are semantically conversation flow. Please notice that the approach has close to each other by construction. However, to pre- been already tested in the CARESSES project with serve coherence when deciding if it is needed to jump more than 20 older people of different nationalities, from one node to another (mechanism 2), proper strate- for about 18 hours of conversation split into 6 ses- gies should be implemented to avoid the feeling that sions: both the quantitative and qualitative analysis the system is “going off-topic” with respect to what performed revealed very positive feedback from the par- the person says: on the one side, the system needs to be ticipants [6][7]. However, the quantitative analysis per- enough responsive to promptly move to another topic if formed with care home residents was not explicitly aimed the person wants to; on the other side, the system needs to avoid jumping from one topic to another, overesti- 9 For instance, https://dialogflow.cloud.google.com/, mating the desire of the person to talk about something that in the current implementation also manages small-talk requests triggered by sentences such as “Hello”, “How are else when he/she would be happier to further explore you”, or similar). the current topic of conversation. Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 11

3.3 Jumping to a different discussion topic

When using the two mechanisms described in the pre- vious section to navigate the DT, to preserve coher- ence in the conversation flow the system uses one of the two following methods: keyword-based topic match- ing and keyword- and category-based topic matching. These methods differ in terms of complexity and in terms of the need to rely on third-party services to an- alyze the semantic content of sentences.

3.3.1 Keyword-Based Topic Matching

The first, and the simpler, method is exclusively based on the detection of keywords in the user sentence (key- words are manually encoded in the Ontology in a corre- sponding Data Property). To match a topic, at least two keywords associated with that topic should be detected in the sentence pronounced by the user, using wildcards to enable more versatility in keyword matching. The use of multiple keywords allows the system to differen- tiate between semantically close topics (i.e., Green Tea rather than the more general concept of Tea). Figure 4 shows a possible implementation of the box Process user sentence with Dialogue Management Fig. 4 Flow chart of the keyword-based topic matching mechanism 1 or 2 in Figure 3 when using the keyword- method. based topic matching method: – If the user’s sentence contains keywords correspond- the keyword-based method, to find the most appropri- ing to a topic in the Ontology (left part of the di- ate topic to continue the conversation. The algorithm agram), the algorithm returns a topic of the DT finds the word “interest” matching with the keyword 1 that matches the keywords. In case more matches of the topic HOBBY. Then it should ideally check if the are found, the algorithm randomly chooses one of sentence contains a word matching with keyword 2 as- the matching topics. sociated with the same topic: however, in this case, it – Otherwise: it turns out that no second keyword is needed since a – If there are still relevant questions to be asked wildcard is used in the topic HOBBY to guarantee that or assertions to be made about that topic, the one keyword is sufficient. As a result, the algorithm will system stays on the same topic. return the only matching node of the DT: the system – Otherwise: will start asking the user about his/her hobbies (which • If not at the bottom of the DT, the Dialogue does not look very appropriate). The second limitation Management System keeps on exploring the related to the behaviour of this algorithm arises when DT along its branches; no keywords are found and we are at the bottom of the • Otherwise the algorithm returns a random DT: in this case, as already mentioned, the algorithm topic immediately below the root. has no information to continue the conversation and This basic approach, despite its simplicity, has been jumps to a random topic close to the root. successfully exploited during the experimental trial in care homes, and its capability to provide (almost al- 3.3.2 Keyword- and Category-Based Topic Matching ways) coherent, knowledge-grounded replies, was con- firmed in public exhibitions. However, when focusing on The second algorithm is meant to solve the issues of the coherence, the approach has obvious limitations. Let’s previous algorithm, or at least reduce their frequency, suppose that the user is having a conversation with the by enabling a more complex understanding of the se- system and at some point, for some reason, the user mantic content of the sentence. For this purpose, this says “My bank account has a high interest”. This sen- algorithm not only exploits the keywords, but it also tence, like any other sentence, is provided as input to takes into account the “category” of the user’s sentence: 12 Lucrezia Grassi et al.

rent Ontology, only 122 topics have not been associated with a category, as none was found to be appropriate. Hence, the topics mapped to at least a CNL Category11 are 2,348. Figure 5 shows a possible implementation of the box Process user sentence with Dialogue Management mechanism 1 or 2 in Figure 3 when using both key- words and categories for topic matching. As a first step, the algorithm checks whether the sentence contains at least 20 tokens: the Content Classification feature of CNL does not work if the input does not satisfy this requirement. If the sentence is not long enough, it is replicated until it contains at least 20 words. Then it proceeds as follows: – If the user’s sentence contains keywords correspond- ing to a topic in the Ontology and CNL returns at least a category associated with the sentence (left part of the diagram), the algorithm selects the topic(s) in the Ontology with the greatest number of cate- gories in common: if there is more than one topic, it picks the closest to the DT root or, if there are mul- tiple topics at the same level, it chooses randomly; Fig. 5 Flow chart of the keyword- and category-based topic – Otherwise, if no keywords match or the sentence matching method. does not have any category: – If there are still relevant questions to be asked in the current implementation, it uses the third-party or assertions to be made about that topic, the services provided by Cloud Natural Language (CNL) system stays on the same topic. API10, an advanced tool for Sentiment Analysis, En- – Otherwise: tity Recognition, and Content Classification, among the • If not at the bottom of the DT, the Dialogue others. Management System keeps on exploring the To exploit the information related to the category DT along its branches; of the sentence, the Content Classification of the topics • Otherwise: contained in the Ontology is a fundamental operation · If there are keywords or CNL categories and needs to be performed in a setup phase to create associated with the sentence (we already a mapping between the topics contained in the Ontol- know there are not both), the algorithm ogy and the CNL category hierarchy. Off-line, before selects the topic(s) with matching key- the system starts, the classification procedure is per- words or the greatest number of cate- formed for all the topics: the algorithm puts together gories in common: if multiple, it picks all the sentences associated with each topic (inserted the one closer to the root in the DT or into the Ontology by experts and then automatically randomly if they are at the same level. composed, or added by the users during the conversa- · Otherwise, the algorithm returns a ran- tion through a mechanism not described in this article dom topic immediately below the root. [52]), sends them to CNL, and associates the returned This second method is more complex and compu- categories to the corresponding topic in the Ontology tationally expensive than the previous one (and, in the and then in the DT (notice that, whilst keywords are current implementation, it requires third-party services), manually encoded in the Ontology, categories are auto- but it provides more stability and consistency when se- matically assigned). On-line, during the conversation, lecting the next conversation topic. Also, it diminishes this mapping will allow the system to find which topics the chances of performing random changes of conver- of the Ontology match best (i.e., have more categories sation topic when at the bottom of the DT and no in common) with the category (if any) of the sentence keywords are found: if the sentence has at least one pronounced by the user according to CNL. In the cur- 11 https://cloud.google.com/natural-language/docs/ 10 https://cloud.google.com/natural-language?hl=en categories Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 13 category associated, recognized by CNL, there is still a 2. The system exploiting the keyword and category- chance of finding a coherent topic to jump to in order based Dialogue Management algorithm; to continue the conversation. 3. A system that chooses the next topic randomly (i.e., regardless of what the user says); 4. A human pretending to be a chatbot (i.e., in a Turing- 4 Materials and Methods test fashion); 5. Replika. Experimental tests with recruited participants have been It shall be reminded that the purpose of the exper- performed for multiple purposes: iments is to test the impact of the Dialogue Manage- – Evaluate the impact of different solutions for Dia- ment solutions we proposed, which emphasize the prob- logue Management we developed, either keyword- lem of switching in-between different topics: Replika is based or both keyword- and category-based, to im- expected to perform better than systems 1, 2, and 3 prove the subjective perception of the system in in generating individual sentences, and this is true in terms of coherence and user satisfaction; the case of a human pretending to be a chatbot. Then, – Compare our solutions with Replika (one of the most the purpose of the comparison is to show that, even advanced commercial social chatbots) as well as with in presence of a more advanced approach for produc- baselines consisting of a system choosing replies ran- ing individual sentences, still our solution for Dialogue domly, and with a human pretending to be a chat- Management may have a significantly positive impact bot. on user evaluation. For reasons due to the Covid-19 pandemic, partic- Among the most famous chatbots mentioned in Sec- ipants could not take the tests with a robot, but they tion 2.3 we chose to use Replika for our experiment as had to interact with their assigned conversational sys- 12 it is available as Android and IOS application, and tem (1, 2, 3, 4, or 5) using their computer, by establish- in a web version. Moreover, as already mentioned, it is ing a connection to a remote server in our laboratory. It 13 backed by Open AI’s sophisticated GPT-2 : a trans- is crucial to mention that the text-based user interface former machine learning model that uses deep learning was identical for all the systems: therefore, participants to generate text output that has been claimed to be assigned group 4 were not aware that they were inter- indistinguishable from that of humans [45]. acting with a human, writing at a terminal, and par- ticipants assigned to group 5 were not aware that they were interacting with Replika (with a human manually 4.1 Participants and Test Groups copying and pasting sentences from Replika to the ter- minal and vice-versa). The participants for this test have been recruited through In total, each session included 20 exchanges (i.e., posts on Social Networks (Facebook and Twitter) as a pair of utterances) between the participant and the well as in Robotics and Computer Science classes at the system. University of Genoa, for a total of 100 volunteer par- ticipants aged between 25 and 65. The only inclusion criteria are that participants shall be able to read and 4.2 Questionnaire write in English and to use a PC with a network con- nection from home. Since the test is anonymous and no Immediately after the conversation session, participants personal data is collected, the whole procedure is fully were asked to fill in a questionnaire, divided into two compliant with the EU General Data Protection Regu- parts. lation. Each time a new participant is recruited, he/she The first section of the questionnaire required to is randomly assigned to a group from 1 to 5 (where each evaluate what we defined as the Coherence of system’s group corresponds to a different conversation system), replies (Section 4.2.1 below), for a total of 20 replies, and given the instructions to perform the test. Based individually scored, in a 7-point Likert scale, where 1 on the group they were assigned, participants had to means not Coherent and 7 means Coherent. The sec- interact with: ond part consisted of the SASSI questionnaire (Section 4.2.2). 1. The system exploiting the keyword-based Dialogue Management algorithm; 4.2.1 Coherence Measure 12 Replika’s Android application has been downloaded more than 5,000,000 times from the Google Play Store. The instructions provided to each participant clarified 13 https://en.wikipedia.org/wiki/GPT-2 what is meant by “evaluating the Coherence”. The aim 14 Lucrezia Grassi et al. is to have a way to determine whether the replies of the chatbot are semantically consistent with what the user says and/or with what the chatbot itself has previously said. Hence, if the user believes that the reply of the chatbot is perfectly consistent with what has just been said, then the user is instructed to assign 7 to that re- ply. Otherwise, if the reply of the chatbot has nothing to share with what has just been said in the previous ex- change, such a reply should be evaluated with 1. Values in between shall be assigned to replies that are loosely consistent with the current conversation topic or the new topic raised by the user. The average Coherence score assigned by a partici- pant is computed over all replies; the average Coherence score of a system (i.e., 1, 2, 3, 4, or 5) is computed over all participants that interacted with it.

4.2.2 SASSI Questionnaire

We decided to use this widely known tool to measure the user experience during the conversation, after per- forming some research and comparing it with other tools commonly used for this purpose [28]. Figure 6 reports the Subjective Assessment of Speech System Interfaces (SASSI) questionnaire. The published version has 34 items distributed across six scales, each item being scored with a 7-point Likert scale from Strongly Disagree to Strongly Agree: Fig. 6 The Subjective Assessment of Speech System Inter- 1. System Response Accuracy (9 items): it refers to faces (SASSI) questionnaire. the user’s perceptions of the system as accurate and therefore doing what they expect. This will relate to the system’s ability to correctly recognise the speech scale, and thoroughly revised in the process [53] to more input, correctly interpret the meaning of the utter- firmly establish its psychometric properties. However, ance, and then act appropriately; to the best of our knowledge, no other validated tools 2. Likeability (9 items): it refers to the user’s ratings exist to capture the constructs that we want to mea- of the system as useful, pleasant, and friendly; sure: then, to strengthen the results provided by the 3. Cognitive Demand (5 items): it refers to the per- SASSI, we introduced the aforementioned Coherence ceived amount of effort needed to interact with the measure, whose correlation with the SASSI scales will system and the feelings resulting from this effort; be presented later in this Section, by computing Pear- 4. Annoyance (5 items): it refers to the extent to which son’s correlation index based on the collected data. users rate the system as repetitive, boring, irritat- ing, and frustrating; 5. Habitability (4 items): it refers to the extent to which 4.3 Data Collection the user knows what to do and knows what the sys- tem is doing; In total, 100 × 20 = 2000 exchanges between the five 6. Speed (2 items): it refers to how quickly the system systems and the participants (400 per group), individ- responds to user inputs. ually scored for their Coherence, have been collected, as well as 100 completed SASSI questionnaires (20 per Controversial opinions about the usage of the SASSI group). To analyze the results of this experiment, an can be found in [22]. It has been pointed out that, be- Excel file containing three sheets has been created. fore being considered as a standard for measuring the quality of speech system interfaces, the SASSI ques- – The first sheet contains all the answers to the first tionnaire should be piloted and validated at a broader part of the questionnaire, concerning the Coherence Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 15

of the replies, divided per group (i.e., 1, 2, 3, 4, and 5); – The second sheet contains all the answers to the second part of the questionnaire, divided per group (i.e., 1, 2, 3, 4, and 5) and per SASSI scale. For the Accuracy, Habitability and Speed scales, the scores assigned to negative statements (2, 3, 4, 5, 29, 31, 32, and 34) have been inverted; while for the Cognitive Demand, as a higher score has a negative meaning, Fig. 7 Histogram reporting the average Coherence and the we inverted the scores assigned to positive state- standard deviation of each group. ments (19, 21, 23); – The third sheet contains a statistical analysis of 5 Results all data. All datasets have been tested for normal- ity both with the Shapiro-Wilk Test and with cri- This section presents the results obtained for the con- teria based on the descriptive statistics (data are sidered seven items: the Coherence and the six scales normally distributed if the absolute values of both of the SASSI. First, we computed the mean value and Skewness and Kurtosis are ≤ 1). the standard deviation for each group, with the aim of comparing them two by two, making the null hy- The internal consistency of SASSI data has been pothesis that the compared groups are equal. To ver- checked by computing the Cronbach’s alpha for all scales: ify whether such hypothesis shall be rejected, we per- for Accuracy, alpha=0.91; for Likeability, alpha=0.91; formed the Shapiro-Wilk test for normality on each of for Cognitive Demand, alpha=0.60, for Annoyance, al- the five datasets: if the distribution was not normal, we pha=0.81; for Habitability, alpha=0.64; for Speed, al- performed only the Mann-Whitney U test, while if the pha=0.92. Most scales/groups present a reliability of distribution was normal, we performed both the Mann- more than 0.80, ranging from good to excellent. How- Whitney U test and the Welch’s t-test. In case both ever, Cognitive Demand and Habitability present a reli- tests were performed, we compared the p-values to ver- ability in the range of 0.6-0.7, which is considered ques- ify if the outcomes were consistent. tionable. We hypothesize that low alpha values can be For each item, we report a histogram with the mean due to the fact that we did not recruit only native En- values and the standard deviations, and a table with glish speakers for our experiments, since our inclusion the U-value and U-critical (corresponding to p = 0.05 criteria only required that participants could read and for the Mann-Whitney U test), and the p-values (com- write in English: the fact that Cognitive Demand and puted with both tests, when appropriate). The green Habitability have some negative and/or complex ques- cells highlight the cases when the null hypothesis is re- tions may have had an impact on the internal consis- jected, with p < 0.05, i.e., there is a significant dif- tency of the collected scores. Results related to such ference between the groups (in many cases, we found scales are reported in Section 5, even if they should be p < 0.01). taken cum grano salis. Figures 7, 9, 11, 13, 15, 17, and 19, report the his- The average scores of every group/scale are com- tograms representing the average and the standard de- puted and pairwise compared. To assess whether there viation for each group respectively for the seven items, is a significant difference between the datasets, the Mann- i.e., Coherence, Accuracy, Likeability, Cognitive Demand, Whitney U test has been used: this test is an alternative Annoyance and Speed. Please notice that, in contrast to the t-test when data are not normally distributed. In with the other scales of the SASSI, a higher value of case the comparison is performed among two normally Cognitive Demand represents a negative aspect. distributed samples, the Welch’s t-test has been addi- Figures 8, 10, 12, 14, 16, 18, and 20, present the ta- tionally performed. bles containing the values computed through the Mann- Whitney U test and Welch’s t-test. Note that, when The Ontology used by systems 1, 2, and 3, including both statistical tests are performed, the p-values are all concepts as well as related sentences and keywords, always consistent. the resulting DT with topics of conversation, the map- ping between CNL categories and topics in the DT, and finally Excel files with individual replies to question- 5.1 Correlation between Coherence and SASSI scales naires as well as data analysis are openly available14 The correlation between Coherence measures and each 14 http://caressesrobot.org/IJSORO2021/ SASSI scale has been computed by considering the av- 16 Lucrezia Grassi et al.

Fig. 8 Pairwise statistical comparison of Coherence for dif- Fig. 12 Pairwise statistical comparison of Likeability for dif- ferent groups. ferent groups.

Fig. 9 Histogram reporting the average Accuracy and the Fig. 13 Histogram reporting the average Cognitive Demand standard deviation of each group. and the standard deviation of each group.

Fig. 10 Pairwise statistical comparison of Accuracy for dif- Fig. 14 Pairwise statistical comparison of Cognitive De- ferent groups. mand for different groups.

Fig. 11 Histogram reporting the average Likeability and the Fig. 15 Histogram reporting the average Annoyance and the standard deviation of each group. standard deviation of each group. erage score totalled by each participant (yielding 100 Demand r=-0.45; with Annoyance r=-0.51; with Hab- values for Coherence and 100 for each SASSI scale), itability r=0.56; with Speed r=-0.24. As expected, a and computing the Pearson’s correlation index by pair- high positive correlation is reported between Coherence ing Coherence measures with the corresponding scores and Accuracy and, to a minor extent, with Likeabil- of each SASSI scale. The value of the correlation found ity and Habitability. A weaker, negative correlation is when comparing Coherence with Accuracy turns out found with Cognitive Demand and Annoyance. An eas- to be r=0.78; with Likeability r=0.69; with Cognitive ily explainable small negative correlation with Speed Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 17

Fig. 16 Pairwise statistical comparison of Annoyance for dif- Fig. 20 Pairwise statistical comparison of Speed for different ferent groups. groups.

6 Discussion

This section highlights and examines the major find- ings.

6.1 Coherence

Fig. 17 Histogram reporting the average Habitability and the standard deviation of each group. Figure 7 shows that the standard deviations of groups 1, 2, and 5 are very similar and not too big: this means that, in general, the evaluations regarding the Coher- ence were quite homogeneous. Looking at the standard deviation of group 3, it is clear that the scores assigned by participants to its Coherence were more dissimilar. This may be because the random answers provided by the system during this test are more coherent during some conversations rather than others, or because some participants may be biased to interpret sentences to re- establish coherence with the context. Eventually, the standard deviation of group 4 is the smallest one: par- Fig. 18 Pairwise statistical comparison of Habitability for ticipants, without knowing that they were interacting different groups. with a real human, gave a homogeneously positive eval- uation of the Coherence of the system’s replies. From Figure 8, and by looking at the averages of the Coherence reported in the histogram in Figure 7, we can state that, as expected, group 3 has the low- est Coherence (with a significant difference with all the others), and group 4 has the highest Coherence (again with a significant difference from the others). The sta- tistical analysis revealed a significant difference between groups 1 and 2 and between groups 2 and 5. Fig. 19 Histogram reporting the average Speed and the stan- Results confirm that the additional cost for category dard deviation of each group. extraction and matching (2) positively impacts the per- ceived Coherence with respect to using keywords only (1), and it is sufficient for beating Replika (5). is reported: system 4 (i.e., a human pretending to be a chatbot) obtains the highest Coherence scores but needs a longer time to reply, whereas system 3 (i.e., 6.2 Accuracy the one that chooses topics randomly) obtains the low- est Coherence scores but has no delays when answering Figure 9 shows that the standard deviations of the Ac- since no reasoning algorithms are involved. curacy corresponding to different groups are very sim- 18 Lucrezia Grassi et al. ilar, except for the one of group 4 which is lower than tems require the same amount of effort during the in- the others: this indicates that the participants evalu- teraction, except the one providing random replies (3). ated more homogeneously the Accuracy of this “sys- A valuable result is that both systems we developed tem” (the human). (1 and 2) present no significant difference, in terms of As for Coherence, the random system (3) is the Cognitive Demand, with respect to a human (4) and worst one. An interesting result is that the system ex- Replika (5). ploiting the keyword and category-based Dialogue Man- agement algorithm (2) is the only one as accurate as a human (4) (no statistically significant difference found), 6.5 Annoyance other than being remarkably more accurate than all the other systems. As shown in Figure 15, the standard deviations are quite high and similar to one another. Again, the lowest standard deviation corresponds to group 4: when par- 6.3 Likeability ticipants unknowingly interacted with a human, their ratings were more homogeneous. Considering the aver- The standard deviations shown in Figure 11 are very ages, as we could expect, the system used with group similar; group 3 reports a slightly higher standard de- 3 appears to be the most annoying, while the lowest viation, which indicates that participants had different Annoyance score is associated with group 4. opinions regarding the utility and the pleasantness of Examining the table in Figure 16 and the histogram the systems. As happened for the Coherence (Section in Figure 15, we can conclude that system 2 (keywords 4.2.1), the higher standard deviation corresponding to plus categories) is significantly less annoying than 1 group 3 may be related to chance, as well as to biases (keywords only), and it is as annoying as 4 (human) of participants to favourably interpret random replies and 5 (Replika). Notice also that 1 (keywords only) is to re-establish missing coherence. significantly more annoying than both 4 (human) and As it could be expected, from the histograms it is 5 (Replika). immediate to notice that group 3 has the lowest Like- ability. As regards the other groups, the average Like- ability scores present less prominent differences. 6.6 Habitability Considering both the table in Figure 12 and the histogram in Figure 11, the most interesting result is From the histograms in Figure 17 it can be seen that that the system that exploits the keyword and category- the standard deviations regarding the Habitability are based Dialogue Management algorithm (2), is signifi- very similar for all groups, with the smallest one being cantly more likeable than the one exploiting keywords that of group 3. only (1) and Replika (5). Moreover, the Likeability of Regarding the averages, the smallest one is that cor- (2) is similar to that of a human (4), while this is not responding to group 3, which presents also the lowest true when comparing the human with the system using average Habitability. This result was expected, as the keywords only (1), fully justifying the additional cost random system had also the lowest Accuracy, Likeabil- for category extraction and matching. ity, and Annoyance, and the highest Cognitive Demand, whose scores are likely not completely independent of one another. 6.4 Cognitive Demand Considering the p-values in Figure 18, and the av- erages in Figure 17, it can be observed that the system The standard deviations in Figure 13 are again very used with group 2 (keywords plus categories) turned similar; group 4 reports a slightly lower standard devi- out to be more habitable than 5 (Replika), while 1 (key- ation which, as always, indicates that the participants’ words only) is less habitable than 5. Very interestingly, evaluation regarding the Cognitive Demand of the sys- both our systems 1 and 2 are not distinguishable from tem is homogeneous. 4 (human) in Habitability. As regards the averages, group 3 deviates from the others. This means that a higher cognitive effort is needed to interact with the system: this is since the 6.7 Speed replies of the system are random, hence the participants found it more difficult to easily interact with it. As it can be seen in Figure 19, the system tested with By looking at the p-values in Figure 14 and the av- group 3 (random replies) is faster than 1 and 2, since erages in Figure 13, we can conclude that all the sys- they are all connected to the CARESSES Cloud, but Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 19 the former does not call the Dialogue Management al- The approach presented in this article has obvious gorithm. However, the standard deviation of group 3 is limitations since it relies on an Ontology of concepts slightly higher than those of groups 1 and 2, even if the and the related sentences to talk about such concepts, average is lower: we conjecture that the extremely low which needs to be manually encoded by experts. This pleasantness of the overall interaction with this system, issue has been addressed in subsequent work, which ex- confirmed by all the results of the previous sections, plored strategies for expanding the knowledge base at negatively influences the perception of Speed. run-time during the interaction with the user [52]. Groups 4 and 5 report a higher standard deviation and a lower average. However, these results cannot be compared with those of the other tests. Since group 4 Conflict of Interest involves the unaware interaction with a human, the final score depends on the typing speed of the human. The The authors declare that they have no conflict of inter- same reasoning applies to group 5, which involved the est. unaware interaction with Replika: a human acted as an intermediary between the participant and Replika, by manually typing Replika’s replies which required some References additional time. 1. T. Belpaeme, J. Kennedy, A. Ramachandran, B. Scas- Considering both the table in Figure 20, and the sellati, and F. Tanaka. Social robots for education: A histogram in Figure 19, the only relevant result is that review. Sci Robot, 3(21), 2018. there are no significant Speed differences between 1, 2, 2. Y. Pan, H. Okada, T. Uchiyama, and K. Suzuki. On the reaction to robot’s speech in a hotel public space. Int J and 3. No conclusions shall be drawn when comparing Soc Robot, 7:911–920, 2015. groups 4 and 5. 3. A. K. Pandey and R. Gelin. A mass-produced sociable : Pepper: the first machine of its kind. IEEE Robot Autom Mag, 25(3):40–48, 2018. 4. M. Niemel¨a,P. Heikkil¨a,H. Lammi, and V. Oksman. A 7 Conclusion social robot in a shopping mall: Studies on acceptance and stakeholder expectations. Social Robots: Technologi- The article proposes a novel system for knowledge-based cal, Societal and Ethical Aspects of Human-Robot Inter- action. Springer, pages 119–144, 2019. conversation designed for Social Robots and other con- 5. J. Broekens, M. Heerink, and H. Rosendal. Assistive so- versational agents. The proposed system relies on an cial robots in elderly care: A review. Gerontechnology, Ontology for the description of all concepts that may 8(2):94–103, 2009. be relevant conversation topics, as well as their mu- 6. C. Papadopoulos, T. Hill, L. Battistuzzi, N. Castro, A. Nigath, G. Randhawa, L. Merton, S. Kanoria, tual relationships. We compare two algorithms, based H. Kamide, N.-Y. Chong, D. Hewson, R. Davidson, and on the Ontology, for Dialogue Management that select A. Sgorbissa. The CARESSES study protocol: test- the most appropriate conversation topics depending on ing and evaluating culturally competent socially assis- the user’s input: the two versions differ in their com- tive robots among older adults residing in long term care homes through a controlled experimental trial. Arch. putational cost and/or the need for third-party NLP Public Health, 78(1):1–20, 2020. services. Moreover, they implement slightly different 7. C. Papadopoulos, N. Castro, A. Nigath, R. Davidson, strategies to ensure a conversation flow that captures N. Faulkes, R. Menicatti, A. A. Khaliq, C. T. Recchi- the user’s intention to drive the conversation in specific uto, L. Battistuzzi, G. Randhawa, L. Merton, S. Kanoria, N. Y. Chong, K. Kamide, D. Hewson, and A. Sgorbissa. directions, while avoiding purely reactive responses. The CARESSES randomised controlled trial: exploring Experiments performed with 100 volunteer partici- the health-related impact of culturally competent artifi- pants, interacting with five different conversational sys- cial intelligence embedded into socially assistive robots tems (one of which is a human pretending to be a chat- and tested in older adult care homes. Int J Soc Robot, 2021. bot in a Turing-test fashion), support our intuitions 8. R. D. Adelman, L. L. Tmanova, D. Delgado, S. Dion, and about the importance of dialogue flow management, the M. S. Lachs. Caregiver burden: a clinical review. JAMA, improvements brought by the proposed solution based 311(10):1052–1060, 2014. on the semantic category of the user’s sentence, as well 9. G.-Z. Yang, B.J. Nelson, R.R. Murphy, H. Choset, H. Christensen, S.H. Collins, P. Dario, K. Goldberg, as the positive correlations between Coherence in flow K. Ikuta, N. Jacobstein, D. Kragic, R.H. Taylor, and management and some scales of the SASSI (Subjective M. McNutt. Combating COVID-19 — the role of robotics Assessment of Speech System Interfaces) questionnaire. in managing public health and infectious diseases. Sci Specifically, one of the proposed algorithms is statisti- Robot, 5(40), 2020. 10. N. Zhu, D. Zhang, W. Wang, X. Li, B. Yang, J. Song, cally superior (and never inferior), in all items, to Rep- X. Zhao, B. Huang, W. Shi, R. Lu, P. Niu, F. Zhan, lika: one of the most popular chatbots worldwide. X. Ma, D. Wang, W. Xu, G. Wu, G. F. Gao, W. Tan, 20 Lucrezia Grassi et al.

and China Novel Coronavirus Investigating and Research 28. A. B. Kocaballi, L. Laranjo, and E. Coiera. Measuring Team. A novel coronavirus from patients with pneumonia user experience in conversational interfaces: A compari- in china. New Eng J Med, 328(8):727–733, 2020. son of six questionnaires. Proceedings of the 32nd Inter- 11. G. Miller. Social distancing prevents infections, but it national BCS Human Computer Interaction Conference can have unintended consequences. Science, 2020. (HCI), 2018. 12. R. M. Schuetzler, G. M. Grimes, and J. S. Giboney. The 29. D. Adiwardana, M.-T. Luong, D. R. So, J. Hall, N. Fiedel, impact of chatbot conversational skill on engagement and R. Thoppilan, Z. Yang, A. Kulshreshtha, G. Nemade, perceived humanness. JIMS, 37(3):875–900, 2020. Y. Lu, and Q. V. Le. Towards a human-like open-domain 13. B. Bruno, N.Y. Chong, H. Kamide, S. Kanoria, J. Lee, chatbot. arXiv preprint arXiv:2001.09977, 2020. Y. Lim, A.K. Pandey, C. Papadopoulos, I. Papadopoulos, 30. S. Roller, E. Dinan, N. Goyal, D. Ju, M. Williamson, F. Pecora, A. Saffiotti, and A. Sgorbissa. Paving the way Y. Liu, J. Xu, M. Ott, K. Shuster, Y.-L. Boureau for culturally competent robots: a position paper. IEEE E. M. Smith, and J. Weston. Recipes for building an International Symposium on Robot and Human Interac- open-domain chatbot. arXiv preprint arXiv:2004.13637, tive Communication (RO-MAN), pages 553–560, 2017. 2020. 14. C. T. Recchiuto and A. Sgorbissa. A feasibility study 31. A. Venkatesh, C. Khatri, A. Ram, F. Guo, R. Gabriel, of culture-aware cloud services for conversational robots. A. Nagar, R. Prasad, M. Cheng, B. Hedayatnia, A. Met- IEEE Robot Autom Lett, 5(4):6559–6566, 2020. allinou, R. Goel, S. Yang, and A. Raju. On evaluat- 15. C. T. Recchiuto, L. Gava, L. Grassi, A. Grillo, M. Lago- ing and comparing conversational agents. arXiv preprint marsino, D. Lanza, Z. Liu, C. Papadopoulos, I. Pa- arXiv:1801.03625, 2018. padopoulos, A. Scalmato, and A. Sgorbissa. Cloud ser- 32. A. Cervone and G. Riccardi. Is this dialogue coherent? vices for culture aware conversation: Socially assistive learning from dialogue acts and entities. Proceedings of robots and virtual assistants. UR, pages 270–277, 2020. the 21th Annual Meeting of the Special Interest Group 16. J. Ware and B. Gandek. Overview of the sf-36 health on Discourse and Dialogue, 2020. survey and the international quality of life assessment 33. R. Higashinaka, T. Meguro, K. Imamura, H. Sugiyama, (iqola) project. J Clin Epidemiol, 51(11):903–12, 1998. T. Makino, and Y. Matsuo. Evaluating coherence in open domain conversational systems. Fifteenth Annual Con- 17. T. Nomura, T. Kanda, and T. Suzuki. Experimental ference of the International Speech Communication As- investigation into influence of negative attitudes toward sociation, 2014. robots on human-robot interaction. AI Soc, 20:138–150, 34. K. Nimavat and T. Champaneria. Chatbots: an overview 2006. types, architecture, tools and future possibilities. Int 18. R. Hays and M. R. DiMatteo. A short-form measure of J Scientific Research & Development, 5(7):1019–1024, loneliness. J Pers Assess, 51(1):69–81, 1987. 2017. 19. N. Mavridis. A review of verbal and non-verbal human- 35. G. S. D’silva, S. Thakare, S. More, and J. Kuriakose. Real robot interactive communication. Rob Auton Syst, 63:2– world smart chatbot for customer care using a software 35, 2015. as a service (saas) architecture. IEEE Int Conf on IoT 20. B. Bruno, C. T. Recchiuto, I. Papadopoulos, A. Saffiotti, in Social, Mobile, Analytics and Cloud (I-SMAC), pages C. Koulouglioti, R. Menicatti, F. Mastrogiovanni, R. Za- 658–664, 2017. ccaria, and A. Sgorbissa. Knowledge representation for 36. M. Nuruzzaman and O. K. Hussain. A survey on chat- culturally competent personal robots: Requirements, de- bot implementation in customer service industry through sign principles, implementation, and assessment. Int J deep neural networks. Paper presented at the 15th Int Soc Robot, 11:515–538, 2019. Conf on e-Business Engineering (ICEBE), pages 54–61, 21. J. R. Lewis and M. L. Hardzinski. Investigating the psy- 2018. chometric properties of the speech user interface service 37. L. Zhou, J. Gao, D. Li, and H.-Y. Shum. The design and quality questionnaire. Int J Speech Tech, 18:479–487, implementation of xiaoice, an empathetic social chatbot. 2015. Comput Linguist, 46(1):53–93, 2020. 22. J. R. Lewis. Standardized questionnaires for voice inter- 38. H.-Y. Shum, X. He, and D. Li. From eliza to xiaoice: action design. AVIxD, 1(1), 2016. Challenges and opportunities with social chatbots. Front 23. E. Dinan, S. Rollerand, K. Shuster, A. Fan, M. Auli, and Inf Technol Electron Eng, 19:10–26, 2018. J. Weston. Wizard of Wikipedia: Knowledge-powered 39. J. Weizenbaum. Eliza - a computer program for the conversational agents. Int Conf on Learning Represen- study of natural language communication between man tations (ICLR), 2018. and machine. Commun ACM, 9(1), 1966. 24. J. Huang, M. Zhou, and D. Yang. Extracting chatbot 40. K. Colby. Artificial paranoia: A computer simulation of knowledge from online discussion forums. Paper pre- paranoid processes. Elsevier, 1975. sented at the 20th Int J Conf on AI (IJCAI’07), pages 41. R. S. Wallace. The anatomy of a.l.i.c.e. parsing the turing 423–428, 2007. test. Springer, pages 181–210, 2009. 25. S. Kim O.-W. Kwon and H. Kim. Knowledge-grounded 42. A. Turing. Computing machinery and intelligence. Mind, chatbot based on dual wasserstein generative adversarial 59(236):433–460, 1950. networks with effective attention mechanisms. Appl Sci, 43. C. Dilmegani. Top 30 successful chatbots of 2021 & rea- 10(9), 2020. sons for their success. https://research.aimultiple. 26. M. Ghazvininejad, C. Brockett, M.-W. Chang, B. Dolan, com/top-chatbot-success/#bots-for-general-use. J. Gao, W.-T. Yih, and M. Galley. A knowledge- Accessed: 2021-04-24. grounded neural conversation model. Paper presented 44. A. Pardes. The emotional chatbots are here to at the Thirty-Second AAAI Conf. on AI, 2018. probe our feelings. https://www.wired.com/story/ 27. A. Ritter, C. Cherry, and W. B. Dolan. Data-driven re- replika-open-source/. Accessed: 2021-04-24. sponse generation in social media. EMNLP, pages 583– 45. A. Alleyne. Chat bots are becoming uncannily human. 593, 2011. can they be our friends? https://edition.cnn.com/ Knowledge-Grounded Dialogue Flow Management for Social Robots and Conversational Agents 21

style/article/tech-loneliness-replika-wellness/ competence. He has also been the local Coordinator for index.html. Accessed: 2021-04-24. the BrainHuRo project, developing Brain-Computer In- 46. J. Weston, E. Dinan, and A. H. Miller. Retrieve and re- terfaces for humanoid robots’ remote control. He is the fine: Improved sequence generation models for dialogue. Proceedings of the 2018 EMNLP Workshop SCAI: The author of more than 40 scientific papers published in 2nd International Workshop on Search-Oriented Con- International Journals and conference proceedings. versational AI, 2018. 47. Y. Zhang, S. Sun, M. Galley, Y.-C. Chen, C. Brockett, X. Gao, J. Gao, J. Liu, and B. Dolan. DIALOGPT: Large-scale generative pre-training for conversational re- Antonio Sgorbissa (Ph.D.) is Associate Professor sponse generation. Proceedings of the 58th Annual Meet- at the University of Genoa, where he teaches Real-Time ing of the Association for Computational Linguistics: Operating Systems, Social Robotics, and Ambient In- System Demonstrations, 2020. 48. E. Dinan, S. Humeau, J. Weston, and B. Chintagunta. telligence in EMARO+, the European Master in Ad- Build it break it fix it for dialogue safety: Robustness vanced Robotics. He is the Coordinator of National and from adversarial human attack. Proceedings of EMNLP- EU research projects, among which the H2020 project IJCNLP 2019 - 2019 Conference on Empirical Methods CARESSES (caressesrobot.org). Also, he is the lo- in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, pages cal Coordinator of the ongoing IENE-10 project, aimed 4537–4546, 2020. at preparing health and social care workers to work 49. N. Guarino. Formal ontology and information systems. with intelligent robots in health and social care envi- FOIS’98 Conf, pages 81–97, 1998. ronments. He received a number of acknowledgements 50. B. Motik, P. F. Patel-Schneider, and B. Parsia. Owl 2 web ontology language structural specification and functional- for his recent work: among the others, CARESSES has style syntax. WWW Consortium, 2008. been acknowledged ”Project of the month” by the EU, 51. M. Carrithers, M. Candea, K. Sykes, M. Holbraad, and its technologies have been acknowledged by the EU S. Venkatesan. Ontology is just another word for cul- Innovation Radar, and the project has been included ture: Motion tabled at the 2008 meeting of the group for debates in anthropological theory. Crit Anthropol, 2010. among the “100 & Automation Sto- 52. L. Grassi, C. T. Recchiuto, and A. Sgorbissa. Knowledge ries” in a report presented by Enel S.p.a. in February triggering, extraction and storage via human-robot verbal 2020. His research focuses on Autonomous Robotic Be- interaction. arXiv preprint arXiv:2104.11170, 2021. haviour, with a focus on Culture-Awareness, Knowledge 53. B. Weiss, I. Wechsung, A. Naumann, and S. M¨oller. Sub- representation, Motion Planning, Wearable, and Ubiq- jective evaluation method for speech-based uni- and mul- timodal applications. Perception in Multimodal Dialogue uitous Robotics. He is the author of about 150 arti- Systems, Springer, pages 285–288, 2008. cles indexed in international databases and has been a member of the Organizing Committee in the top- ranked robotic conference as well as Associate Editor for the International Journal of Advanced Robotic Sys- Lucrezia Grassi (M.D.) is a Ph.D. student in Bio- tems edited by SAGE and Intelligent Service Robotics engineering and Robotics at the University of Genova. by Springer. She got her master’s degree in Robotics Engineering in 2020 with the thesis “A Knowledge-Based Conversa- tion System for Robots and Smart Assistants”, and she is currently pursuing her Ph.D. on multiparty interac- tion between humans and artificial agents. Her interests include Social Robotics, autonomous conversation sys- tems, mixed and virtual reality.

Carmine Tommaso Recchiuto (Ph.D.) is Assis- tant Professor at the University of Genoa, where he teaches Experimental Robotics, ROS programming, and Computer Science. His research interests include Hu- manoid and Social Robotics (with a specific focus on knowledge representation and human-robot interaction), wearable sensors, and Aerial Robotics. He has been the Coordinator of software integration and Head of Soft- ware Development for the CARESSES project, aimed at endowing social robots for older adults with cultural