<<

Human-Guided Learning of Social Action Selection for Robot-Assisted Therapy

Emmanuel Senft, Paul Baxter, Tony Belpaeme Centre for Robotics and Neural Systems, Cognition Institute Plymouth University, U.K. emmanuel.senft, paul.baxter, tony.belpaeme @plymouth.ac.uk { }

Abstract this early learning of an action selection mechanism for autonomous robot interaction with a human, by tak- This paper presents a method for progres- ing advantage of the expert knowledge of a third-party sively increasing autonomous action selec- human supervisor to prevent the robot from exploring tion capabilities in sensitive environments, in an inappropriate manner. We first present the for- where random exploration-based learning is mal framework in which our action selection strategy not desirable, using guidance provided by a learning takes place (section 2), then illustrate this human supervisor. We describe the global with a case study from the domain of Robot Assisted framework and a simulation case study based Therapy for children with Autism Spectrum Disorder on a scenario in Robot Assisted Therapy (ASD), where the incorrect selection of actions can for children with Autism Spectrum Disorder. lead to an unacceptable impact on the goals of the This simulation illustrates the functional fea- interaction (section 3). tures of our proposed approach, and demon- strates how a system following these princi- State (S) Impact ples adapts to di↵erent interaction contexts Environment

while maintaining an appropriate behaviour Taken into account for learning for the system at all times. Algorithm Update M based A' Supervisor A on history Select A based on Agent C, S, A', K Execute Action A Suggest A' and history 1 Introduction Based on inputs and model M Humans are interacting increasingly with machines, and robots will be progressively more important part- Context ners in the coming years. Human-human interactions Context (C) Impact involve high dimensionality signals and require com- plex processing: this results in a large quantity of data that ideally needs to be processed by an autonomous Figure 1: The supervised online learning of au- robot. One potential solution is the application of ma- tonomous action selection mechanism. chine learning techniques. Specifically, online learning is desirable, however, some level of initial knowledge and competencies are required to avoid pitfalls in the 2 Supervised Emergent Autonomous early phases of the learning process, particularly in Decision Making contexts where random exploration could lead to un- desirable consequences. 2.1 Framework In this paper, we propose an approach inspired by, but separate from, learning from demonstration to guide The situation considered involves a robotic agent, a human supervisor of the agent, and a human with Appearing in Proceedings of the 4th Workshop on Ma- which the agent, but not the supervisor, should in- chine Learning for Interactive Systems (MLIS) 2015, Lille, teract. The agent proposes actions that are accepted France. JMLR: W&CP volume 40. Copyright 2015 by the or rejected by the supervisor prior to executing them. authors. The method proposed in this paper aims at enabling Human-Guided Learning of Social Action Selection for Robot-Assisted Therapy the agent to progressively and autonomously approxi- select a new one according to , and makes the agent T mate the ideal behaviour as specified by the supervisor. execute the resulting action A0. The environment will change to a new state S and the context will be up- Our framework has five components: an agent and an 1 dated to C according to . Based on S , S , C , C , environment interacting with each other, a supervi- 1 1 0 1 0 A , and A , the algorithmE will update its model to sor, the algorithm controlling the agent and a context 0 0 M . The process can then be repeated based on the characterising the interaction between the agent and 1 updated model. the environment. The agent has a defined set of avail- able actions . The environment could be a human, a robot, or a computerA for example and is characterised 2.2 Related Work by a n-dimensional vector n, which is time vary- ing. The context mSgives2R a set of parameters The approach we take here necessarily requires the ap- defining higher-levelC aspects,2R such as goals or the state, plication of , but we do not commit at of the interaction, see figure 1. The supervisor and the this stage to a single algorithmic approach; the specific agent have direct access to the context, but may ignore requirements for our application include online learn- the real value of the state and see it only through ob- ing, deferring to an external supervisor, and being able servations. to handle a dynamic environment. The principal constraints are that the interaction has A widely used method to transfer knowledge from a one or more high level goals, and some available ac- human to a robot is Learning from Demonstration tions can have a negative impact on these goals if ex- (LfD), see [2] for a survey. In the case of policy learn- ecuted in specific states. This should be avoided, so ing, a teacher provides the learning algorithm with cor- algorithms depending on randomness to explore the rect actions for the current state and repeats this state- environment state space are inappropriate. action mapping for enough di↵erent states to give the algorithm a general policy. LfD is usually combined In order to simplify the system, we are making the with : trying directly to map out- following assumptions. Firstly, that the environment, puts and inputs from a teacher, see [12] for a list of while dynamic, is consistent: it follows a defined set of algorithms that can be used in supervised learning. rules which also describe how the context is updated. The other important point is how the demonstrations Secondly,E the supervisor is omniscient (complete T are generated, a first approach is using batch learn- knowledge of the environment), constant (does not ing: the teacher trains the algorithm during a training adapt during the interaction), and coherent (will react phase after which the robot is used in full autonomy with the same action if two sets of inputs are identi- [11]. Or there may be no explicit training phase; us- cal). Finally, the supervisor has some prior knowledge ing online learning the demonstrations are given dur- of the environment . K ing the execution if required: the robot can request The algorithm has a model of the supervisor and a demonstration for the uncertain states, e.g. when the environment and will updateM it through online a confidence value about the action to perform is too learning following the learning method . is it- low [6]. L M eratively updated based on supervisor feedback to ap- Another method is : the algo- proximate and , in this way progressively approxi- T E rithm tries to find a policy maximising the expected mating the action that the supervisor would have cho- reward [3, 10]. However, this implies the presence of sen, and what impact this would have on the environ- a reward function, which may not be trivial to de- ment. Equation 1 describes the update of each part of scribe in domains (such as social interaction) that do the framework from the step n to n + 1. not lend themselves to characterisation. Consequently Abbeel and Ng proposed to use Inverse Reinforcement Learning by using an expert to generate the reward Mn :(C0 n,S0 n,A0 n 1,A0 n 1) An0 ! ! ! ! ! function [1], subsequently extended to use partially- :(C0 n,S0 n,A0 n,A n 1, ) An observable MDPs [8], although expert-generated re- T ! ! ! ! K ! (1) :(C0 n,S0 n,A0 n) (Sn+1,Cn+1) wards also pose problems on the human side [17]. E ! ! ! ! :(C0 n+1,S0 n+1,A0 n,A0 n) Mn+1 The goal of our proposed approach di↵ers from these L ! ! ! ! ! alternative existing methods. The intention is to pro- At the start of the interaction, the environment is in a vide a system that can take advantage of expert hu- state S0 with the context C0 and the algorithm has a man knowledge to progressively improve its competen- model M0. Applying M0 to C0 and S0, the algorithm cies without requiring manual intervention on every will select an action A0 and propose it to the super- interaction cycle. This is achieved by only asking the visor. The supervisor can either accept this action or human supervisor to intervene with corrective infor- Emmanuel Senft, Paul Baxter, Tony Belpaeme mation when the proposed action of the robot agent is child’s state (as provided by therapists for example). deemed inappropriate (e.g. dangerous) prior to actual In principle, while precise determinations are likely to execution. This allows the robot to learn from con- be problematic, we assume that some aspects of these strained exploration; a consequence of this is that the variables can be estimated using a set of sensors (e.g. load on the supervisor should reduce over time as the cameras and RGBD sensors to capture the child’s gaze robot learns. The supervisor nevertheless retains con- and position; the way the child interacts with the touch trol of the robot, and as such we characterise the robot screen; etc). For the remainder of this paper, however, as having supervised autonomy. Contrary to the active we assume that a direct estimation of internal child learning approach used by, for example, Cakmak and states are available to the system. Thomaz [5] the robot is not asking a question and re- quiring a supervisor response, it is proposing an action 3.1 Proof of concept which may or may not be corrected by the supervisor. A minimal simulation was constructed to illustrate the 3 Case Study: Application to Therapy case study described above. The state is defined us- ing three variables: the child’s performance,S engage- One potential application area is Robot Assisted Ther- ment and motivation in the interaction. The robot has the following set of actions : encouragement (give a apy (RAT) for children with Autism Spectrum Disor- A ders (ASD). Children with ASD generally lack typical motivating feedback to the child), waving (perform a social skills, and RAT can help them to acquire these gesture to catch the child’s attention), and proposi- competencies, with a certain degree of success, e.g. tion (inviting the child to make a classification). In this minimal example, the environment is the child [7, 14]. However, these experiments typically use the E Wizard of Oz (WoZ) paradigm [13], which necessitates model. A minimal model of the child was constructed a heavy load on highly trained human operators. that encompassed both processes that were dependent on the robot behaviour (e.g. responding to a request We propose the use of supervised autonomy [15, 16], for action), and processes that were independent of where the robot is primarily autonomous, but the ther- the robot behaviour (e.g. a monotonic decrease of en- apeutic goals are set by a therapist who maintains gagement and motivation over time independently of oversight of the interaction. Having a supervised au- other events). The reaction of the model follows a tonomous robot would reduce the workload on the rule-based system, but the amplitude of the response therapist.Both the therapist and the robot would be is randomly drawn from a normal distribution to rep- present in the interaction, the robot interacting with resent the stochastic aspect of the child’s reaction and the child and the therapist supervising the interaction the potential influence of non-defined variables in the and guiding the robot while it is learning its action state. A number of simplifications are necessary, such selection policy. as the assumption of strict turn-taking and interac- The formalism described above (section 2.1) can be di- tions in discrete time. The minimal child model is rectly applied to this scenario. In this case, the context summarised in figure 2. is the state of the task selected by the therapist to help State: Performance the child develop certain social competencies, for ex- Motivation Engagement ample, a collaborative categorisation game [4] intended Encouragement Waving Proposition to allow the child to practice turn taking or emotion Motivation>0.6 Motivation+=Norm(0.1,0.05) Engagement+=Norm(0.1,0.05) recognition. The state may be defined using multiple Engagement>0.6 variables such as motivation, engagement, and perfor- Yes No mance exhibited by the child during the interaction, Perf += 0.05 Perf -= 0.05 the time elapsed since the last child’s action, and their Motivation = 0.5 last move (correct or incorrect). The robot could have Engagement = 0.5 a set of actions related to the game, such as proposing Motivation -= Norm(0.01,0.001) that the child categorises an image, or giving encour- Engagement -= Norm(0.01,0.001) agement to the child. In this scenario, the goal would be to allow the child to Figure 2: Model of the child used in the minimal sim- improve their performance on the categorisation task, ulation; random numbers are drawn from normal dis- and this would be done by selecting the appropriate tributions. diculty level and finding a way to motivate the child to play the proposed game. We can expect the child to Formally, the minimal simulation follows the frame- react to the robot action and that these reactions can work established above (equation 1), with the simpli- be captured by the di↵erent variables that define the fication that a history of prior states, contexts, and Human-Guided Learning of Social Action Selection for Robot-Assisted Therapy

actions is not used in the learning algorithm. This re- 1.4 Child Performance sults in a setup where the system makes a suggestion 1.2 Child Engagement Child Motivation Threshold of an action to take, which the supervisor can either 1 accept or reject, in which case an alternative action is 0.8 State Values chosen (figure 3). 0.6 0.4 100 105 110 115 120 125 130 135 140 145 150 This allows the supervisor to take a more passive ap- Time Steps

3 proach when the algorithm selects an acceptable action proposition encouragement 2 waving since they will only have to manually select a correc- Intervention tive action when this is needed. If the learning method 1 0 is e↵ective, the number of corrective actions should de- -1

crease over time, decreasing the workload on the ther- Output of Neural Network -2 100 105 110 115 120 125 130 135 140 145 150 apist over the interaction. Time Steps

A = alternate action Figure 4: Subset of an interaction from step 100 to

No 150.

State (S) Does Yes, A = A' Learning M M proposes Agent executes & Supervisor (T) agree using L action A' action A Context (C) with A'? ent models of a child and for a random action selection Change in Environment (E) scheme. The di↵erence in the child models in the three first graphs is the value of the thresholds required for Figure 3: Description of agent’s action selection pro- a good classification action, high reactive child: 0.6 cess: the agent proposes actions that are validated by and 0.6, asymmetrically: 0.9 for encouragement and the supervisor prior to execution. 0.6 for engagement, and low reactive: 0.9 and 0.9. Be- low these thresholds, a proposition would lead to a bad The learning model is a MultiLayer action decreasing the performance. It can be observed (MLP), with three inputM nodes for the input states, that the algorithm learns di↵erent strategies for each three output nodes for the three possible actions and child and that there is more learning apparent at the nine nodes in the hidden . The model is trained start of the interaction than at the end (the rate of using (as ), the true labels are given interventions is decreasing over time), indicating that by the supervisor decision:L output of 1 for the action the system is choosing the appropriate action at the selected by the supervisor (A) and 1 for the other appropriate time, and that the workload on the super- ones. A Winner-Takes-All process is applied on the visor (necessity to provide these corrective actions) de- output of the MLP to select the action suggested by crease over time. The last plot demonstrates a random the robot (A’). action selection with a high reactive child. Contrary Figure 4 shows a subset of a run from step 100 to to the other cases, the child’s performance decreases 150. With this approach, there is no distinct learning over time, and the number of interventions increases. and testing phases, but in the first part of the interac- Here, a bad action only decreases the performance, but tion (before step 100), the supervisor had to produce in reality it may result in the termination of the inter- multiple corrective actions to train the network to ex- action, which must be avoided. press the desired output. The strategy used by the supervisor is the following: if the motivation is lower 4 Discussion than 0.6 the supervisor enforces the action ‘encourage- ment’, else if the engagement is below 0.6 ‘waving’ is While demonstrating promise, there are a number of enforced, and if both are above 0.6 then a proposition limitations to the framework as presented. The as- is made. The first graph presents the evolution of the sumptions described in section 2.1 are typically vio- state over time, and the second one the output of the lated when working with humans. Firstly the children MLP for each action. The vertical red lines represent are all di↵erent, and a method learned for one child an intervention from the supervisor, i.e. a case where may often not be suited when working with another. the supervisor enforces a di↵erent action than the one Furthermore, the same child may have non-consistent suggested by the MLP. The action actually executed behaviour between the sessions and even within a sin- is represented by a cross with the same colour as the gle session. There is no perfect solution to solve this respective curves. problem, but we can expect that with enough training Figure 5 shows a comparison of the cumulative total sessions and a more complex learning algorithm, the of the di↵erent actions suggested and of the interven- system would be able to capture patterns and react to tion as well as the child performance for three di↵er- the di↵erent behaviours appropriately. Since it is ex- Emmanuel Senft, Paul Baxter, Tony Belpaeme

High reactive child Asymmetrically reactive child Low reactive child Random action selection 300 300 300 300 proposition encouragement 4 4 4 4 waving 250 Intervention 250 250 250 Child Performance

2 2 2 2 200 200 200 200

150 0 150 0 150 0 150 0

100 100 100 100 -2 -2 -2 -2 and intervention Child Performance

50 50 50 50

Number of actions suggested -4 -4 -4 -4

0 0 0 0 0 100 200 300 400 0 100 200 300 400 0 100 200 300 400 0 100 200 300 400 Time Steps

Figure 5: Comparison of the cumulative total of the di↵erent actions suggested, the supervisor interventions required, and child performance for three di↵erent models of a child (highly responsive; asymmetrically increased responsiveness to engagement than motivation; low responsiveness), and a random action selection scheme. pected that in a real application of such an approach a posed action: for example increasing the time available therapist who knows the child will always be present, if the confidence in the proposed action is low. This we propose that for a new child the algorithm will use modulation of the load on the supervisor’s attention a generic strategy based on previous interactions with according to confidence should result in the supervisor other children, with subsequent fine-tuning under su- being able to increasingly pay attention to the child pervision. directly, rather than to the robot system, as training progresses. Another assumption that is likely to be violated is that of a perfect supervisor. As explained in [6] humans are not always consistent nor omniscient, but meth- ods presented in the literature can be used to cope 5 Conclusion with these inconsistencies if enough data is gathered. Further mitigating solutions could be employed, such as the robot warning the therapist if it is about to We have presented a general framework to progres- select an action which had negative consequences in sively increase the competence of an autonomous ac- a previous interaction (even if for a di↵erent child). tion selection mechanism that takes advantage of the Furthermore, it may not be possible to measure the expert knowledge of a human supervisor to prevent in- true internal states of the child in the real world, with appropriate behaviour during training. This method imperfect estimations of these states being more likely is particularly applicable to application contexts such accessible. In this case, inspiration from [9] can be as robot-assisted therapy, and our case study has pro- used to mixed the POMDP framework with the help vided preliminary support for the utility of the ap- of an exterior oracle. Another problem which will have proach. While the simulation necessarily only pro- to be addressed in the future is the di↵erence in in- vided a minimal setup, and thus omitted many of the puts between the robot and the therapist: the thera- complexities present in a real-world setup, we have pist will have access to language, more subtle visual nevertheless shown how the proposed method resulted features and their prior experience, whereas the robot in the learning of distinct action selection strategies may have direct and precise access to some aspects of given di↵ering interaction contexts, although further the child’s overt behaviour (such as timings of touch- refinement is required for real-world application. In- screen interaction). deed, given real-world supervisor knowledge limita- tions, we suggest it will furthermore be possible for In the currently implemented case study, we assume a suitably trained action selection mechanism of this that the supervisor responds to the action proposed type to aid the supervisor in complex and highly dy- by the robot within some predetermined fixed time, namic scenarios. whether this response is accept or reject (figure 3). This, in principle, allows the supervisor to only ac- tively respond if a proposed action is clearly inappro- priate. In further developments, we will incorporate Acknowledgements a measure of certainty (given prior experience) into the time allowed the supervisor to respond to the pro- This work is supported by the EU FP7 DREAM project (grant 611391). Human-Guided Learning of Social Action Selection for Robot-Assisted Therapy

References [12] S. B. Kotsiantis, I. Zaharakis, and P. Pintelas. Supervised machine learning: A review of classi- [1] P. Abbeel and A. Y. Ng. Apprenticeship learning fication techniques, 2007. via inverse reinforcement learning. Proceedings of the 21st International Conference on Machine [13] L. Riek. Wizard of Oz Studies in HRI: A Sys- Learning (ICML), pages 1–8, 2004. tematic Review and New Reporting Guidelines. Journal of Human-Robot Interaction, 1(1):119– [2] B. D. Argall, S. Chernova, M. Veloso, and 136, Aug. 2012. B. Browning. A survey of robot learning from demonstration. Robotics and autonomous sys- [14] B. Robins, K. Dautenhahn, R. T. Boekhorst, and tems, 57(5):469–483, 2009. A. Billard. Robotic assistants in therapy and ed- ucation of children with autism: can a small hu- [3] A. G. Barto. Reinforcement learning: An intro- manoid robot help encourage social interaction duction. MIT press, 1998. skills? Universal Access in the Information Soci- ety, 4(2):105–120, July 2005. [4] P. Baxter, R. Wood, and T. Belpaeme. A touchscreen-based sandtrayto facilitate, mediate [15] E. Senft, P. Baxter, J. Kennedy, and T. Bel- and contextualise human-robot social interaction. paeme. When is it better to give up?: Towards In Human-Robot Interaction (HRI), 2012 7th autonomous action selection for robot assisted ACM/IEEE International Conference on, pages asd therapy. In Proceedings of the Tenth Annual 105–106. IEEE, 2012. ACM/IEEE International Conference on Human- Robot Interaction Extended Abstracts, HRI’15 Ex- [5] M. Cakmak and A. L. Thomaz. Designing robot tended Abstracts, pages 197–198, New York, NY, learners that ask good questions. In Proceedings USA, 2015. ACM. of the seventh annual ACM/IEEE international conference on Human-Robot Interaction, pages [16] S. Thill, C. A. Pop, T. Belpaeme, T. Ziemke, 17–24. ACM, 2012. and B. Vanderborght. Robot-assisted therapy for autism spectrum disorders with (partially) au- [6] S. Chernova and M. Veloso. Interactive pol- tonomous control: Challenges and outlook. Pala- icy learning through confidence-based auton- dyn, 3(4):209–217, Apr. 2013. omy. Journal of Artificial Intelligence Research, 34(1):1, 2009. [17] A. L. Thomaz and C. Breazeal. Teachable robots: Understanding human teaching behavior to build [7] K. Dautenhahn. Robots as social actors: Au- more e↵ective robot learners. Artificial Intelli- rora and the case of autism. In Proc. CT99, The gence, 172(6-7):716–737, 2008. Third International Cognitive Technology Confer- ence, August, San Francisco, volume 359, page 374, 1999.

[8] F. Doshi, J. Pineau, and N. Roy. Reinforcement learning with limited reinforcement: Using bayes risk for active learning in pomdps. In Proceedings of the 25th international conference on Machine learning, pages 256–263. ACM, 2008.

[9] F. Doshi, J. Pineau, and N. Roy. Reinforcement learning with limited reinforcement: Using bayes risk for active learning in pomdps. In Proceedings of the 25th international conference on Machine learning, pages 256–263. ACM, 2008.

[10] L. P. Kaelbling, M. L. Littman, and A. W. Moore. Reinforcement learning: A survey. Journal of ar- tificial intelligence research, pages 237–285, 1996.

[11] W. B. Knox, S. Spaulding, and C. Breazeal. Learning social interaction from the wizard: A proposal. In Workshops at the Twenty-Eighth AAAI Conference on Artificial Intelligence, 2014.