
Safe Imitation Learning via Fast Bayesian Reward Inference from Preferences Daniel S. Brown 1 Russell Coleman 1 2 Ravi Srinivasan 2 Scott Niekum 1 Abstract demonstrations, it is important for an agent to be able to Bayesian reward learning from demonstrations provide high-confidence bounds on its performance with re- enables rigorous safety and uncertainty analysis spect to the demonstrator; however, while there exists much when performing imitation learning. However, work on high-confidence off-policy evaluation in the rein- Bayesian reward learning methods are typically forcement learning (RL) setting, there has been much less computationally intractable for complex control work on high-confidence policy evaluation in the imitation problems. We propose Bayesian Reward Ex- learning setting, where the reward samples are unavailable. trapolation (Bayesian REX), a highly efficient Prior work on high-confidence policy evaluation for im- Bayesian reward learning algorithm that scales itation learning has used Bayesian inverse reinforcement to high-dimensional imitation learning problems learning (IRL) (Ramachandran & Amir, 2007) to allow an by pre-training a low-dimensional feature encod- agent to reason about reward uncertainty and policy gen- ing via self-supervised tasks and then leveraging eralization error (Brown et al., 2018). However, Bayesian preferences over demonstrations to perform fast IRL is typically intractable for complex problems due to the Bayesian inference. Bayesian REX can learn to need to repeatedly solve an MDP in the inner loop, resulting play Atari games from demonstrations, without in high computational cost as well as high sample cost if a access to the game score and can generate 100,000 model is not available. This precludes robust safety and un- samples from the posterior over reward functions certainty analysis for imitation learning in high-dimensional in only 5 minutes on a personal laptop. Bayesian problems or in problems in which a model of the MDP is REX also results in imitation learning perfor- unavailable. We seek to remedy this problem by propos- mance that is competitive with or better than state- ing and evaluating a method for safe and efficient Bayesian of-the-art methods that only learn point estimates reward learning via preferences over demonstrations. Pref- of the reward function. Finally, Bayesian REX en- erences over trajectories are intuitive for humans to provide ables efficient high-confidence policy evaluation (Akrour et al., 2011; Wilson et al., 2012; Sadigh et al., 2017; without having access to samples of the reward Christiano et al., 2017; Palan et al., 2019) and enable better- function. These high-confidence performance than-demonstrator performance (Brown et al., 2019b;a). To bounds can be used to rank the performance and the best of our knowledge, we are the first to show that risk of a variety of evaluation policies and provide preferences over demonstrations enable both fast Bayesian a way to detect reward hacking behaviors. reward learning in high-dimensional, visual control tasks as well as efficient high-confidence performance bounds. 1. Introduction We first formalize the problem of high-confidence policy evaluation (Thomas et al., 2015) for imitation learning. We arXiv:2002.09089v4 [cs.LG] 17 Dec 2020 It is important that robots and other autonomous agents can then propose a novel algorithm, Bayesian Reward Extrapola- safely learn from and adapt to a variety of human prefer- tion (Bayesian REX), that uses a pairwise ranking likelihood ences and goals. One common way to learn preferences to significantly increase the efficiency of generating samples and goals is via imitation learning, in which an autonomous from the posterior distribution over reward functions. We agent learns how to perform a task by observing demonstra- demonstrate that Bayesian REX can leverage neural network tions of the task (Argall et al., 2009). When learning from function approximation to learn useful reward features via 1 self-supervised learning in order to efficiently perform deep Computer Science Department, The University of Texas Bayesian reward inference from visual demonstrations. Fi- at Austin. 2Applied Research Laboratories, The University of Texas at Austin. Correspondence to: Daniel Brown <ds- nally, we demonstrate that samples obtained from Bayesian [email protected]>. REX can be used to solve the high-confidence policy evalua- tion problem for imitation learning. We evaluate our method th Proceedings of the 37 International Conference on Machine on imitation learning for Atari games and demonstrate that Learning, Online, PMLR 119, 2020. Copyright 2020 by the au- we can efficiently compute high-confidence bounds on pol- thor(s). Safe Imitation Learning via Fast Bayesian Reward Inference from Preferences icy performance, without access to samples of the reward than-demonstrator performance even in the absence of ex- function. We use these high-confidence performance bounds plicit preference labels. However, despite their successes, to rank different evaluation policies according to their risk deep learning approaches to IRL typically only return a point and expected return under the posterior distribution over estimate of the reward function, precluding uncertainty and the unknown ground-truth reward function. Finally, we pro- robustness analysis. vide evidence that bounds on uncertainty and risk provide a useful tool for detecting reward hacking/gaming (Amodei 2.2. Safe Imitation Learning et al., 2016), a common problem in reward inference from demonstrations (Ibarz et al., 2018) as well as reinforcement While there has been much interest in imitation learning, learning (Ng et al., 1999; Leike et al., 2017). less attention has been given to problems related to safety. SafeDAgger (Zhang & Cho, 2017) and EnsembleDAgger (Menda et al., 2019) are extensions of DAgger that give 2. Related work control to the demonstrator in states where the imitation 2.1. Imitation Learning learning policy is predicted to have a large action difference from the demonstrator. Other approaches to safe imitation Imitation learning is the problem of learning a policy from learning seek to match the tail risk of the expert as well as demonstrations and can roughly be divided into techniques find a policy that is indistinguishable from the demonstra- that use behavioral cloning and techniques that use in- tions (Majumdar et al., 2017; Lacotte et al., 2019). verse reinforcement learning. Behavioral cloning methods (Pomerleau, 1991; Torabi et al., 2018) seek to solve the imi- Brown & Niekum(2018) propose a Bayesian sampling tation learning problem via supervised learning, in which approach to provide explicit high-confidence performance the goal is to learn a mapping from states to actions that bounds in the imitation learning setting, but require an MDP mimics the demonstrator. While computationally efficient, solver in the inner-loop. Their method uses samples from the P (RjD) these methods suffer from compounding errors (Ross et al., posterior distribution to compute sample efficient 2011). Methods such as DAgger (Ross et al., 2011) and probabilistic upper bounds on the policy loss of any evalu- DART (Laskey et al., 2017) avoid this problem by repeat- ation policy. Other work considers robust policy optimiza- edly collecting additional state-action pairs from an expert. tion over a distribution of reward functions conditioned on demonstrations or a partially specified reward function, but Inverse reinforcement learning (IRL) methods seek to solve these methods require an MDP solver in the inner loop, lim- the imitation learning problem by estimating the reward iting their scalability (Hadfield-Menell et al., 2017; Brown function that the demonstrator is optimizing (Ng & Russell, et al., 2018; Huang et al., 2018). We extend and generalize 2000). Classical approaches repeatedly alternate between a the work of Brown & Niekum(2018) by demonstrating, reward estimation step and a full policy optimization step for the first time, that high-confidence performance bounds (Abbeel & Ng, 2004; Ziebart et al., 2008; Ramachandran can be efficiently obtained when performing imitation learn- & Amir, 2007). Bayesian IRL (Ramachandran & Amir, ing from high-dimensional visual demonstrations without 2007) samples from the posterior distribution over reward requiring an MDP solver or model during reward inference. functions, whereas other methods seek a single reward func- tion that induces the demonstrator’s feature expectations 2.3. Value Alignment and Active Preference Learning (Abbeel & Ng, 2004), often while also seeking to maximize the entropy of the resulting policy (Ziebart et al., 2008). Safe imitation learning is closely related to the problem of value alignment, which seeks to design methods that prevent Most deep learning approaches for IRL use maximum en- AI systems from acting in ways that violate human values tropy policy optimization and divergence minimization be- (Hadfield-Menell et al., 2016; Fisac et al., 2020). Research tween marginal state-action distributions (Ho & Ermon, has shown that difficulties arise when an agent seeks to align 2016; Fu et al., 2017; Ghasemipour et al., 2019) and are its value with a human who is not perfectly rational (Milli related to Generative Adversarial Networks (Finn et al., et al., 2017) and there are fundamental impossibility results 2016a). These methods scale to complex control problems regarding value alignment unless the objective is represented by iterating between reward learning and
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages24 Page
-
File Size-