JournalofMachineLearningResearch13(2012)1333-1371 Submitted 10/09; Revised 6/11; Published 5/12
Transfer in Reinforcement Learning via Shared Features
George Konidaris [email protected] Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology 32 Vassar Street Cambridge MA 02139
Ilya Scheidwasser [email protected] Department of Mathematics Northeastern University 360 Huntington Avenue Boston MA 02115
Andrew G. Barto [email protected] Department of Computer Science University of Massachusetts Amherst 140 Governors Drive Amherst MA 01003
Editor: Ronald Parr
Abstract We present a framework for transfer in reinforcement learning based on the idea that related tasks share some common features, and that transfer can be achieved via those shared features. The framework attempts to capture the notion of tasks that are related but distinct, and provides some insight into when transfer can be usefully applied to a problem sequence and when it cannot. We apply the framework to the knowledge transfer problem, and show that an agent can learn a portable shaping function from experience in a sequence of tasks to significantly improve performance in a later related task, even given a very brief training period. We also apply the framework to skill transfer, to show that agents can learn portable skills across a sequence of tasks that significantly improve performance on later related tasks, approaching the performance of agents given perfectly learned problem-specific skills. Keywords: reinforcement learning, transfer, shaping, skills
1. Introduction One aspect of human problem-solving that remains poorly understood is the ability to appropriately generalize knowledge and skills learned in one task and apply them to improve performance in another. This effective use of prior experience is one of the reasons that humans are effective learners, and is therefore an aspect of human learning that we would like to replicate when designing machine learning algorithms. Although reinforcement learning researchers study algorithms for improving task performance with experience, we do not yet understand how to effectively transfer learned skills and knowledge from one problem setting to another. It is not even clear which problem sequences allow transfer, which do not, and which do not need to. Although the idea behind transfer in reinforcement learning