
Risk Bounds and Rademacher Complexity in Batch Reinforcement Learning Yaqi Duan 1 Chi Jin 2 Zhiyuan Li 3 Abstract algorithms including support vector machines (Cortes & Vapnik, 1995; Suykens & Vandewalle, 1999), boosting (Fre- This paper considers batch Reinforcement Learn- und et al., 1996; Schapire, 1999), as well as many success- ing (RL) with general value function approxima- ful applications in fields such as computer vision (Szeliski, tion. Our study investigates the minimal assump- 2010; Forsyth & Ponce, 2012), speech recognition (Juang tions to reliably estimate/minimize Bellman error, & Rabiner, 1991; Jelinek, 1997), and bioinformatics (Baldi and characterizes the generalization performance et al., 2001). by (local) Rademacher complexities of general function classes, which makes initial steps in Notably, in the area of supervised learning, a considerable bridging the gap between statistical learning the- amount of effort has been spent on obtaining sharp risk ory and batch RL. Concretely, we view the Bell- bounds. These are valuable, for instance, in the problem man error as a surrogate loss for the optimality of model selection—choosing a model of suitable complex- gap, and prove the followings: (1) In double sam- ity. Typically, these risk bounds characterize the excess pling regime, the excess risk of Empirical Risk risk—the suboptimality of the learned function compared Minimizer (ERM) is bounded by the Rademacher to the best function within a given function class, via proper complexity of the function class. (2) In the single complexity measures of that function class. After a long sampling regime, sample-efficient risk minimiza- line of extensive research (Vapnik, 2013; Vapnik & Chervo- tion is not possible without further assumptions, nenkis, 2015; Bartlett et al., 2005; 2006), risk bounds are regardless of algorithms. However, with com- proved under very weak assumptions which do not require pleteness assumptions, the excess risk of FQI and realizability—the prespecified function class contains the a minimax style algorithm can be again bounded ground-truth. The complexity measures for general function by the Rademacher complexity of the correspond- classes have also been developed, including but not limited ing function classes. (3) Fast statistical rates can to metric entropy (Dudley, 1974), VC dimension (Vapnik & be achieved by using tools of local Rademacher Chervonenkis, 2015) and Rademacher complexity (Bartlett complexity. Our analysis covers a wide range of & Mendelson, 2002). (See e.g. (Wainwright, 2019) for a function classes, including finite classes, linear textbook review.) spaces, kernel spaces, sparse linear features, etc. Concurrently, batch reinforcement learning (Lange et al., 2012; Levine et al., 2020)—a branch of Reinforcement Learning (RL) that learns from offline data, has been in- 1. Introduction dependently developed. This paper considers the value Statistical learning theory, since its introduction in the late function approximation setting, where the learning agent 1960’s, has become one of the most important frameworks aims to approximate the optimal value function from a re- in machine learning, to study problems of inference or func- stricted function class that encodes the prior knowledge. tion estimation from a given collection of data (Hastie et al., Batch RL with value function approximation provides an 2009; Vapnik, 2013; James et al., 2013). The development important foundation for the empirical success of modern of statistical learning has led to a series of new popular RL, and leads to the design of many popular algorithms such as DQN (Mnih et al., 2015) and Fitted Q-Iteration with *Equal contribution 1Department of Operations Re- neural networks (Riedmiller, 2005; Fan et al., 2020). search and Financial Engineering, Princeton University 2Department of Electrical and Computer Engineering, Despite being a special case of supervised learning, batch Princeton University 3Department of Computer Science, RL also brings several unique challenges due to the addi- Princeton University. Correspondence to: Yaqi Duan tional requirement of learning the rich temporal structures <[email protected]>, Chi Jin <[email protected]>, Zhiyuan Li <[email protected]>. within the data. Addressing these unique challenges has been the main focus of the field so far (Levine et al., 2020). Proceedings of the 38 th International Conference on Machine Consequently, the field of statistical learning and batch RL Learning, PMLR 139, 2021. Copyright 2021 by the author(s). Risk Bounds and Rademacher Complexity in Batch Reinforcement Learning have been developed relatively in parallel. In contrast to the Farahmand et al., 2016; Le et al., 2019) are canonical and mild assumptions required and the generic function class popular approaches. When using a linear function space, the allowed in classical statistical learning theory, a majority of sample complexity for batch RL is shown to depend on the batch RL results (Munos & Szepesvari´ , 2008; Antos et al., dimension (Lazaric et al., 2012). When it comes to general 2008; Lazaric et al., 2012; Chen & Jiang, 2019) remain function classes, several complexity measures of function under rather strong assumptions which rarely hold in prac- class such as metric entropy and VC dimensions have been tice, and are applicable only to a restricted set of function used to bound the performance of fitted value iteration and classes. This raises a natural question: can we bring the policy iteration (Munos & Szepesvari´ , 2008; Antos et al., rich knowledge in statistical learning theory to advance our 2008; Farahmand et al., 2016). understanding in batch RL? Throughout the existing theoretical studies of batch RL, This paper makes initial steps in bridging the gap between people commonly use concentrability, realizability and com- statistical learning theory and batch RL. We investigate the pleteness assumptions to prove polynomial sample com- minimal assumptions required to reliably estimate or mini- plexity. Chen & Jiang(2019) justify the necessity of low mize the Bellman error, and characterize the generalization concentrability and hold a debate on realizability and com- performance of batch RL algorithms by (local) Rademacher pleteness. Xie & Jiang(2020a) develop an algorithm that complexities of general function classes. Concretely, we only relies on the realizability of optimal Q-function and establish conditions when the Bellman error can be viewed circumvents completeness condition. However, they use as a surrogate loss for the optimality gap in values. We a stronger concentrability assumption and the error bound then bound the excess risk measured in Bellman errors. We has a slower convergence rate. While the analyses in Chen prove the followings: & Jiang(2019) and Xie & Jiang(2020a) are restricted to discrete function classes with a finite number of elements, • In the double sampling regime, the excess risk of a Wang et al.(2020a) investigate value function approxima- simple Empirical Risk Minimizer (ERM) is bounded tion with linear spaces. It is shown that data coverage and by the Rademacher complexity of the function class, realizability conditions are not sufficient for polynomial under almost no assumptions. sample complexity in the linear case. • In the single sampling regime, without further assump- Off-policy evaluation Off-policy evaluation (OPE) refers tions, no algorithm can achieve small excess risk in to the estimation of value function given offline data (Precup, the worse case unless the number of samples scales up 2000; Precup et al., 2001; Xie et al., 2019; Uehara et al., polynomially with respect to the number of states. 2020; Kallus & Uehara, 2020; Yin et al., 2020; Uehara et al., • In the single sampling regime, under additional com- 2021), which can be viewed as a subroutine of batch RL. pleteness assumptions, the excess risks of Fitted Q- Combining OPE with policy improvement leads to policy- Iteration (FQI) algorithm and a minimax style algo- iteration-based or actor-critic algorithms (Dann et al., 2014). rithm can be again bounded by the Rademacher com- OPE is considered as a simpler problem than batch RL and plexity of the corresponding function classes. its analyses cannot directly translate to guarantees in batch RL. • Fast statistical rates can be achieved by using tools of local Rademacher complexity. Online RL RL in online mode is in general a more dif- ficult problem than batch RL. The role of value function Finally, we specialize our generic theory to concrete ex- approximation in online RL remains largely unclear. It re- amples, and show that our analysis covers a wide range quires better tools to measure the capacity of function class of function classes, including finite classes, linear spaces, in an online manner. In the past few years, there are some kernel spaces, sparse linear features, etc. investigations in this direction, including using Bellman rank (Jiang et al., 2017) and Eluder dimension (Wang et al., 1.1. Related Work 2020b) to characterize the hardness of RL problem. We restrict our discussions in this section to the RL results 1.2. Notation under function approximation. For any integer K > 0, let [K] be the collection of Batch RL There exists a stream of literature regarding 1; 2;:::;K. We use 1[·] to denote the indicator function. finite sample guarantees for batch RL with value function For any function q(·) and any measure ρ over the domain of 2 2 approximation. Among the works, fitted value iteration q, we define
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages11 Page
-
File Size-