A Self-Supervised Terrain Roughness Estimator for Off-Road Autonomous
Total Page:16
File Type:pdf, Size:1020Kb
A Self-Supervised Terrain Roughness Estimator for O®-Road Autonomous Driving David Stavens and Sebastian Thrun Stanford Arti¯cial Intelligence Laboratory Computer Science Department Stanford, CA 94305-9010 fstavens,[email protected] Abstract 1 INTRODUCTION Accurate perception is a principal challenge In robotic autonomous o®-road driving, the primary of autonomous o®-road driving. Percep- perceptual problem is terrain assessment in front of tive technologies generally focus on obsta- the robot. For example, in the 2005 DARPA Grand cle avoidance. However, at high speed, ter- Challenge (DARPA, 2004), a robot competition orga- rain roughness is also important to control nized by the U.S. Government, robots had to identify shock the vehicle experiences. The accuracy drivable surface while avoiding a myriad of obstacles required to detect rough terrain is signi¯- { cli®s, berms, rocks, fence posts. To perform ter- cantly greater than that necessary for obsta- rain assessment, it is common practice to endow ve- cle avoidance. hicles with forward-pointed range sensors. Terrain is We present a self-supervised machine learn- then analyzed for potential obstacles. The result is ing approach for estimating terrain rough- used to adjust the direction of vehicle motion (Kelly ness from laser range data. Our approach & Stentz, 1998a; Kelly & Stentz, 1998b; Langer et al., compares sets of nearby surface points ac- 1994; Urmson et al., 2004). quired with a laser. This comparison is chal- lenging due to uncertainty. For example, at When driving at high speed { as in the DARPA Grand range, laser readings may be so sparse that Challenge { terrain roughness must also dictate vehicle signi¯cant information about the surface is behavior because rough terrain induces shock propor- missing. Also, a high degree of precision tional to vehicle speed. The e®ect of shock can be is required when projecting laser readings. detrimental (Brooks & Iagnemma, 2005). To be safe, This precision may be unavailable due to la- a vehicle must sense terrain roughness and slow ac- tency or error in pose estimation. We model cordingly. The accuracy needed for assessing terrain these sources of error as a multivariate poly- roughness exceeds that required for obstacle ¯nding nomial. The coe±cients of this polynomial by a substantial margin { rough patches are often just are obtained through a self-supervised learn- a few centimeters in height. This makes design of a ing process. The \labels" of terrain rough- competent terrain assessment function di±cult. ness are automatically generated from actual In this paper, we present a method that enables a shock, measured when driving over the tar- vehicle to acquire a competent roughness estimator get terrain. In this way, the approach pro- for high speed navigation. Our method uses self- vides its own training labels. It \transfers" supervised machine learning. This allows the vehicle the ability to measure terrain roughness from to learn to detect rough terrain while in motion and the vehicle's inertial sensors to its range sen- without human training input. Training data is ob- sors. Thus, the vehicle can slow before hit- tained by a ¯ltered analysis of inertial data acquired ting rough terrain. at the vehicle core. This data is used to train (in a Our experiments use data from the 2005 supervised fashion) a classi¯er that predicts terrain DARPA Grand Challenge. We ¯nd our ap- roughness from laser data. In this way, the learning proach is substantially more e®ective at iden- approach \transfers" the capability to sense rough ter- tifying rough surfaces and assuring vehicle rain from inertial sensors to environment sensors. The safety than previous methods { often by as resulting module detects rough terrain in advance, al- much as 50%. lowing the vehicle to slow. Thus, the vehicle avoids high shock that would otherwise cause damage. line (Rusinkiewicz & Levoy, 2001), it is not well-suited to autonomous driving. At signi¯cant range, laser We evaluate our method using data acquired in the data is quite sparse. Thus, although the terrain may 2005 DARPA Grand Challenge with the vehicle shown be scanned by multiple di®erent lasers, scans rarely in Fig. 1. Our experiments measure ability to predict cover precisely the same terrain. This e®ectively shock and e®ect of such predictions on vehicle safety. breaks the correspondence of ICP. Therefore, we We ¯nd our method is more e®ective at identifying believe that accurate recovery of a full 3-D world rough surfaces than previous techniques derived from model from this noisy, incomplete data is impossible. obstacle avoidance algorithms. The self-supervised ap- Instead we use a machine learning approach to de¯ne proach { whose training data emphasizes the distin- tests that indicate when terrain is likely to be rough. guishing characteristics between very small disconti- Other work uses machine learning in lieu of recovering nuities { is essential to making this possible. a full 3-D model. For example, (Saxena et al., 2005) Furthermore, we ¯nd our method reduces vehicle shock and (Michels et al., 2005) use learning to reason about signi¯cantly with only a small reduction in average ve- depth in a single monocular image. Although full re- hicle speed. The ability to slow before hitting rough covery of a world model is impossible from a single terrain is in stark contrast to the speed controller used image, useful data can be extracted using appropriate in the race (Thrun et al., 2006b). That controller mea- learned tests. Our work has two key di®erences from sured vehicle shock exclusively with inertial sensing. these prior papers: the use of lasers rather than vision Hence, it sometimes slowed the vehicle after hitting and the emphasis upon self-supervised rather than re- rough terrain. inforcement learning. We present the paper in four parts. First, we de¯ne In addition, other work has used machine learning in the functional form of the laser-based terrain estima- a self-supervised manner. The method is an impor- tor. Then we describe the exact method for generating tant component of the DARPA LAGR Program. It training data from inertial measurements. Third, we was also used for visual road detection in the DARPA train the estimator with the learning. Finally, we ex- Grand Challenge (Dahlkamp et al., 2006). The \trick" amine the results experimentally. of using self-supervision to generate training data au- tomatically, without the need for hand-labeling, has been adapted from this work. However, none of these 2 RELATED WORK approaches addresses the problem of roughness esti- mation and intelligent speed control of fast moving There has been extensive work on Terramechanics vehicles. As a result, the source of the data and the un- (Bekker, 1956; Bekker, 1969; Wong, 1989), the guid- derlying mathematical models are quite di®erent from ance of autonomous vehicles through rough terrain. those proposed here. Early work in the ¯eld relies on accurate a priori mod- els of terrain. More recent work (Iagnemma et al., 2004; Brooks & Iagnemma, 2005; Shimoda et al., 2005) 3 THE 3-D POINT CLOUD addresses the problem of online terrain roughness es- timation. However, sensors used in that work are pro- Our vehicle uses a scanning laser mounted on the roof prioceptive and require the robot to drive over terrain as illustrated in Fig. 2. The speci¯c laser generates for classi¯cation. In this paper we predict roughness from laser data so the vehicle can slow in advance of hitting rough terrain. Numerous techniques for laser perception exist in the literature, including its application to o®-road driv- ing (Shoemaker et al., 1999; Urmson et al., 2004). Those that address error in 3-D point cloud acquisi- tion are especially relevant for our work. The iterative closest point (ICP) algorithm (Besl & McKay, 1992) is a well-known approach for dealing with this type of error. ICP relies on the assumption that terrain is scanned multiple times. Any overlap in multiple Figure 1: Stanley won the 2005 DARPA Grand Chal- scans represents an error and that mismatch is used lenge by completing a 132 mile desert route in just for alignment. under 7 hours. Data from this race is used to evaluate Although ICP can be implemented on- the terrain perception method described in this paper. Figure 2: The left graphic shows a single laser scan. The scanning takes place in a single plane tilted down- ward. Reliable perception requires integration of scans over time as shown in the right graphic. Figure 3: Pose estimation errors can generate false surfaces that are problematic for navigation. Shown here is a central laser ray traced over time, sensing a relatively flat surface. range data for 181 angular positions at 75Hz with .5 degree angular resolution. The scanning takes place in a single plane, tilted slightly downward, as indicated 4 SURFACE CLASSIFICATION in the ¯gure. In the absence of obstacles, a scanner produces a line orthogonal to the vehicle's direction of This section describes the function for evaluating 3-D travel. Detecting obstacles, however, requires the third point clouds. The point clouds are scored by compar- dimension. This is achieved through vehicle motion. ing adjacent measurement points. The function for As the vehicle moves, 3-D measurements are integrated this comparison considers a number of features (in- over time into a 3-D point cloud. cluding the elapsed time) which it weights using a Integration of data over time is not without problems. number of parameters. The optimization of these pa- It requires an estimate of the coordinates at which rameters is discussed in Sect. 6. a measurement was taken and the orientation of the After projection as described in Sect. 3, an individ- sensor { in short, the robot pose.