Teleoperation of Autonomous Vehicle with 360° Camera Feedback
Total Page:16
File Type:pdf, Size:1020Kb
Teleoperation of Autonomous Vehicle With 360° Camera Feedback Master’s thesis in Systems, Control and Mechatronics OSCAR BODELL ERIK GULLIKSSON Department of Signals and Systems CHALMERS UNIVERSITY OF TECHNOLOGY Gothenburg, Sweden 2016 Master’s thesis in systems, control and mechatronics Teleoperation of Autonomous Vehicle With 360° Camera Feedback OSCAR BODELL ERIK GULLIKSSON Department of Signals and Systems Division of Control, Automation and Mechatronics Chalmers University of Technology Gothenburg, Sweden 2016 Teleoperation of Autonomous Vehicle With 360° Camera Feedback OSCAR BODELL ERIK GULLIKSSON © OSCAR BODELL, ERIK GULLIKSSON 2016. Master’s Thesis 2016:43 Department of Signals and Systems Division of Control, Automation and Mechatronics Chalmers University of Technology SE-412 96 Gothenburg Sweden Telephone +46 (0)31 772 1000 Cover: Teleoperation setup with stitched camera view & graphical user interface. Chalmers Reproservice Gothenburg, Sweden 2016 Teleoperation of Autonomous Vehicle With 360° Camera Feedback Master’s thesis in Systems, Control and Mechatronics OSCAR BODELL ERIK GULLIKSSON Department of Signals and Systems Division of Control, Automation and Mechatronics Chalmers University of Technology Abstract Teleoperation is using remote control from outside line of sight. The operator is often assisted by cameras to emulate sitting in the vehicle. In this report a system for tele- operation of an autonomous Volvo FMX truck is specified, designed, implemented and evaluated. First a survey of existing solutions on the market is conducted find- ing that there are a few autonomous and teleoperation solutions available, but they are still new and information is sparse. To identify what types of requirements are needed for such a system and how to design it a literature study is performed. The system is then designed from the set requirements in a modular fashion using the Robot Operating System as the underlying framework. Four cameras are mounted on the cab and in software the images are stitched together into one 360◦ image that the operator can pan around in. The system is designed so that the operator at any time can pause the autonomous navigation and manually control the vehicle via teleoperation. When the operators intervention is completed the truck can resume autonomous navigation. A solution for synchronization between manual and autonomous mode is specified. The truck is implemented in a simulation where the functionality and requirements of the system is evaluated by a group of test subjects driving a test track. Results from simulation show that latencies higher than 300 ms lead to difficulties when driving, but having a high frame rate is not as critical. The benefit of a full 360◦ camera view compared to a number of fixed cameras in strategic places is not obvious. The use of a head mounted display together with the 360◦ video would be of interest to investigate further. Keywords: Teleoperation, autonomous vehicle, surround view, remote steering. i Acknowledgements First of all we would like to thank CPAC Systems for letting us become part of the family and our supervisors at CPAC Annegret Render and Michael Pettersson. Thank you also Knut Åkesson for being our supervisor and examiner at Chalmers University of Technology and helping us with the report. A final big thank you to Josef Kindberg for your help and expertise in ROS. Oscar Bodell, Erik Gulliksson, Gothenburg, June 2016 iii Contents Abstracti Acknowledgements iii Contentsv Nomenclature ix List of Figures xi List of Tables xiii 1 Introduction1 1.1 Purpose & Objective........................... 1 1.2 Main Research Questions......................... 2 1.3 Boundaries ................................ 2 1.4 Method .................................. 2 1.5 Thesis Contribution............................ 3 1.6 Thesis Outline............................... 4 2 Application Overview5 2.1 Level of Autonomy............................ 5 2.2 Evaluation vehicle............................. 6 2.3 Cameras and stitching .......................... 6 2.4 Visualization and Operator Support................... 6 2.5 User Controls............................... 7 2.6 Autonomous Functions.......................... 7 2.7 Implementation.............................. 8 3 Existing Solutions and Standards9 3.1 Existing standards for remote control.................. 9 3.2 Original Equipment Manufacturer Solutions .............. 9 3.3 Retrofit Solutions............................. 11 3.4 Outcome of Market Survey........................ 12 4 System Requirements 13 4.1 Identification of Requirements...................... 13 4.2 Video Feedback.............................. 14 4.3 Latency Effects.............................. 15 v Contents 4.4 Sensor Data Presentation......................... 16 4.4.1 Map ................................ 16 4.5 Driver Control Inputs........................... 16 4.6 Autonomous Synchronization ...................... 17 4.7 Communication Interface......................... 17 4.8 Communication Link........................... 17 4.8.1 Free-space Path Loss....................... 18 4.8.2 Alternatives to Wireless ..................... 18 4.8.3 5th Generation Wireless Systems ................ 19 4.8.4 Link Requirement......................... 19 4.9 Safety Requirements ........................... 20 5 System Design 21 5.1 System Architecture ........................... 21 5.1.1 Vehicle............................... 21 5.1.2 User Interface........................... 22 5.2 Framework - Robot Operating System ................. 23 5.3 Subsystems ................................ 23 5.3.1 Cameras.............................. 24 5.3.2 Image Processing......................... 24 5.3.2.1 Image Stitching..................... 25 5.3.2.2 Information Overlay .................. 27 5.3.3 Maps ............................... 28 5.3.4 Control Inputs .......................... 29 5.3.4.1 Haptic Feedback .................... 30 5.3.5 Additional Sensors ........................ 30 5.3.5.1 Light Detection And Ranging............. 30 5.3.5.2 Global Navigation Satellite System.......... 30 5.3.5.3 Inertial Measurement Unit .............. 31 5.4 System Coordinator............................ 31 5.4.1 Graphical User Interface..................... 32 5.4.2 Autonomous Functions...................... 32 5.4.2.1 Navigation ....................... 33 5.4.2.2 Synchronization..................... 33 5.4.2.3 Paths .......................... 33 5.5 Gazebo Simulation ............................ 34 5.5.1 Visual ............................... 34 5.5.2 Mass and Inertia ......................... 35 5.5.3 Collision.............................. 36 5.5.4 Propulsion............................. 36 5.5.5 Steering.............................. 36 5.5.6 Sensors............................... 37 5.5.7 World ............................... 38 5.5.8 Performance............................ 38 6 Results 39 6.1 Standards and System Layout...................... 39 vi Contents 6.2 Autonomous synchronization ...................... 40 6.3 Evaluation................................. 41 6.3.1 Driving Experience and Support Functions........... 41 6.3.2 Impact of Latency ........................ 43 6.4 Results Summary............................. 45 7 Conclusion & Future Work 47 Bibliography 49 vii Contents viii Nomenclature ∗B Nr. of Bytes ∗b Nr. of bits 5G 5th Generation Wireless Systems AHS Autonomous Haulage System CAN Controller Area Network CPU Central Processing Unit FOV Field Of View FPS Frames Per Second FSPL Free-Space Path Loss GNSS Global Navigation Satellite System GPS Global Positioning System (see GNSS) GPU Graphics Processing Unit GUI Graphical User Interface HMD Head Mounted Display HSV Hue Saturation Value IEEE Institute of Electrical and Electronics Engineers IMU Inertial Measurement Unit IP Internet Protocol LAN Local Area Network LiDAR Light Detection And Ranging LOS Line Of Sight OEM Original Equipment Manufacturer OpenCV Open source Computer Vision library PoE Power over Ethernet RGB Red Green Blue ROI Region Of Interest ROS Robot Operating System RTK Real Time Kinematic RTSP Real Time Streaming Protocol RTT Round Trip Time SLAM Simultaneous Localization And Mapping Volvo FMX Volvo Forward control Medium Xtreme VR Virtual Reality Wi-Fi IEEE 802.11xx wireless communication WLAN Wireless Local Area Network ix x List of Figures 2.1 Screenshot of the evaluation vehicle in the Gazebo simulation..... 6 2.2 Screenshot from the stitched video feed overlaid with map, speedome- ter and align assistance for autonomous paths.............. 7 2.3 Screenshot from the operator interface controlling autonomous func- tions and maps............................... 8 5.1 System overview of control center and vehicle with communication nodes.................................... 22 5.2 Four cameras with a 120◦ FOV. Mounted to capture a complete 360◦ view. ................................... 24 5.3 IP camera used for surround view of the vehicle............ 25 5.4 Illustration of the effects of the homography matrix........... 26 5.5 110◦ Stitched camera view from vehicle in simulation with map and offset to chosen path............................ 28 5.6 Overview maps with surroundings and recorded paths. 29 5.7 LiDAR sensor data presentation. .................... 29 5.8 Screenshot of the GUI .......................... 32 5.9 The model of the Volvo FMX truck in Gazebo simulation. 35 5.10 The inertial model of the truck...................... 35 5.11 The collision model of the truck. .................... 36 5.12