A Reinforcement Learning Control Approach for Underwater Manipulation Under Position and Torque Constraints

A Reinforcement Learning Control Approach for Underwater Manipulation Under Position and Torque Constraints

A reinforcement learning control approach for underwater manipulation under position and torque constraints Ignacio Carluchoa, Mariano De Paulab, Corina Barbalataa, Gerardo G. Acostab [email protected], mariano.depaula@fio.unicen.edu.ar, [email protected], ggacosta@fio.unicen.edu.ar a Department of Mechanical Engineering, Louisiana State University, Baton Rouge, USA b INTELYMEC Group, Centro de Investigaciones en F´ısica e Ingenier´ıa del Centro CIFICEN – UNICEN – CICpBA – CONICET, 7400 Olavarr´ıa, Argentina Abstract—In marine operations underwater manipulators play computational requirements. However, when it is used for con- a primordial role. However, due to uncertainties in the dy- trolling manipulators arms, it must cope with highly non-linear namic model and disturbances caused by the environment, low- systems. This issue is aggravated in underwater environments level control methods require great capabilities to adapt to change. Furthermore, under position and torque constraints where unknown disturbances affect the behaviour of the arm. the requirements for the control system are greatly increased. Furthermore, for underwater manipulators, controllers are used Reinforcement learning is a data driven control technique that under the assumption that the arm will move slowly and as can learn complex control policies without the need of a model. such it is possible to decouple each degree of freedom, some- The learning capabilities of these type of agents allow for great thing that is not true for every application [5]. Researchers adaptability to changes in the operative conditions. In this article we present a novel reinforcement learning low-level controller for have also turned to non-linear optimal control techniques as a the position control of an underwater manipulator under torque viable option, since they allow to optimize a cost function and position constraints. The reinforcement learning agent is under different metrics. One of these techniques is Model based on an actor-critic architecture using sensor readings as Predictive Control (MPC) [6], used successfully for controlling state information. Simulation results using the Reach Alpha 5 a different number of underwater robots [7], [8]. However, underwater manipulator show the advantages of the proposed control strategy. one of the drawbacks of this technique is that it requires an Index Terms—Underwater manipulation, Reinforcement learn- accurate model of the plant in order to work properly, not a ing, Neural networks, Intelligent control, Deep Deterministic trivial matter in underwater robotics [9]. Policy Gradient Data driven control techniques have appeared as an alter- native for systems with complex or unknown models. One I. INTRODUCTION of these techniques is Reinforcement Learning (RL) [10]. In the RL framework, the robot arm control problem can be In the last decades the importance of underwater manip- formulated as a Markov Decision Process (MDP) [11]. Solving ulators in marine operations has grown continuously. Most a RL problem consists in iteratively learning a task from robotic underwater industrial applications are conducted with interactions to achieve a goal. During learning, an artificial Remotely Operated Vehicles (ROV) where a human operator agent (controller) interacts with the target system (arm) by is tasked with the remote operation of the manipulator [1]. taking an action (torque command), that makes the robot n arXiv:2011.12360v1 [cs.RO] 24 Nov 2020 However, due to the limited number of expert operators and evolve from its current state xt 2 X ⊆ R to xt+1. The the high cost of operations, the industry is migrating towards agent then receives a numerical signal rt, called reward, which Autonomous Underwater Vehicles (AUV) [2]. In this type provides a measure of how good (or bad) the action taken of scenario, a manipulator, usually electric, is mounted on at time t is in terms of the observed state transition. Many the AUV and operates autonomously, however, this requires works have used the RL paradigm for controlling AUVs in a robust and adaptable control system. Furthermore, in au- underwater environment [12]. However, this technique has not tonomous missions different types of operational constraints yet been applied to underwater manipulators. may appear, such as specific joint constraints that must be fol- The main contribution of this work is the development of a lowed in order to avoid collisions [3], or decreased joint torque reinforcement learning based control system for the low-level due to faulty motors. These constraints increase the need for control of an electric underwater manipulator under position designing complex control systems for robust manipulation. and torque constraints. Our reinforcement learning formulation One of the most used low-level controllers for manipulators is based on the Deep Deterministic Policy Gradient (DDPG) is the classical Proportional Integrative Derivative (PID) con- algorithm [13]. The proposed method uses an actor critic troller [4]. This is due mostly to its simplicity of use and low structure, where the actor is a function that maps system 1 states to actions and the critic is a function that assess actions chosen by the actor. Deep Neural Networks (DNN) are used as function approximators for the actor and critic. Results in simulation show the advantages of our proposal when controlling a simulated version of the Reach 5 Alpha manipulator, shown in Fig. 1. The proposed controller is compared with a MPC, showing that the RL controller is able to outperform the MPC. The article is structured as follows, Section II presents an overview of related works followed by Section III that introduces the basics of RL control utilized in our formulation. In Section IV the details of our implementations are described, in Section V we present the results obtained with our proposed control scheme, and finally Section VI presents the overall conclusions of the proposed work. II. RELATED WORKS Designing control systems under constrains considerations for robotic manipulators appeared due to the need of robots to interact with the environment. Some of the most fundamental Fig. 1. Reach 5 Alpha underwater manipulator [26] approaches focused on designing motion/interaction control systems by using a hybrid control formulation [14], [15]. combination with a Deep Neural Network. Specifically for In this approach, the constrains are expressed based on the the underwater environments, RL has been used as a control end-effector’s working space, and are used to decide the techniques in several previous works [23]–[25]. However, the type of control law (either a motion control law or a force majority of these works focus on the of AUVs. Works utilizing regulator). Nevertheless, constraints are not imposed only by RL in underwater manipulators are lacking in the literature. the interaction with the environment, but are also required for cases when the robotic manipulator has to adjust its working III. REINFORCEMENT LEARNING BASED CONTROL space due to obstacles in the environment, or faults in the From the point of view of classical and modern control, robotic system. In [16] a passivity-based kinematic control the design strategy of a control system is based on the law is proposed under joint velocity limits considerations. The complete knowledge of the dynamics of the system under method proposed can be adapted to different feedback control study. Assuming that there is a model with the adequate laws, but can be applied only for redundant systems. An capacity to describe the dynamics of the system, generally adaptive neural-network control for robotic manipulators with through a set of differential equations, the control problem is parametric uncertainties and motion constraints is proposed reduced to design a controller (or agent) capable of generating in [17]. The simulation and experimental results with a 2 the adequate control actions for the system to achieve a degree of freedom (DOF) planar manipulator show the velocity given objective, goal, task or specific desired behavior. In this constraints always being respected, but steady-state errors way, the performance capabilities of the conventional designed are present. Deep neural-network approaches have become control systems are excessively dependent on the mathematical popular in the past years. An example is given in [18] where models used to describe the behavior of the dynamic systems an obstacle avoidance control law is designed for redundant to be controlled. However, underwater manipulation is a manipulators. The problem is reformulated as a Quadratic complex decision making problem in which, the presence of Programming (QP) problem in the speed level, and a deep uncertainty in dynamics is ubiquitous and, consequently, it is recurrent neural network is designed to solve the QP problem of paramount importance designing and using controllers with in an online way. Although the simulation results show that suitable adaptation capabilities. the robot is capable of avoiding the obstacles while tracking Markov decision processes are models for sequential deci- the predefined trajectories, an experimental evaluation is not sion making problems when outcomes are uncertain [11]. In presented. our formulation, we consider a finite-horizon Markov decision The requirements of adaptability and the difficulties with process with a 1; 2; :::; T decisions and T − 1 visited stages modeling have lead research towards intelligent control meth- [27]. That is, the decision

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    7 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us