A Camera-Based Position Correction System for Autonomous Production Line Inspection

Total Page:16

File Type:pdf, Size:1020Kb

A Camera-Based Position Correction System for Autonomous Production Line Inspection sensors Article A Camera-Based Position Correction System for Autonomous Production Line Inspection Amit Kumar Bedaka 1,* , Shao-Chun Lee 1, Alaa M. Mahmoud 1, Yong-Sheng Cheng 1 and Chyi-Yeu Lin 1,2,3,* 1 Department of Mechanical Engineering, National Taiwan University of Science and Technology, Taipei 106, Taiwan; [email protected] (S.-C.L.); [email protected] (A.M.M.); [email protected] (Y.-S.C.) 2 Center for Cyber-Physical System Innovation, National Taiwan University of Science and Technology, Taipei 106, Taiwan 3 Taiwan Building Technology Center, National Taiwan University of Science and Technology, Taipei 106, Taiwan * Correspondence: [email protected] (A.K.B.); [email protected] (C.-Y.L.) Abstract: Visual inspection is an important task in manufacturing industries in order to evaluate the completeness and quality of manufactured products. An autonomous robot-guided inspection system was recently developed based on an offline programming (OLP) and RGB-D model system. This system allows a non-expert automatic optical inspection (AOI) engineer to easily perform inspections using scanned data. However, if there is a positioning error due to displacement or rotation of the object, this system cannot be used on a production line. In this study, we developed an automated position correction module to locate an object’s position and correct the robot’s pose and position based on the detected error values in terms of displacement or rotation. The proposed Citation: Bedaka, A.K.; Lee, S.-C.; module comprised an automatic hand–eye calibration and the PnP algorithm. The automatic hand– Mahmoud, A.M.; Cheng, Y.-S.; Lin, eye calibration was performed using a calibration board to reduce manual error. After calibration, the C.-Y. A Camera-Based Position PnP algorithm calculates the object position error using artificial marker images and compensates for Correction System for Autonomous the error to a new object on the production line. The position correction module then automatically Production Line Inspection. Sensors 2021, 21, 4071. https://doi.org/ maps the defined AOI target positions onto a new object, unless the target position changes. We 10.3390/s21124071 performed experiments that showed that the robot-guided inspection system with the position correction module effectively performed the desired task. This smart innovative system provides a Academic Editors: Abir Hussain, novel advancement by automating the AOI process on a production line to increase productivity. Dhiya Al-Jumeily, Hissam Tawfik and Panos Liatsis Keywords: automatic optical inspection (AOI); offline programming; position correction; image processing; production line; manipulator; camera-based system Received: 30 May 2021 Accepted: 12 June 2021 Published: 13 June 2021 1. Introduction Publisher’s Note: MDPI stays neutral The development of industrial automation became quite mature by the 1970s, and with regard to jurisdictional claims in many companies have been gradually introducing automated production technology published maps and institutional affil- to assist production lines since [1]. The automation of the production line has enabled iations. the manufacturing process to be more efficient, data-based, and unified. Automation technologies have heavily influenced many processes, from loading and unloading objects to sorting, assembly, and packaging [2]. In 2012, the German government proposed the concept of Industry 4.0 [3]. Many Copyright: © 2021 by the authors. companies have started to develop automation technologies, from upper layer software Licensee MDPI, Basel, Switzerland. or hardware providers and middle layer companies that store, analyze, manage, and This article is an open access article provide solutions, to lower layer companies, so that they can apply these technologies distributed under the terms and in their factories [4]. In addition, a smarter integrated sensing and control system based conditions of the Creative Commons Attribution (CC BY) license (https:// on existing automation technologies has begun to evolve as part of the development of a creativecommons.org/licenses/by/ highly automated or even fully automated production model [5]. 4.0/). Sensors 2021, 21, 4071. https://doi.org/10.3390/s21124071 https://www.mdpi.com/journal/sensors Sensors 2021, 21, 4071 2 of 19 In modern automation technology, development is mainly based on a variety of algorithms and robot arm control systems [6]. The robotic arm has the advantages of high precision, fast-moving speed, and versatile motion. In advanced applications, ex- tended sensing modules, such as vision modules, distance sensors, or force sensors, can be integrated with the robotic arm to give human sense to robotic systems [7]. Robot systems in general have the ability to “see” things through various vision systems [8]. These help robots to detect and locate objects, which speeds up the technical process. In addition, they improve the processing accuracy and expand the range of applications for automated systems. The type of data obtained can be approximately divided into a 2D vision system (RGB/Mono Camera system), which receives planar images [9], and a 3D vision system (RGB-D Camera system), which receives depth images in the form of vision modules [10]. These two systems have their own advantages and disadvantages and suitable application scenarios. In the field of industrial manufacturing automation, the 2D vision system is commonly used to meet the needs of high precision, ease of installation, and use. The robotic arm with a 2D vision camera is the most commonly used setup in today’s industrial automation applications. Pick and place [11] and random bin picking [12] are among the most frequent applications that use an industrial camera with a manipulator. Several open-source and commercial toolboxes, such as OpenCV [13] and Halcon [14], are already available for use in vision systems. However, regardless of how mature the software support is, a lot of hardware is still needed with human intervention to complete the process. Camera calibration is needed before operation and is applied to various image processing algorithms so that important parameters, such as focal length, center of image, intrinsic parameter, extrinsic parameter, and lens distortion [15,16], can be learnt. Furthermore, camera calibration is a challenging and time-consuming process for an operator unfamiliar with the characteristics of the camera on the production line. Camera calibration is a major issue that is not easily solved in the automation industry and complicates the introduction of production line automation. Traditional vision-based methods [17–20] require 3D fixtures corresponding to a reference coordinate system to calibrate a robot. These methods are time-consuming, inconvenient, and may not be feasible in some applications. A camera installed in the working environment or mounted on a robotic arm can be categorized as an eye-to-hand (camera-in-hand) calibration or stand-alone calibration [21]. The purpose of the hand–eye correction is similar to that of the robot arm end-point tool calibration (TCP calibration), which obtains a convergence homogeneous matrix between the robot end-effector and the tool end-point [22]. However, unlike TCP calibration, it can be corrected by using tools to touch fixed points in different positions. To obtain a hand–eye conversion matrix, the visual systems use different methods, such as the parametrization of a stochastic mode [23] and dual-quaternion parameterization [24], since the actual image center cannot be used. For self-calibration methods, the camera is rigidly linked to the robot end-effector [25]. A vision-based measurement device and a posture measuring device have been used in a system that captures robot position data to model manipulator stiffness [26] and estimate kinematic parameters [27–29]. The optimization technique is based on the end-effector’s measured positions. However, these methods require offline calibration, which is a limitation. In such systems, accurate camera model calibration and robot kinematics model calibration are required for accurate positioning. The camera calibration procedure required to achieve high accuracy is, therefore, time-consuming and expensive [30]. Positioning of the manufactured object is an important factor in industrial arm appli- cations. If the object is not correctly positioned, it may cause assembly failure or destroy the object. Consequently, the accuracy of object positioning often indirectly influences the processing accuracy of the automated system. Although numerous studies have been conducted to define object positioning accurately based on vision systems [31–33], no system has been found with an offline programing platform to perform AOI inspection Sensors 2021, 21, 4071 3 of 19 on a production line. Table1 shows a comparison between the system we propose here and to existing vision-based position correction systems in terms of performance. The proposed system expedites the development of a vision-based object position correction module for a robot-guided inspection system. This allows the robot-guided inspection system to complete AOI tasks automatically in a production line regardless of the object’s position. In the proposed system, the AOI targets can be automatically mapped onto a new object for the inspection task, whereas existing vision systems have been developed to locate the object’s position but not for the production line. To operate these systems, the operator
Recommended publications
  • China's Space Robotic Arm Programs
    SITC Bulletin Analysis October 2013 China’s Space Robotic Arm Programs Kevin POLLPETER Deputy Director, Study of Innovation and Technology in China Project UC Institute on Global Conflict and Cooperation On July 20, 2013, China launched three satellites on a Long March 4C launch vehicle, ostensibly to test space debris observation and space robotic arm technologies. The three satellites, Chuangxin-3, Shiyan-7, and Shijian-15, drew the attention of satellite tracking enthusiasts when two of them began conducting orbital maneuvers with each other and an additional satellite that had been launched in 2005. The maneuvers began on August 1 and involved one satellite acting as the target and another satellite, most likely equipped with a robotic arm, grappling the target satellite. Exactly which two of the three satellites were involved in the maneuvers is unknown. Based on data from the U.S. Strategic Command’s Space-Track.org website, however, the largest satellite of the three, possibly the Shijian-15, fired its thrusters to move to the smallest of the three satel- lites, possibly the Chuangxin-3, which remained in a set orbit.1 The third satellite, possibly the Shiyan-7, does not appear to be involved in the test. These maneuvers continued until August 17 and resulted in the largest satellite closing in on and then away from the smallest satellite. On August 18, the largest satellite changed orbits and closed in on a completely separate satellite, the Shijian-7, that had been launched in 2005. These maneuvers have caused concern that the tests go beyond the stated objectives and are actually a cover for testing on-orbit anti-satellite (ASAT) technologies.
    [Show full text]
  • 2.1: What Is Robotics? a Robot Is a Programmable Mechanical Device
    2.1: What is Robotics? A robot is a programmable mechanical device that can perform tasks and interact with its environment, without the aid of human interaction. Robotics is the science and technology behind the design, manufacturing and application of robots. The word robot was coined by the Czech playwright Karel Capek in 1921. He wrote a play called “Rossum's Universal Robots” that was about a slave class of manufactured human-like servants and their struggle for freedom. The Czech word robota loosely means "compulsive servitude.” The word robotics was first used by the famous science fiction writer, Isaac Asimov, in 1941. 2.1: What is Robotics? Basic Components of a Robot The components of a robot are the body/frame, control system, manipulators, and drivetrain. Body/frame: The body or frame can be of any shape and size. Essentially, the body/frame provides the structure of the robot. Most people are comfortable with human-sized and shaped robots that they have seen in movies, but the majority of actual robots look nothing like humans. Typically, robots are designed more for function than appearance. Control System: The control system of a robot is equivalent to the central nervous system of a human. It coordinates and controls all aspects of the robot. Sensors provide feedback based on the robot’s surroundings, which is then sent to the Central Processing Unit (CPU). The CPU filters this information through the robot’s programming and makes decisions based on logic. The same can be done with a variety of inputs or human commands.
    [Show full text]
  • Design and Construction of a Robotic Arm for Industrial Automation
    Published by : International Journal of Engineering Research & Technology (IJERT) http://www.ijert.org ISSN: 2278-0181 Vol. 6 Issue 05, May - 2017 Design and Construction of a Robotic Arm for Industrial Automation Md. Tasnim Rana*, Anupom Roy Department of Mechanical Engineering, Department of Mechanical Engineering, Khulna University of Engineering & Technology, Khulna- Khulna University of Engineering & Technology, Khulna- 9203, BANGLADESH 9203, BANGLADESH Abstract - The main concentration of the work was to make a assembly. In some circumstances, close emulation of the cost efficient autonomous robotic arm in terms of industrial human hand is desired, as in robots designed to conduct bomb automation. It is a type of mechanical arm, usually disarmament and disposal. In case of firefighting or rescue programmable, with similar functions to a human arm; the arm operation where human life is in danger robotic arm can be may be a unit mechanism or may be a part of a more complex used as a rescue device. This can be functioned as required robotic process. The end effector or robotic hand can be designed to perform any desired task such as welding, gripping, and can do works risky for human being. In case of rapid spinning etc., depending on the application. For detective production the time limit for production will be shorten with investigations and bomb disposal it can be used as an essential the use of robotic arm. machine. In industry any kind of work which should be accurate and works continuously, normal programming algorithms and 2. BACKGROUND mechanical function can do the job perfectly .It can sense the co- At first robot was developed by Leo nartho the vence.
    [Show full text]
  • Design, Manufacturing and Analysis of Robotic Arm with SCARA Configuration
    International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 Design, Manufacturing and Analysis of Robotic Arm with SCARA Configuration Kaushik Phasale1, Praveen Kumar2, Akshay Raut3, Ravi Ranjan Singh4, Amit Nichat5 1,2,3,4Students, 5Assitant Proffesor, Deparatment of Mechanical Engineering, JSPM’s Bhivarabai Sawant Institute of Technology & Research, Wagholi, Pune, Maharasthra, India. ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract - This paper deals with the “Design, The use of robots in library is becoming more popular in Manufacturing and Analysis of Robotic Arm with SCARA recent years. The trend seems to continue as long as the Configuration”. In the modern world, robotics has become robotics technology meets diverse and challenging needs in popular, useful and has achieved great successes in several educational purpose. The prototype consists of robotic arm along with grippers capable of moving in the three axes and fields of humanity. Every industrialist cannot afford to an ATMEGA 8 microcontroller. Software such as AVR Studio transform his unit from manual to semi-automatic or fully is used for programming, PROTESUS is used for simulation automatic as automation is not that cheap in India. The basic and PROGISP is used for dumping the program. RFID is used objective of this project is to develop a versatile and low cost for identifying the books and it has two IR Sensors for robotic arm which can be utilized for Pick and Place detecting the path. This robot is about 4 kg in weight and it is operation. Here controlling of the robot has been done by capable of picking and placing a book of weight one kg.s.
    [Show full text]
  • Computer Vision Based Robotic Arm Controlled Using Interactive GUI
    Intelligent Automation & Soft Computing Tech Science Press DOI:10.32604/iasc.2021.015482 Article Computer Vision Based Robotic Arm Controlled Using Interactive GUI Muhatasim Intisar1, Mohammad Monirujjaman Khan1,*, Mohammad Rezaul Islam1 and Mehedi Masud2 1Department of Electrical and Computer Engineering, North South University, Bashundhara, Dhaka-1229, Bangladesh 2Department of Computer Science, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia ÃCorresponding Author: Mohammad Monirujjaman Khan. Email: [email protected] Received: 24 November 2020; Accepted: 19 December 2020 Abstract: This paper presents the design and implementation of a robotic vision system operated using an interactive Graphical User Interface (GUI) application. As robotics continue to become a more integral part of the industrial complex, there is a need for automated systems that require minimal to no user training to operate. With this motivation in mind, the system is designed so that a beginner user can operate the device with very little instruction. The application allows users to determine their desired object, which will be picked up and placed by a robotic arm into the target location. The application allows users to filter objects based on color, shape, and size. The filtering along the three parameters is done by employing a Hue-Saturation-Value (HSV) mode color detection algorithm, shape detection algorithm, size determining algorithm. Once the target object is identi- fied, a centroid detection algorithm is employed to find the object’s center coor- dinates. An inverse kinematic algorithm is used to ascertain the robotic arm’sjoint positions for picking the object. The arm then goes through a set of preset posi- tions to pick up the object, place the object, and then return the arm to the initial position.
    [Show full text]
  • A Review of Motion Planning Algorithms for Robotic Arm Systems
    This is a repository copy of A Review of Motion Planning Algorithms for Robotic Arm Systems. White Rose Research Online URL for this paper: https://eprints.whiterose.ac.uk/168146/ Version: Accepted Version Proceedings Paper: Liu, Shuai and Liu, Pengcheng orcid.org/0000-0003-0677-4421 (Accepted: 2020) A Review of Motion Planning Algorithms for Robotic Arm Systems. In: 8th International Conference on Robot Intelligence Technology and Applications (proceedings). The 8th International Conference on Robot Intelligence Technology and Applications, 11-13 Dec 2020, Cardiff. IEEE , GBR . (In Press) Reuse Items deposited in White Rose Research Online are protected by copyright, with all rights reserved unless indicated otherwise. They may be downloaded and/or printed for private study, or other acts as permitted by national copyright laws. The publisher or other rights holders may allow further reproduction and re-use of the full text version. This is indicated by the licence information on the White Rose Research Online record for the item. Takedown If you consider content in White Rose Research Online to be in breach of UK law, please notify us by emailing [email protected] including the URL of the record and the reason for the withdrawal request. [email protected] https://eprints.whiterose.ac.uk/ A Review of Motion Planning Algorithms for Robotic Arm Systems Shuai Liu[0000-0001-6939-494X] and Pengcheng Liu[0000-0003-0677-4421] Department of Computer Science, University of York, York YO10 5GH, United Kingdom [email protected] Abstract. Motion planning plays a vital role in the field of robotics.
    [Show full text]
  • An Engineer's Guide to Industrial Robot Designs
    e-book An Engineer’s Guide to Industrial Robot Designs A compendium of technical documentation on robotic system designs ti.com/robotics Q2 | 2020 Table of Contents/Overview 1. Introduction 3. Robot arm and driving system (manipulator) 1.1 An introduction to an industrial robot system. 3 3.1.1 How to protect battery power management systems from thermal damage. 45 2. Robot system controller 3.1.2 Protecting your battery isn’t as hard as 2.1 Control panel you think.....................................46 2.1.1 Using Sitara™ processors for Industry 3.1.3 Position feedback-related reference designs 4.0 servo drives.......................9 for robotic systems............................47 2.2 Servo drives for robotic systems 4. Sensing and vision technologies 2.2.1 The impact of an isolated gate driver. 13 4.1 TI mmWave radar sensors in robotics 2.2.2 Understanding peak source and applications..................................48 sink current parameters . 17 4.2 Intelligence at the edge powers autonomous 2.2.3 Low-side gate drivers with UVLO versus factories .....................................53 BJT totem poles ......................19 4.3 Use ultrasonic sensing for graceful robots. 55 2.2.4 An external gate-resistor design guide for 4.4 How sensor data is powering AI in robotics. 57 gate drivers.......................... 20 4.5 Bringing machine learning to embedded 2.2.5 High-side motor current monitoring for systems .....................................61 overcurrent protection. 22 4.6 Robots get wheels to address new 2.2.6 Five benefits of enhanced PWM rejection for challenges and functions. 65 in-line motor control. 24 4.7 Vision and sensing-technology reference 2.2.7 How to protect control systems from thermal designs for robotic systems.
    [Show full text]
  • Comprehensive Review on Modular Self-Reconfigurable Robot Architecture
    International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 Comprehensive Review on Modular Self-Reconfigurable Robot Architecture Muhammad Haziq Hasbulah1, Fairul Azni Jafar2, Mohd. Hisham Nordin2 1Centre for Graduate Studies, Universiti Teknikal Malaysia Melaka, Hang Tuah Jaya, 76100 Durian Tunggal, Melaka, Malaysia 2Faculty of Manufacturing Engineering, Universiti Teknikal Malaysia Melaka, Hang Tuah Jaya, 76100 Durian Tunggal, Melaka, Malaysia ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract - Self-reconfigurable modular robot is a new film directed by William Don Hall. In this movie, a character approach of robotic system which involves a group of identical named Hiro create a lot of Microbots that able to be robotic modules that are connecting together and forming controlled by neurotransmitter. They are designed by Hiro to structure that able to perform specific tasks. Such robotic connect together to form various shapes and perform tasks system will allows for reconfiguration of the robot and its cooperatively Hall and Williams [3]. The idea of that movie structure in order to adapting continuously to the current concept is multiple robots that able to change shape in group needs or specific tasks, without the use of additional tools. being controlled by human thought. Nowadays, the use of this type of robot is very limited because The MSR robot is build based on the electronics components, it is at the early stage of technology development. This type of computer processors, and memory and power supplies, and robots will probably be widely used in industry, search and also they might have a feature for the robot to have an ability rescue purpose or even on leisure activities in the future.
    [Show full text]
  • Canada's Robotic Arms
    CANADA’S ROBOTIC ARMS Murtaza Syed Aqeel Introduction I picked this because it had a big impact on Canada and Canada’s economics. The Canadarm inspired many more robots that are not just in space but have helped us here on earth. Thank you for all the people who created Canadarm! Canadarm Canadarm’s specifications The Canadarm also known as Shuttle Remote Manipulator System (for short SRMS) was remote controlled by astronauts on the shuttle. It weighs 410 kg and its length is 15 m. It has 2 cameras 1 on the elbow and 1 on the wrist Who made it? The fabulous Canadarm cost 110 million dollars. The companies that made Canadarm were Spar Aerospace, it also had CAE Electronics Ltd. And DSMA Acton Ltd. The Canadarm was made on November 13, 1981. What can it do? The Canadarm can lift very heavy things, that is one reason space workers need the Canadarm. It can lift to 30,000 kg on earth and in space it can lift up to 266,000 kg in space. A couple of things Canadarm did were repairing satellites, deploying satellites, and capturing satellites. It did these and other things because it was too dangerous for astronauts to do. How is it repaired? It can be repaired in earth unlike some of the other robots Retirement The Canadarm has done 30 years of great work and is now retired and is being displayed at the Canada Aviation and space Museum in Ottawa, Ontario. it retired on July 2011. Controlling the Canadarm Chris Hadfield was the first Canadian astronaut to control the Canadarm.
    [Show full text]
  • Characterization of Semi-Autonomous On-Orbit Assembly Cubesat Constellation
    SSC19-WP1-09 Characterization of Semi-autonomous On-orbit Assembly CubeSat Constellation John M. Gregory, Jin S. Kang, Michael Sanders, Dakota Wenberg United States Naval Academy 590 Holloway Rd., MS 11B, Annapolis, MD 21402; 410-293-6416 [email protected] Ronald M. Sega Colorado State University 2545 Research Blvd., Fort Collins, CO 80526; 970-491-7067 [email protected] ABSTRACT Demand for more complex space systems is ever increasing as the scale of the future missions expands. Accordingly, much focus has been given recently to innovations in on-orbit assembly and servicing to ensure those missions are executed in a time-efficient manner. The past on-orbit servicing demonstrations have involved large satellites that were designed to dock/berth and service specific client satellites, and did not leverage the current advancements in small satellite technology. The U.S. Naval Academy (USNA) is contributing to advancing the on- orbit servicing and assembly technology with a next-generation robotic arm Intelligent Space Assembly Robot (ISAR) system, which is envisioned to operate independently or as a constellation of 3U CubeSats and seeks to demonstrate semi-autonomous robotic assembly capabilities on-orbit on a nano-satellite scale. This paper will present an overview of the ISAR system, outline design, operation, and demonstration modifications for the on-orbit demonstrator, analyze the results from the ground test platform, and discuss the interfacing between existing robotic operations structures and advanced sensors. It will also focus on the analysis of cost effectiveness of the proposed mission architecture by characterizing the operation envelope of CubeSat-based assembly satellite constellations and volumetric efficiency analysis of on-orbit assembly using “Bin of Parts”.
    [Show full text]
  • Advanced Use Robotic Arm AURA Soft As a Human Touch
    EN AURA Advanced Use Robotic Arm AURA Soft as a Human Touch 2 Comau Aura - Advanced Use Robotic Arm The Culture of Automation Designing advanced automation solutions means thinking about the industry in a new way, developing new scenarios, designing innovative products and creating ways to streamline production processes. It requires more than technical competence; it requires a team of professionals whose vision is rooted in a culture of excellence. It also requires a combination of talent, passion and experience that unite to define new trends in automation. Here at Comau our passion for our work reflects who we are. 3 Comau Aura - Advanced Use Robotic Arm Comau HUMANufacturing Industry 4.0. The factory is changing to become a network of flexible, modular and scalable automation systems that are capable of operating autonomously or in a secure synergistic way with operators, always connected and under control. With AURA, automation is no longer confined within barriers, but it collaborates with human beings: this is what we like to call Comau HUMANufacturing. 4 Comau Aura - Advanced Use Robotic Arm Discover AURA: almost human The skin which covers this Comau robot recalls human skin sensitivity. AURA supports humans as they perform manual operations by optimizing the work process. Speed is tuned, depending on the device signals, which makes it possible for the robot to move in open space or in contact operations by following programmed trajectories or learning from the operator via manual guidance. Why AURA is unique: • compared to collaborative
    [Show full text]
  • Robolink: Modular Assistive Robot Arm
    Project Number: RBE.TP1.CPS2 Robolink: Modular Assistive Robot Arm A Major Qualifying Project Report Submitted to the faculty of the WORCESTER POLYTECHNIC INSTITUTE In partial fulfillment of the requirements for the Degree of Bachelor of Science Advisor: Taskin Padir Submitted By: Cassiopia Hudson Gabriel Morell-Pacheco 1 Acknowledgments The team would like to extend our thanks to several people who helped make this project possible: Taskin Padir, for his help and guidance as our advisor, Ty Tremblay, for configuring the Robolink and getting us started, Lillian Walker, for contributing her time and CAD experience, Joe St. Germain, for his constant and helpful advice, James Fleming, for his expertise on the Cyber Physical Systems project, and Thomas Angelotti and Patrick Morrison, for opening the ECE shop to us. 2 2 Abstract The goal of this project is to utilize the igus® Robolink arm five degree of freedom modular robot arm, to complete useful tasks for persons with no or limited mobility. These tasks include driving the joystick of a wheelchair, flipping a light switch, and turning the pages of a book. This is done through designing and building a modular interface for mounting the Robolink arm onto an existing wheelchair project and implementing a universal control interface in the software for future expansion of tasks and control methods. 3 3 Table of Contents 1 Acknowledgments ................................................................................................................... 2 2 Abstract ...................................................................................................................................
    [Show full text]