Open-Source Drone Programming Course for Distance Engineering Education
Total Page:16
File Type:pdf, Size:1020Kb

Load more
Recommended publications
-
Signature Redacted
Long-range Outdoor Monocular Localization with Active Features for Ship Air Wake Measurement by Brandon J. Draper B.S., Aerospace Engineering, Univ. of Maryland, College Park (2016) Submitted to the Department of Aeronautics and Astronautics in partial fulfillment of the requirements for the degree of M•s-tttr BBJchdM of Science in Aerospace Engineering at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY February 2019 © Massachusetts Institute of Technology 2019. All rights reserved. Author..-Signature redacted . ................ Departme~t of Aeronautics and Astronautics October 1, 2019 Signature redacted Certified by .......... Jonathan P. How R. C. Maclaurin Professor of Aeronautics and Astronautics, MIT Thesis Supervisor --- . / Signature redacted Accepted by .... \..._..,/ Sertac Karaman Associate Professor of Aeronautics and Astronautics, MIT MASSACHUSETTSINSTITUTE Chair, Graduate Program Committee OF TECHNOLOGY MAR 12 2019 LIBRARIES ARCHIVES 2 Long-range Outdoor Monocular Localization with Active Features for Ship Air Wake Measurement by Brandon J. Draper Submitted to the Department of Aeronautics and Astronautics on October 1, 2019, in partial fulfillment of the requirements for the degree of Bachelor of Science in Aerospace Engineering Abstract Monocular pose estimation is a well-studied aspect of computer vision with a wide ar- ray of applications, including camera calibration, autonomous navigation, object pose tracking, augmented reality, and numerous other areas. However, some unexplored areas of camera pose estimation remain academically interesting. This thesis provides a detailed description of the system hardware and software that permits operation in one application area in particular: long-range, precise monocular pose estimation in feature-starved environments. The novel approach to pose extraction uses special hardware, including active LED features and a bandpass-interference optical filter, to significantly simplify the image processing step of the Perspective-n-Point (PnP) problem. -
Team Members: Pooja Bansal (Pbansal2) Suraj Shanbhag (Smshanbh)
ARDUROVER with BBBmini Project made for ECE 785 Advanced Computer Design Team Members: Pooja Bansal (pbansal2) Suraj Shanbhag (smshanbh) OVERVIEW: The aim of the project was to interface the beaglebone with the APM rover in order to impart autonomy to the vehicle. The idea was to make the beaglebone as the gateway between the remote control and the vehicle and also interface with the sensors and communicate with the Ground Control Station(GCS) to give the commands to the rover. The GCS software is called Mission Planner and it runs on Windows OS. This report contains the details about the hardware and software required and procedure to be followed to build and launch the system. HARDWARE: The basic hardware consists of a rover built to be compatible with the APM arduover software (we used the APM rover provided to us), Beagle Bone Black which forms the heart of the system, the BBB mini which is a cape on the BBB to interface with the other sensors and the Remote Control for providing the motion commands to the rover. The hardware of the BBBMini requires the following sensors: 1. MPU-9250 IMU 2. MS5611 barometer 3. 3Dr GPS 4. Wifi adapter 5. RC remote (Taranis capable of combined PWM output on one pin) When all these components are connected to the beaglebone, an external power supply is required. In this project a battery pack with usb output of 5V and 2A was used. The connections are as shown below. The output to servo and ESC has to be on RC2(servo) and RC4(throttle). -
Dragon Bee University of Central Florida EEL 4915C Senior Design II Spring 2016 Group 34
Dragon Bee University of Central Florida EEL 4915C Senior Design II Spring 2016 Group 34 Group Members Ayoub Soud(CpE) Younes Enouiti (EE) Akash Patel(CpE) Nishit Dave(EE) Table of Content 1 Executive Summary…………………………………………………………..……1 2 Project Description ………………………………………………………..………2 2.1 Project Motivation……………………………………………….…………2 2.2 Goals and Objectives…………………………………………..………….3 2.3 Specifications………………………………………………………….……4 2.4 Design Constraints and Standards……………….…………………….. 5 2.4.1 Battery/Power Consumption……………………………...…….11 2.4.2 Video Streaming………………………………………………....12 2.4.3 Budget Allocation………………………………..……………….13 3 Research……………………………………………………………………….……..14 3.1 Quad Rotor Frame…………………………………………………………14 3.1.1 Quad Rotor Frame……………………………………………….14 3.1.2 Motors………………………………………………………..……16 3.1.3 Propellers……………………………………………….…….......20 3.1.4 Electric Speed Controllers (ESC)……………………………... 21 3.1.5 Flight Controller/ArduPilot……………………………………….22 3.2 Microcontroller………………………………………………..…………….24 3.2.1 PCB………………………………………………………………..26 3.3 Communication Technologies…………………..……………………….. 28 3.3.1 Wi-Fi Module for MSP430……………………………………….30 3.3.2 Ardupilot/MSP430 Communication………………..……………31 3.4 Position Detection Sensors………………………..………………………35 3.4.1 Ultrasonic Sensors………………………..………………………35 3.5 Wi-Fi Camera………………………………………….……………………..36 3.6 GPS…………………………………………………….……………………. 38 3.7 Power Consumption Management…………………………………………41 3.7.1 Batteries…………………………………………………………….43 3.7.2 Power Module ……………………………..………………………46 3.7.3 Power Distribution Board …………………………….…………..47 -
Player-Stage Based Simulator for Simultaneous Multi-Robot Exploration and Terrain Coverage Problem
International Journal of Artificial Intelligence & Applications (IJAIA), Vol.2, No.4, October 2011 PLAYER -STAGE BASED SIMULATOR FOR SIMULTANEOUS MULTI -ROBOT EXPLORATION AND TERRAIN COVERAGE PROBLEM K.S. Senthilkumar and K. K. Bharadwaj School of Computer and Systems Sciences Jawaharlal Nehru University, New Delhi 110067 [email protected], [email protected] ABSTRACT One of the possible ways of offering assistance without risking additional human lives during hazardous situations is by deploying a robot team, equipped with various sensors and actuators. Working with intelligent robotics requires a large investment in both money and time. There is a general purpose, open source simulator called Player/Stage, which provides a hardware abstraction layer to several popular robot platforms, and is commonly used by robotics community in research and university teaching, today. This simulator tends to be very simple and task-specific. Player, which is a distributed device server for robots, sensors and actuators, can control either a real or simulated robot thus allowing direct application of developed algorithms to real-life scenarios. Hence, we believe that Player/Stage, when coupled with robust hardware, is a viable paradigm for our Simultaneous MSTC(S-MSTC) algorithm. This paper gives details of our experience in running Player/Stage during the implementation of our online S-MSTC algorithm for multi-robot being implemented in C++ and we use Player/Stage middleware for validation and testing. In addition, the experience with Player/Stage can help us to do research in more complicated situations. KEYWORDS Multi-robot, Exploration, Terrain Coverage, Player/Stage Simulator 1. INTRODUCTION In recent past, there has been an increasing interest in the software side of robotics such as simulators. -
Downloaded Model to the Curved Shape of the IRIS+ Belly
Abstract This project involves the design of a vision-based navigation and guidance system for a quadrotor unmanned aerial vehicle (UAV), to enable the UAV to follow a planned route specified by navigational markers, such as brightly colored squares, on the ground. A commercially available UAV is modified by attaching a camera and an embedded computer called Raspberry Pi. An image processing algorithm is designed using the open-source software library OpenCV to capture streaming video data from the camera and recognize the navigational markers. A guidance algorithm, also executed by the Raspberry Pi, is designed to command with the UAV autopilot to move from the currently recognized marker to the next marker. Laboratory bench tests and flight tests are performed to validate the designs. Fair Use Disclaimer: This document may contain copyrighted material, such as photographs and diagrams, the use of which may not always have been specifically authorized by the copyright owner. The use of copyrighted material in this document is in accordance with the “fair use doctrine”, as incorporated in Title 17 USC S107 of the United States Copyright Act of 1976. I Acknowledgments We would like to thank the following individuals for their help and support throughout the entirety of this project. Project Advisor: Professor Raghvendra Cowlagi II Table of Authorship Section Author Project Work Background Real World Applications AH N/A Research AH N/A Navigation, Guidance, and Control JB N/A GPS AH N/A Python and OpenCV JB, KB N/A System Design Image Processing -
A Step-By-Step Guidance to Build a Drone from Scratch Using Ardupilot APM Navio2 Flight Controller
A step-by-step guidance to build a drone from scratch using Ardupilot APM Navio2 Flight controller Prepared by Sim Kai Sheng Yew Chang Chern 1 Table of Contents Full Component List ............................................................................................................................... 1 Navio2 Emlid Flight Controller ................................................................................................ 2 Raspberry Pi3 Model B ............................................................................................................ 2 Airframe .................................................................................................................................. 3 Motors 2216/950KV ................................................................................................................ 4 Electronic Speed Controller (ESC) ........................................................................................... 5 Propellers ................................................................................................................................ 6 GNSS receiver with antenna ................................................................................................... 7 Transmitter ............................................................................................................................. 8 Receiver ................................................................................................................................... 8 MicroSD card .......................................................................................................................... -
Modelling, Simulation and Control of a Humanoid Robot
Heriot-Watt University School of Engineering and Physical Sciences M odelling, Simulation and control of a humanoid robot Inserting a pin to a hole using DARWINOP Group members: Roshenac Mitchell, Markus Lampert, Nurgaliyev Shakh- Izat, Samir Mammadov, Hugo Sardinha, Waqar Khadas Table of Contents 1 Introduction .................................................................................................................... 4 1.1 System overview .................................................................................................................. 4 1.2 Specification ........................................................................................................................ 5 2 Robot design and simulation ........................................................................................... 5 2.1 The Robotics Simulators ....................................................................................................... 6 2.1.1 Webot Robot Simulator .......................................................................................................... 6 2.1.2 Features .................................................................................................................................. 7 2.2 Technical information .......................................................................................................... 7 2.3 The Robotics World .............................................................................................................. 7 2.4 Simulation with the control of -
Implementation of a Trusted I/O Processor on a Nascent Soc-FPGA Based Flight Controller for Unmanned Aerial Systems
Implementation of a Trusted I/O Processor on a Nascent SoC-FPGA Based Flight Controller for Unmanned Aerial Systems Akshatha J Kini Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the requirements for the degree of Master of Science in Computer Engineering Cameron D. Patterson, Chair Paul E. Plassmann Lynn A. Abbott February 21, 2018 Blacksburg, Virginia Keywords: SoC, UAV, FPGA, ZYNQ, MicroBlaze, autopilot, ArduPilot, ArduPlane, GPS, Mission Planner, Sensor, Vulnerabilities, Mailbox Copyright 2018, Akshatha J Kini Implementation of a Trusted I/O Processor on a Nascent SoC-FPGA Based Flight Controller for Unmanned Aerial Systems Akshatha J Kini (ABSTRACT) Unmanned Aerial Systems (UAS) are aircraft without a human pilot on board. They are comprised of a ground-based autonomous or human operated control system, an unmanned aerial vehicle (UAV) and a communication, command and control (C3) link be- tween the two systems. UAS are widely used in military warfare, wildfire mapping, aerial photography, etc primarily to collect and process large amounts of data. While they are highly efficient in data collection and processing, they are susceptible to software espionage and data manipulation. This research aims to provide a novel solution to enhance the secu- rity of the flight controller thereby contributing to a secure and robust UAS. The proposed solution begins by introducing a new technology in the domain of flight controllers and how it can be leveraged to overcome the limitations of current flight controllers. The idea is to decouple the applications running on the flight controller from the task of data validation. -
Design and Implementation of an Autonomous Robotics Simulator
DESIGN AND IMPLEMENTATION OF AN AUTONOMOUS ROBOTICS SIMULATOR by Adam Carlton Harris A thesis submitted to the faculty of The University of North Carolina at Charlotte in partial fulfillment of the requirements for the degree of Master of Science in Electrical Engineering Charlotte 2011 Approved by: _______________________________ Dr. James M. Conrad _______________________________ Dr. Ronald R. Sass _______________________________ Dr. Bharat Joshi ii © 2011 Adam Carlton Harris ALL RIGHTS RESERVED iii ABSTRACT ADAM CARLTON HARRIS. Design and implementation of an autonomous robotics simulator. (Under the direction of DR. JAMES M. CONRAD) Robotics simulators are important tools that can save both time and money for developers. Being able to accurately and easily simulate robotic vehicles is invaluable. In the past two decades, corporations, robotics labs, and software development groups have released many robotics simulators to developers. Commercial simulators have proven to be very accurate and many are easy to use, however they are closed source and generally expensive. Open source simulators have recently had an explosion of popularity, but most are not easy to use. This thesis describes the design criteria and implementation of an easy to use open source robotics simulator. SEAR (Simulation Environment for Autonomous Robots) is designed to be an open source cross-platform 3D (3 dimensional) robotics simulator written in Java using jMonkeyEngine3 and the Bullet Physics engine. Users can import custom-designed 3D models of robotic vehicles and terrains to be used in testing their own robotics control code. Several sensor types (GPS, triple-axis accelerometer, triple-axis gyroscope, and a compass) have been simulated and early work on infrared and ultrasonic distance sensors as well as LIDAR simulators has been undertaken. -
Conceptual Design Document
Aerospace Senior Projects ASEN 4018 2017 University of Colorado Department of Aerospace Engineering Sciences Senior Projects – ASEN 4018 SHAMU Search and Help Aquatic Mammals UAS Conceptual Design Document 2 October 2017 1.0 Information Project Customers Jean Koster James Nestor 872 Welsh Ct. Louisville, CO 80027 San Francisco, CA Phone: 303-579-0741 Phone: Email: [email protected] Email: [email protected] Team Members Severyn V. Polakiewicz Samuel N. Kelly [email protected] [email protected] 818-358-5785 678-437-6309 Michael R. Shannon Ian B. Barrett [email protected] [email protected] 720-509-9549 815-815-5439 Jesse R. Holton Grant T. Dunbar [email protected] [email protected] 720-563-9777 720-237-6294 George Duong Brandon Sundahl [email protected] [email protected] 720-385-5828 303-330-8634 Benjamin Mellinkoff Justin Norman [email protected] [email protected] 310-562-3928 303-570-5605 Lauren Mcintire [email protected] 267-664-0889 Conceptual Design Assignment Conceptual Design Document 2017 Aerospace Senior Projects ASEN 4018 1.1 List of Acronyms AP Autopilot API Application Programming Interface ARM Advanced RISC Machine A(M)SL Above (Mean) Sea Level CETI Cetacean Echolocation Translation Initiative CLI Command Line Interface CPU Central Processing Unit CONOPS Concept of Operations COTS Commercial Off The Shelf EPS Expanded Polystyrene FBD Functional Block Diagram GCS Ground Control Station GPS Global Positioning System GPU Graphics Processing -
Reef Rover: a Low-Cost Small Autonomous Unmanned Surface Vehicle (USV) for Mapping and Monitoring Coral Reefs
Article Reef Rover: A Low-Cost Small Autonomous Unmanned Surface Vehicle (USV) for Mapping and Monitoring Coral Reefs George T. Raber 1,* and Steven R. Schill 2,* 1 The University of Southern Mississippi, School of Biological, Environmental, and Earth Sciences, Hattiesburg, MS 39406, USA 2 The Nature Conservancy, Caribbean Division, Coral Gables, FL 33134, USA * Correspondence: [email protected] (G.T.R.); [email protected] (S.R.S.); Tel.: +1-1601-434-9489 (G.T.R.); +1-1435-881-7838 (S.R.S.) Received: 8 March 2019; Accepted: 15 April 2019; Published: 17 April 2019 Abstract: In the effort to design a more repeatable and consistent platform to collect data for Structure from Motion (SfM) monitoring of coral reefs and other benthic habitats, we explore the use of recent advances in open source Global Positioning System (GPS)-guided drone technology to design and test a low-cost and transportable small unmanned surface vehicle (sUSV). The vehicle operates using Ardupilot open source software and can be used by local scientists and marine managers to map and monitor marine environments in shallow areas (<20 m) with commensurate visibility. The imaging system uses two Sony a6300 mirrorless cameras to collect stereo photos that can be later processed using photogrammetry software to create underwater high-resolution orthophoto mosaics and digital surface models. The propulsion system consists of two small brushless motors powered by lithium batteries that follow pre-programmed survey transects and are operated by a GPS-guided autopilot control board. Results from our project suggest the sUSV provides a repeatable, viable, and low-cost (<$3000 USD) solution for acquiring images of benthic environments on a frequent basis from directly below the water surface. -
Autonomous Landing of a Multicopter Using Computer Vision
Autonomous Landing of a Multicopter Using Computer Vision Thesis of 30 ECTS submitted to the Department of Computer Science in partial fulfillment of the requirements for the degree of Master of Science in Computer Science Joshua Springer Reykjavík University; Department of Computer Science Mälardalen University; School of Innovation, Design, and Engineering June 12, 2020 Acknowledgements I would like to thank my family for their undying love, support, and encouragement which have gotten me this far. My thanks go to Mälardalen University, Reykjavík University, and the Nordic Council, without whose generosity in the MDH Scholarship and the Nordic Master Programme, I would not have had the opportunity to study my passion in two wonderful countries. 2 Abstract Takeoff, normal flight, and even specialized tasks such as taking pictures, are essentially solved problems in autonomous drone flight. This notably excludes landing, which typically requires a pilot because of its inherently risky nature. This project attempts to solve the problem of autonomous drone landing using computer vision and fiducial markers - and specifically does not use GPS as a primary means of navigation during landing. The system described in this thesis extends the functionality of the widely-used, open-source ArduPilot software which runs on many drones today, and which has only primitive functionality for autonomous landing. This system is implemented as a set of ROS modules which interact with ArduPilot for control. It is designed to be easily integrated into existing ArduPilot drone systems through the addition of a companion board and gimbal-mounted camera. Results are presented to evaluate the system’s performance with regard to pose estimation and landing capabilities within Gazebo 9 simulator.