Analysis and Deployment of an OCR - SSD Deep Learning Technique for Real-Time Active Car Tracking and Positioning on a Quadrotor

Luiz Gustavo Miranda Pinto1

1 Institute of Mathematics and Computation – IMC, Federal University of Itajubá. Brazil 2

ABSTRACT

This work presents a deep learning solution object tracking and object detection in images and real-time license plate recognition implemented in F450 quadcopter in autonomous flight. The solution uses Python programming language, OpenCV library, remote PX4 control with MAVSDK, OpenALPR, neural network using Caffe and TensorFlow.

1. INTRODUCTION

A drone can follow an object that updates its route all the time. This is called active tracking and positioning, where an autonomous vehicle needs to follow a goal without assistance from human intervention. There are some approaches to this mission with drones, but it is rarely used for object detection and OCR due to resource consumption. State-of-the-art algorithms can identify the class of a target object being followed. This work presents and analyzes a technique that grants control to a drone during an autonomous flight, using real-time tracking and posi- tioning through an OCR system for deep learning model of plate detection and ob- ject detection.

2. MATERIALS AND METHODS

The following will present the concepts, techniques, models, materials and methods used in the proposed system, in addition to the structures and platforms used for its creation.

2.1. Rotary Wing UAV

This project used an F-450 quadcopter drone for outdoor testing and a Typhoon H480 octorotor for the simulation. A quadcopter is an aircraft made up of 4 rotors carrying the controller board in the middle and the rotors at the ends. It is con- trolled by changing the angular speeds of the rotors that are rotated by electro- magnetic motors, where you can have 6 degrees of freedom, as seen in Figure 1, with x, y and z as the transition movement, and roll, pitch and yaw as the rotation- al movement.

The altitude and position of the drone can be modified by adjusting engine speeds. The same applies to pitch control but controlling the rear or front engines. 3

Figure 1 – Six degrees of freedom. Source: (STRIMEL; BARTHOLOMEW; KIM, 2017)

2.1.1. Pixhawk Fligh Control Hardware This project was implemented using the Pixhawk flight controller (Figure 2), an independent open hardware flight controller. Pixhawk supports manual and au- tomatic operations, being suitable for research, because it is inexpensive and com- patible with most remote control transmitters and receivers. Pixhawk is built with a dual processor with 32-bit computing capacity STM32f427 Cortex M4 MHz processor / 256 cores 168KB of RAM / 2MB of flash bit. The current version is the Pixhawk 4.

Figure 2 – Pixhalk 2.4.8 flight controller. Source: (KANTUE; PEDRO, 2019)

4

2.1.2. PX4 Flight Control Software In this work, the PX4 flight control software was used, because is the Pix- hawk's native software, avoiding compatibility problems. PX4 is open source flight control software for drones and other unmanned vehicles that provides a set of tools to create customized solutions.

There are several internal types of frames with their own flight control parame- ters, including engine assignment and numbering, which includes the F450 used in this project. These parameters can be modified to obtain refinement during a flight. In the case of PX4, it uses Proportional, Integral, Derivative (PID) control- lers, which are one of the most widespread control techniques.

In the PID controllers, the P (proportional) gain is used to minimize the track- ing error. It is responsible for a quick response and therefore should be as high as possible. Gain (derivative) is used to moisten. It is necessary, but only the maxi- mum necessary to avoid overtaking must be defined. Gain I (integral) maintains an error memory. The term “I” increases when desired the rate has not been reached for some time. The idea is to parameterize the in-flight board model using GCS (Ground Control Station) software, where it is possible change these parame- ters and check their effects on each of the drone´s degrees of freedom.

2.2. Ground Control Station

A ground control station (GCS), described in the previous chapter, needs to check the behavior of the drone during the flight and was used to update the drone's firmware, adjust its parameters and calibrate your sensors. Running on a base computer, communicate with the UAV wirelessly, such as telemetry or Wi- Fi, display real-time data on the performance and position of the UAV, and show data from many instruments present on a conventional plane or helicopter.

This work used the QGroundControl GCS due to its affinity with the PX4 and PID tools, which allowed changes to the PID while the drone was still in the air. QGroundControl has the ability to read telemetry data simultaneously from multi- ple aircrafts if they are using the same version of MAVLink, while still being more common features such as telemetry logging, visual display of the GUI, a mission planning tool, the ability to adjust the PID gains during the flight, as shown in Figure 3, and the ability to display vital information flight data infor- mation. 5

Figure 3 – QGroundControl PID tuning Source: (QGROUNDCONTROL, 2019b)

2.3. Drone Software Development Kit (SDK)

During the development of this project, some SDKs were used to obtain auton- omous control of software on drones. They were used in different parts of the pro- ject and the reason it was the versatility of each other. The idea was to choose the SDK that had the best balance between robustness and simplicity. The selection included MAVROS, MAVSDK and DroneKit because of their popularity. All SDKs included the MAVLink protocol, responsible for controlling the drone.

DroneKit presented the best simplicity, but it did not have the best support for the PX4. External control that accepts remote commands via the programming was developed for ArduPilot applications 10.

The MAVROS package running ROS presented the best system in terms of ro- bustness, but it was complex to manage. MAVSDK presented the best result. It has full support for PX4 applications and is not complex to manage, being the sub- ject of choice for this project.

2.3.1. MAVLink and MAVSDK MAVLink is a very useful messaging protocol for exchanging information with a drone and between the controller board components. During its use, data flows, such as telemetry status, are published as topics, while subprotocol, like those used for mission or parameter control, are point to point with retransmission.

6

MAVSDK is a MAVLink library with APIs implemented in different lan- guages, such as ++ and Python. MAVSDK can be used on a computer embed- ded in a vehicle, at a land or mobile station device, which has significantly more processing power than an ordinary flight controller, be able to perform tasks such as computer vision, obstacle prevention and route planning. MAVSDK was the chosen framework, due to its affinity with the PX4 and its simplicity.

2.3.2. ROS and MAVROS The Robot (ROS), also presented in the previous chapter, is a structure widely used in robotics. ROS offers features such as distributed com- puting, message passing and code reuse.

ROS allows the robot control to be divided into several tasks called nodes and are processes that perform calculations, allowing modularity. The nodes exchange messages with each other in an editor-subscriber system, where a topic acts as an intermediate store for some of the nodes to publish its content while others sub- scribe to receive this content.

ROS has a general manager called “master”. The ROS master, as seen in Figure 4, is responsible for providing names and records to services and nodes in the ROS system. It tracks and directs editors and subscribers to topics and Services. The role of the master is to guide the individual ROS nodes to locate each from others. After the nodes are located, they define their communication via point to point.

Figure 4 – The ROS Master Source: (HASAN, 2019)

To support collaborative development, the ROS system is organized in packag- es, which are simply directories that contain XML files that describe the package 7

and presenting any dependencies. One of these packages is the MAVROS, which is an extensible communication node MAVLink with proxy for GCS. The package provides a driver for communication between a variety of autopilots with MAVLink communication protocol and a UDP MAVLink bridge for GCS. The MAVROS package allows MAVLink communication between different comput- ers running ROS, being currently the official supported bridge between ROS and MAVLink.

2.3.3. DroneKit DroneKit is an SDK built with development tools for drones. It allows to create applications which runs on a host computer and allows communication with Ar- duPilot flight controllers. Applications can to insert a level of intelligence into the vehicle's behavior, and can perform tasks with high computational performance they cost or depend in real time, such as computer vision or path planning.

Currently, the PX4 is not yet fully compatible, being more suitable for ArduPi- lot applications.

2.4. Simulation

In this project, the Gazebo platform was the choice to imitate environment through PX4 SITL. However, it was not the only option, since the PX4 SITL is available for other platforms, such as jMAVSim, AirSim and X-Plane. JMAVSim was not easy to integrate obstacles or Extra sensors, such as cameras, are discard- ed mainly for this purpose reason. AirSim was also discarded because, while real- istic, it requires a powerful GPU (Graphics Processing Unit), which could com- pete for resources during the object detection phase.

The X-Plane, also discarded, is realistic and has a variety of UAV models and environments already implemented, however, it depends on the licensing for its use. Thus, Gazebo was the option chosen, due to its simulation environment with a variety of online resource models, the ability to import meshes from other model- ing software, such as Solidworks or Inkscape, and the free license.

2.4.1. Gazebo Gazebo, already presented in the previous chapter, is an open source simulation platform that allows integration with ROS. Gazebo is currently only supported on Linux, but there are some speculation about a version of Windows 8

2.4.2. PX4 SITL The PX4 firmware offers a complete hardware simulation, with a response that provides the entry of the environment using its own SITL (Software in the loop). The simulation reacts to the simulation data given exactly how it would react in reality and evaluates the total production energy required in each rotor.

The PX4 SITL allows you to simulate the same software as on a real platform, rigorously replicating the behaviors of an autopilot. Can simulate the same autopi- lot used on a real drone and its MAVLink protocol, which generalizes direct use of a real drone. The greatest the PX4 SITL is that a flight controller cannot distin- guish whether it is running on simulation or inside a real vehicle, allowing the simulation code to be imported directly for commercially available UAV plat- forms.

2.5. Deep Learning Frameworks

Deep learning is a type of machine learning, generally used for classification, regression and resource extraction tasks, with multiple layers of representation and abstraction. For object detection, resource extraction tasks are required and can be achieved using convolutional neural networks (CNN), a class of deep neural net- works that apply filters at various levels to extract and classify visual information from a source, such as an image or video. This project used a CNN to detect visual targets using a camera.

2.5.1. Caffe Framework In this project was used the Caffe deep learning framework. Caffe provides a complete toolkit for training, testing, fine-tuning and model deployment, with well-written documentation examples for these tasks. It is developed under a free BSD license, being built with the C++ language and maintaining Python and MATLAB links for training and deploy general purpose convolutional neural net- works and many other deep models efficiently.

9

2.6. Object Detection

The most common way to interpret the location of the object is to create a bounding box around the detected object, as seen in Figure 5.

Figure 5 – Bounding boxes from object detection Source: (GANESH, 2019)

Object detection, detailed in the previous chapter, was the first stage in this tracking system, as it focuses on the object to be tracked.

2.6.1. Convolutional Neural Networks (CNN) A convolutional neural network (CNN), as described in the previous chapter, is a variation of the so called multilayer perceptron networks and was inspired by the biological process of data processing.

2.6.2. Single-Shot MultiBox Detectot – SSD SSD is an object detection algorithm that uses a deep learning model for neural networks. It was designed for real-time applications, like this one. It is lighter than other models, as it speeds up the process of inferring new bounding boxes reuse of pixels or feature maps, which are the result of convolutional blocks and represen- tation of the dominant characteristics of the image at different scales.

Its core was built around a technique called MultiBox, which is a method for proposals for coordinates of class agnostic bounding boxes. Regarding its perfor- mance and accuracy, for applicability in object detection, it has a score above 74% mAP at 59 frames per second in data sets like COCO and PascalVOC. 10

2.6.2.1. MultiBox MultiBox is a method for bounding box regression that achieves dimensionality reduction, as it consists of branches of convolutional layers, as seen in Figure 6 that resize images over the network, maintaining the original width and height.

Figure 6 – MultiBox architecture Source: (FORSON, 2017)

The magic behind the MultiBox technique is the interaction between two criti- cal assessment factors: loss of confidence (CL) and loss of location (LL). CL is a factor that measures how confident the class selection is made, which means whether it is the correct class of the object, using categorical cross entropy in rela- tion to entropy. We can consider cross entropy as a received response that is not optimal. Entropy on the other hand, it represents the ideal answer. Therefore, knowing entropy, all entropy received can be measured in terms of how far these responses are from the ideal.

2.6.2.2. SSD Architecture The SSD is composed by the extraction of maps of resources, through an in- termediate neural network called a resource extractor and the application of con- volution filters to detect objects. The SSD architecture consists of 3 main compo- nents: basic network, extra layers for resources extraction and forecasting layers.

11

Figure 7 – VG-16 architecture Source: (FROSSARD, 2016) In the basic network, extraction is performed using a convolutional neural net- work called VGG-16 - the resource extractor, as seen in Figure 7, where it is made up of combinations of convolution layers with ReLU for fully connected layers. In addition, in addition, it has layers of maximum pool and a final layer with a soft- max activation function. The output of this network is a resource map with dimen- sions 19x19x1024. Right after the basic network, 4 additional additional convolu- tional layers are added to continue reducing the size of the resource map until its final dimensions are 1 x 1 x 256. Finally, the forecast layers, a crucial element of the SSD, use a variety of resource maps representing various scales to predict class scores and bounding box coordinates. The final composition of the SSD can be seen in Figure 8.

Figure 8 – SSD architecture Source: (DALMIA, 2019)

2.7. Optical Character Recognition - OCR

In this work, the identification of the vehicle license plate was important for the drone be able to follow a car with the correct license plate. This problem occurred with the optical character recognition system (OCR). OCR is a technique respon- sible for recognizing optically drawn characters. OCR is a complex problem to deal with due to the variety of languages, fonts and styles in which characters and information can be written, including the complex rules for each of these lan- guages.

2.7.1. Inference Process An example of the steps in the OCR technique is shown in Figure 9. The steps are as follows: Acquisition, Pre-processing, Segmentation, Resource Extraction and Recognition. 12

 Acquisition: a recorded image is fed into the system.

Figure 9 – OCR inference process steps Source: (ADAMS, 1854)

 Pre-processing: eliminates color variations by smoothing and normalizing pixels. Smoothing applies convolution filters to the image to remove noise and smooth the edges. Normalization finds a uniform size, slope and rotation for all characters in the image.

 Segmentation: find the words written inside the image.

 Resource extraction: extracts the characteristics of the symbols.

 Recognition: actually identifies the characters and classifies them, search- ing the lines, word for word, converting the images for character streams representing letters of recognized words.

2.7.2. Teressact OCR This work used a plate detection called OpenALPR that uses Google , an open source OCR framework to train networks for different languages and scripts. It convert the image into binary images. Identifies and extracts character outlines. Transforms Blobs contours, which are small regions isolated from digital images. Divide text into words using techniques like cloudy spaces and defined spaces. Finally, it recognizes the text by classifying and storing each recognized word.

13

2.7.3. Automatic License Plate Recognition - ALPR ALPR is a way to detect characters that makes up the license plate of the vehi- cle and uses OCR for most of the process. It combines object detection, image processing and pattern recognition. It is used in real-life applications, such as au- tomatic toll collection, traffic law enforcement, access control to parking lots and road traffic monitoring. The four steps of the ALPR are shown in Figure 10.

Figure 10 – ALPR steps Source: (SARFRAZ; AHMED; GHAZI, 2003)

2.7.3.1. OpenALPR This project used OpenALPR, an open source ALPR library built with the C ++ language and has links in C #, Java, Node.js, and Python. The library receives im- ages and video streams for analysis in order to identify registrations and generates a text representing the enrollment characters. It is based on OpenCV, an open source computer vision library for image analysis and Tesseract OCR.

14

3. IMPLEMENTED MODEL

This project used data sets, SSD training in TensorFlow and Caffe frameworks, image pre-processing for OCR and tracking and motion functions.

3.1. Supervioned Learning

Supervised Learning (SL) is a kind of Machine Learning (ML) training in which the model is provided with labeled data during your training phase. The La- belImg tool was used, with PascalVOC format like the standard XML annotation (Figure 11), which includes class labels, coordinates of the bounding boxes, image path, image size and name and other tags.

Figure 11 – PascalVOC’s XML example Source: (EVERINGHAM et al.,2010)

3.1.1. Datasets Three sets were used: INRIA Person, Stanford Cars and GRAZ-02.

The Stanford Cars data set allowed SSD to identify cars along the quadcopter trajectory. This dataset contains 16,185 images from 196 classes of cars, divided into 8,144 training images and 8,041 test images, already noted in terms of make, model and year.

15

The INRIA person data set is a collections of digital images to highlight peo- ple, taken over a long period of time, and some web images taken from Google Images. About 2500 images were collected from that data set. The class of person was added because it was the most common false positive found in single-class training, along with a background class, to help improve the model discernment between different environments.

3.1.2. Caffe and TensorFlow Training Unlike TensorFlow, Caffe does not have a complete object detection API, which makes it more complicated when starting a new training. Caffe does not in- clude a direct visualization tool like Tensorboard. However, it includes in its tools subdirectory, the parse log.py script, which extracts all relevant training infor- mation from the log file and makes it suitable for plotting. Using a Linux plotting tool called gnuplot, in addition to an analysis script, it was possible to build a real- time plotting algorithm (Figure 12). The latest log messages indicated that the model reached an overall mAP of 95.28%, which is an accuracy gain of 3.55% compared to the first trained model.

Figure 12 – Results from gnuplot custom script Source: (Author)

3.2. Image Preprocessing for ALPR

To help ALPR identify license plates, the quality of the images has been im- proved through the use of two techniques of image preprocessing brightness varia- tion and sharp mask. The variation in brightness is controlled in the color system 16

used was RGB, where the color varies according to the levels of red, green and blue provided. There are 256 possible values for each level, ranging from 0 to 255. To change the brightness, just you need to add or subtract a constant value from each level. For brighter images, the value is add, while for darker images the val- ues are subtracted, as seen in Figure 13, providing a total -255 to 255 values. The only necessary care is to check whenever the addition or subtracted value will be greater than 255 and less than 0.

Figure 13 – Effect of brightness on a colleted image Source: (Author)

After applying the brightness, the sharpness mask, also called the sharpness fil- ter, was it is necessary to highlight the edges, thus improving the characters of the plate. Figure 14 shows an example of the result of this process.

Figure 14 – Effect of sharpness filter application Source: (Author)

3.3. Drone Control Algorithm

The algorithm for controlling the drone was built with the assistance of librar- ies in the Python language, which include OpenALPR, Caffe, TensorFlow, MAVSDK, OpenCV and others. Algorithm 1 is the main aogorithm of the solu- 17

tion and evaluates updated data from object detection and OCR at each speed on the x, y, z axes of the three-dimensional real-world environment.

3.3.1. Height Centralization Function Height centering is the only control function shared between views. It positions the drone at a standard altitude (Figure 14).

Figure 15 – Height centralization Source: (Author)

3.3.2. 2D-Centralization Function The drone has its camera pointed at the ground, the captured image is analo- gous to a 2D Cartesian system and the centralization is oriented using the x and y coordinates, as shown in Figure 16. The idea is to reduce the two Values ∆x and ∆y, which represent the detection x and y distances respectively central point (pink) to the central point of the frame (blue). 18

Figure 16 – 2D-centralization Source: (Author)

3.3.3. Yaw Alignment Function In the rear view approach, the alignment function is used to center the drone horizontally, according to the position of the car. The idea is to keep the drone fac- ing the car, as shown in Figure 17. These values are calculated using the a ∆yaw distance between the frame and the central detection point on the Y axis.

Figure 17 – Yaw alignment Source: (Author)

3.3.4. Approximation Function In the rear view approach, the zoom function was the most difficult to find, and used the distance from the object in the camera, the speed of the car and the bal- ance between a safe distance from the car and the minimum distance for the OCR to work. The distance from the object was calculated using the relationship be- 19

tween the object camera field of view (FOV) and its sensor dimension, as seen in Figure 18.

Figure 18 – Object distance from camera Source: (Author)

4. EXPERIMENTS AND RESULTS

To evaluate the algorithm, a notebook was used as Ground Station, coordinat- ing the acquisition of the frame, the processing of the frame by the SSD, the val- ues calculation and drone positioning command. Its specifications were an Ubuntu v18.04 64 OS, Intel R CoreTM i7 with 2.2 GHz, 16 GB of DDR3 SDRAM and an Nvidia GeForce GTX 1060 graphics card, with 6 GB of G-DDR5 memory.

4.1. The Simuled Drone

For the simulated experiment, a Typhoon H-480 model was used in the Gaze- bo, as shown in Figure 19. It is available in the PX4 Firmware Tools directory on Github, available at: https://github.com/PX4/Firmware/tree/master/Tools, ready to use on SITL environment. It was a handful, as it had a built-in gimbal and camera. The gimbal allowed the images to be captured in a very stable way, avoiding compromising the detections.

20

Figure 19 – Typhoon H-480 in Gazebo Source: (Author) 4.2. Simulated World

In the simulated experiments, a customized city (Figure 20) was created with precompilation models in the Gazebo model database, available at: https://bitbucket.org/osrf/gazebo models.

Figure 20 – Custom world in Gazebo Source: (Author)

4.3. CAT Vehicle

The CAT vehicle is a simulated distributed autonomous vehicle that is part of the ROS project in order to support research on autonomous driving technology. CAT has complete configurable simulated sensors and actuators imitating a real- world vehicle capable of autonomous driving, which includes a steering speed control implemented in real time simulation.

In the experiments the drone was adjusted some distance from the rear end of the car, as seen in Figure 21, and followed it, capturing its complete image and al- lowing the algorithm to process the recognized card. In addition, it allowed the plate to be personalized with a Brazilian model, making it very convenient for this project.

21

Figure 21 – The CAT Vehicle and Typhoon H-480 in Gazebo Source: (Author) 4.4. Simulated Experiments

The experiments related that in an urban scene, the most cars could be detected within a range of 0.5 to 4.5m from the camera, as shown in the green area in Fig- ure 22.

Figure 22 – Car detection simulation Source: (Author)

The detection range was 1.5 to 3 m. The balance between the main number of detections with the total of correct information extracted is represented in the green area of Figure 23.

Figure 23 – Plate detection and processing results Source: (Author) 22

Figure 24 shows a collision hazard zone represented by a red area.

Figure 24 – Car rear following results Source: (Author)

The ideal safety height and image quality vary between 4m and 6m. Figure 25 shows the red area where the height of other possible objects to be found, such as people and other vehicles, making it difficult to identify the moving vehicle as an object to be tracked.

Figure 25 – Car top following results Source: (Author)

4.5. The Real Quadrotor

A custom F450 (Figure 26) was used for the outdoor experiments. The chal- lenge was to use cheap materials to achieve reasonable results.

The Pixhawk 4, the most expensive component, was the flight controller cho- sen board, as the idea was to use the PX4 flight stack as the control board configu- ration.

23

The numbers shown in Figure 26 are reference indexes for each of the compo- nents shown in Table 1. For the acquisition of the frame, an EasyCap capture de- vice was connected to the computer, previously connected to the video receiver.

Figure 26 – Customized F450 built Source: (Author)

Table 1 – Customized F450 specifications

* Not added to weight sum, since it was used in the ground station ** The price is equivalent to USD 190,77 on March 08th, 2020

4.6. Outdoor Experiment

For the outdoor experiment on Campus (Figure 27), the model was changed to detect people instead of cars. To avoid colliding with the person who served as a target, everything was coordinated very slowly.

Another experiment used a video as images source (Figure 28), to check how many detections, plates and right information extracted the technique could ac- 24

quire. For the video, a record from Marginal Pinheiros was used, being one of the busiest highways in the city of São Paulo.

Figure 27 – Outdoor experiment in Campus Source: (Author)

Figure 28 – Experiment on record Source: (OLIVEIRA NA ESTRADA, 2018)

The experiment produced 379 vehicle detections of the 500 existing in a video clip, where 227 plaques were found, 164 with the correct information extracted (Figure 29).

Figure 29 – Experiment on record results Source: (author)

25

5. CONCLUSIONS AND FUTURE WORKS

In this project it was observed that a Single Class training for object detection models did not meet the needs of real time object detection for the experiments carried out. Multi Class allowed us to achieve good results.

The distance and the brightness level determined the limits in the tests per- formed, being an aspect to work on future improvements. A high definition cam- era should be used to prevent noise and vibrations in the captured images.

The mathematical functions used to calculate the drone's speed were useful in understanding the drone's behavior.

A different approach to position estimation or PID controller can be used to de- termine the object's route.

REFERENCES 3D ROBOTICS. DroneKit. 2015. Accessed on: December 21st, 2019. Available at: .

3D ROBOTICS. About DroneKit. 2016. Accessed on: December 21st, 2019. Available at: .

ABADI, M. et al. Tensor ow: Large-scale machine learning on heterogeneous dis- tributed systems. ArXiv, abs/1603.04467, 2015.

. Tensor ow: A system for large-scale machine learning. In: Proceedings of the 12th USENIX Conference on Operating Systems Design and Implementation. USA: USENIX Association, 2016. (OSDI’16), p. 265{283. ISBN 9781931971331.

ADAMS, H. G. Nests and Eggs of Familiar Birds. 1st ed. 5 Paternoster Row, London, England: London: Groombridge and Sons, 1854. Vol. IV.

26

ALLOUCH, A. et al. Mavsec: Securing the mavlink protocol for ardupilot/px4 unmanned aerial systems. 2019 15th International Wireless Communications &Mobile Computing Conference (IWCMC), IEEE, Jun 2019. Available at: .

ALY, M. Survey on Multiclass Classi cation Methods. 1200 East California Boulevard Pasadena, California, USA, 2005.

AMIT, Y.; FELZENSZWALB, P. Object detection. In: Computer Vision. Spring- er US, 2014. p. 537{542. Available at: .

ANAGNOSTOPOULOS, C.-N. et al. License plate recognition from still images and video sequences: A survey. IEEE Transactions on Intelligent Transportation Systems, Institute of Electrical and Electronics Engineers (IEEE), v. 9, n. 3, p. 377{391, sep 2008. Available at: .

ANDERSON, M. J.; WHITCOMB, P. J. Two-level factorial design. In: DOE Simpli ed: Practical Tools for E ective Experimentation. 3rd ed. [N.a.]: Productivi- ty Press, 2017. Chapter 3.

ARDUPILOT. APM Planner 2 Home. 2019. Accessed on: December 16th, 2019. Available at: .

. Choosing a Ground Station: Overview. 2019. Accessed on: December 10th, 2019. Available at: .

. Mission Planner Overview. 2019. Accessed on: December 16th, 2019. Available at: .

. Archived: APM 2.5 and 2.6 Overview. 2020. Accessed on: February 26th, 2020. Available at: .

. ArduPilot. 2020. Accessed on: February 26th, 2020. Available at: .

BAHRAMPOUR, S. et al. Comparative study of ca e, neon, theano, and torch for deep learning. ArXiv, abs/1511.06435, 2015.

27

BARTAK, R.; VYKOVSKY, A. Any object tracking and following by a ying drone. In: 2015 Fourteenth Mexican International Conference on Arti cial Intelli- gence (MICAI). IEEE, 2015. Available at: .

BARTON, T. E. A.; AZHAR, M. A. H. B. Forensic analysis of popular uav sys- tems. In: 2017 Seventh International Conference on Emerging Security Technolo- gies (EST).IEEE, 2017. Available at: .

BENDEA, H. et al. Low cost uav for post-disaster assessment. The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sci- ences, v. 37, 01 2008.

BHADANI, R. K.; SPRINKLE, J.; BUNTING, M. The CAT vehicle testbed: A simulator with hardware in the loop for autonomous vehicle applications. Elec- tronic Proceedings in Theoretical Computer Science, Open Publishing Associa- tion, v. 269, p. 32{47, apr 2018. Available at: .

BHATIA, R. Tensor ow vs ca e: Which machine learning framework should you opt for? Analytics India Magazine, Analytics India Magazine Pvt Ltd, aug 2018. Accessed on: January 14th, 2020. Available at: .

BOUREZ, C. About loss functions, regularization and joint losses : multinomial logistic, cross entropy, square errors, euclidian, hinge, Crammer and Singer, one versus all, squared hinge, absolute value, infogain, L1 / L2 - Frobenius / L2,1 norms, connectionist temporal classi cation loss. 2016. Accessed on: December 08th, 2019. Available at: .

BRADSKI, G.; KAEHLER, A. Learning : Computer vision with the opencv library. In: rst. Sebastopol, CA, USA: O’Reilly, 2008. book 1.

BRAGA, R. G. et al. Collision avoidance based on reynolds rules: A case study using quadrotors. In: Advances in Intelligent Systems and Computing. Springer International Publishing, 2017. p. 773{780. Available at: .

BREEDLOVE, L. An insider’s look at the rise of drones: Industry veteran Lon Breedlove gives his perspective on the evolution and future of the drone industry. 28

2018. Accessed on: November 15th, 2019. Available at: .

BRITO, P. L. de et al. A technique about neural network for passageway detec- tion. In: . 16th International Conference on Information Technology-New Genera- tions (ITNG 2019). Springer International Publishing, 2019. p. 465{470. Availa- ble at: .

BUHUS, E. R.; TIMIS, D.; APATEAN, A. Automatic parking access using openalpr on raspberry pi3. In: Journal of ACTA TECHNICA NAPOCENSIS Electronics and Telecommunications. Cluj, Romania: Technical University of Cluj-Napoca, 2016. (3, v. 57), p. 10{15.

CABEBE, J. Google Translate for Android adds OCR. 2012. Available at: .

CARDAMONE, A. IMPLEMENTATION OF A PILOT IN THE LOOP SIMULATION ENVIRONMENT FOR UAV DEVELOPMENT AND TESTING. Doctoral Thesis (Graduation Project) | Scuola di Ingegneria Industriale e dell’Informazione, Politecnico di Milano, Milano, Lombardia, Italia, jul 2017.

CHAPMAN, A. Types of drones: Multi-rotor vs xed-wing vs single rotor vs hy- brid vtol. DRONE Magazine, I, n. 3, Jun 2016.

CHIOU, Y.-C. et al. Optimal locations of license plate recognition to enhance the origin-destination matrix estimation. Proceedings of the Eastern Asia Society for Transportation Studies, v. 2011, p. 297{297, 2011.

COUSINS, S. ROS on the PR2 [ROS topics]. IEEE Robotics & Automation Mag- azine, Institute of Electrical and Electronics Engineers (IEEE), v. 17, n. 3, p. 23{25, set. 2010. Available at: .

CUETO, M.; SILVA, A. da; ZANETTI, M. Desenvolvimento de software para automatizac~ao da operac~ao de aeronave remotamente pilotada pautado pelo rtca do-178c e do-278a. In: 1o Congresso Aeroespacial Brasileiro { Edic~ao Especial Pl^eiade. Foz do Iguacu, PR, Brazil: [n.p.], 2018.

DAI, J. et al. R-fcn: Object detection via region-based fully convolutional net- works. In: Proceedings of the 30th International Conference on Neural Infor- mation Processing Systems. Red Hook, NY, USA: Curran Associates Inc., 2016. (NIPS’16), p. 379{387. ISBN 9781510838819.

29

DALAL, N.; TRIGGS, B. Histograms of oriented gradients for human detection. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05). IEEE, 2005. Available at: .

DALMIA, A. Real-time object detection: Understanding SSD. 2019. Accessed on: November 28th, 2019. Available at: .

DENG, L. Deep learning: Methods and applications. Foundations and Trends R in Signal Processing, Now Publishers, v. 7, n. 3-4, p. 197{387, 2014. Available at: .

DIPIETRO, R. A Friendly Introduction to Cross-Entropy Loss. 2016. Accessed on: December 06th, 2019. Available at: .

DJI. NAZA-M. 2020. Accessed on: February 26th, 2020. Available at: .

DOGRA, A.; BHALLA, P. Image sharpening by gaussian and butterworth high pass lter. Biomedical and Pharmacology Journal, Oriental Scienti c Publishing Company, v. 7, n. 2, p. 707{713, dec 2014. Available at: .

DRONEKIT. DRONEKIT: Your Aerial Platform. 2015. Accessed on: December 21st, 2019. Available at: .

DU, S. et al. Automatic license plate recognition (ALPR): A state-of-the-art re- view. IEEE Transactions on Circuits and Systems for Video Technology, Institute of Electrical and Electronics Engineers (IEEE), v. 23, n. 2, p. 311{325, feb 2013. Available at: .

EIKVIL, L. OCR - Optical Character Recognition. Gaustadalleen 23, P.O. Box 114 Blindern, N-0314 Oslo, Norway, 1993.

EVERINGHAM, M. et al. The pascal visual object classes (voc) challenge. Inter- national Journal of Computer Vision, v. 88, n. 2, p. 303{338, jun. 2010.

FAIRCHILD, C. Getting started with ros. In: ROS robotics by example : bring life to your robot using ROS robotic applications. Birmingham, England: Packt Pub- lishing, 2016. Chapter 1, p. 5{14. ISBN 978-1-78217-519-3.

30

FATHIAN, K. et al. Vision-Based Distributed Formation Control of Unmanned Aerial Vehicles. 2018.

FENG, L.; FANGCHAO, Q. Research on the hardware structure characteristics and EKF ltering algorithm of the autopilot PIXHAWK. In: 2016 Sixth Interna- tional Conference on Instrumentation & Measurement, Computer, Communication and Control (IMCCC). IEEE, 2016. Available at: .

FERDAUS, M. M. et al. Fuzzy clustering based nonlinear system identi cation and controller development of pixhawk based quadcopter. In: 2017 Ninth International Conference on Advanced Computational Intelligence (ICACI). IEEE, 2017. Available at: .

FISHER, R. et al. Unsharp Filter. 10 Crichton Street, Edinburgh, EH8 9AB, Scot- land, UK: The University of Edinburgh, 2003. Hypermedia Image Processing Reference (HIPR), School of Informatics, The University of Edinburgh. Accessed on: January 14th, 2020. Available at: .

FORSON, E. Understanding SSD MultiBox | Real-Time Object Detection In Deep Learning. 2017. Accessed on: November 26th, 2019. Available at: .

FROSSARD, D. VGG in TensorFlow: Model and pre-trained parameters for VGG16 in TensorFlow. 2016. Accessed on: November 28th, 2019. Available at: .

FULTON, W. Math of Field of View (FOV) for a Camera and Lens. 2020. Ac- cessed on: January 16th, 2020. Available at: .

GANESH, P. Object Detection : Simpli ed. 2019. Accessed on: January 02nd, 2020. Available at: .

GAO, H. Mobilenet Based Single Short Multi-box Detector in Pytorch, ONNX and Ca e2. 2018. Accessed on: January 10th, 2020. Available at: .

31

GARCIA, S. I. L0 Norm, L1 Norm, L2 Norm & L-In nity Norm. 2018. Accessed on: December 06th, 2019. Available at: .

GEDOUIN, P.-A. et al. Experimental comparison of classical PID and model-free control: Position control of a shape memory alloy active spring. Control Engineer- ing Practice, Elsevier BV, v. 19, n. 5, p. 433{441, may 2011. Available at: .

GOOGLE DEVELOPERS. Classi cation: True vs. False and Positive vs. Nega- tive.2019. Accessed on: December 08th, 2019. Available at: .

. What is Supervised Learning? 2019. Accessed on: January 13th, 2020. Available at: .

GREGERSEN, H. Integrating physical devices with IOTA | Car-IOTA Part 1. 2019. Accessed on: January 10th, 2020. Available at: .

HAMPAPUR, A. et al. Smart video surveillance: exploring the concept of mul- tiscale spatiotemporal tracking. IEEE Signal Processing Magazine, Institute of Electrical and Electronics Engineers (IEEE), v. 22, n. 2, p. 38{51, mar. 2005. Available at: .

HASAN, K. S. B. What, Why and How of ROS. 2019. Accessed on: December 19th, 2019. Available at: .

HEIDARYSAFA, M. et al. From videos to urls: A multi-browser guide to extract user’s behavior with optical character recognition. Advances in Computer Vision, Springer International Publishing, p. 503{514, Apr 2019. ISSN 2194-5365. Avail- able at: .

HENTATI, A. I. et al. Simulation tools, environments and frameworks for UAV systems performance analysis. In: 2018 14th International Wireless Communica- tions & Mobile Computing Conference (IWCMC). IEEE, 2018. Available at: .

32

HOLDEN, D.; SAITO, J.; KOMURA, T. A deep learning framework for character motion synthesis and editing. ACM Transactions on Graphics, Association for Computing Machinery (ACM), v. 35, n. 4, p. 1{11, jul. 2016. Available at: .

HONG, Y.; FANG, J.; TAO, Y. Ground control station development for autono- mous UAV. In: Intelligent Robotics and Applications. Springer Berlin Heidelberg, 2008. p. 36{44. Available at: .

HOWARD, A. G. et al. MobileNets: E cient Convolutional Neural Networks for Mobile Vision Applications. 2017.

HUANG, J. et al. Speed/accuracy trade-o s for modern convolutional object detec- tors In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

IEEE, 2017. Available at: .

HUDSON, J. impload - Simple CLI tool for uploading misson plans to iNav. 2018. Accessed on: December 16th, 2019. Available at: .

HUI, J. mAP (mean Average Precision) for Object Detection. 2018. Accessed on: December 08th, 2019. Available at: .

. SSD Object Detection: Single Shot Multibox Detector for Real-Time Processing. 2018. Accessed on: November 27th, 2019. Available at: .

HULSTAERT, L. A Beginner’s Guide to Object Detection. 2018. Accessed on: March

07th, 2020. Available at: .

HUNG, D. et al. Autonomous ground reconnaissance drone using robot operating system (ros). Proceedings of the International Telemetering Conference, 1 2017. ISSN 0884-5123.

HONEYWELL INC. Airborne drone formation control system. 1994. US5521817A. 33

ISLAM, N.; ISLAM, Z.; NOOR, N. A survey on optical character recognition sys- tem. In: Journal of Information & Communication Technology-JICT. 06010 UUM Sintok Kedah Darul Aman, Malaysia: Universiti Utara Malaysia Press, 2016. (Is- sue. 2, v. 10).

JAIN, Y. Tensor ow or PyTorch : The force is strong with which one? 2018. Ac- cessed on: March 04th, 2020. Available at: .

JESUS, L. D. de et al. Greater autonomy for RPAs using solar panels and taking advantage of rising winds through the algorithm. In: 16th International Conference on Information Technology-New Generations (ITNG 2019). Springer Internation- al Publishing, 2019. p. 615{616. Available at: .

JIA, Y.; SHELHAMER, E. Ca e. 2019. Accessed on: January 02nd, 2020. Availa- ble at: .

. Ca e Model Zoo. 2020. Accessed on: January 14th, 2020. Available at: .

JIA, Y. et al. Ca e. In: Proceedings of the ACM International Conference on Mul- timedia- MM ’14. ACM Press, 2014. Available at: .

JOSEPH, L. Why should we learn ros?: Introduction to ros and its package man- agement. In: Mastering ROS for robotics programming : design, build, and simu- late complex robots using Robot Operating System and master its out-of-the-box functionalities. Birmingham, England: Packt Publishing, 2015. Chapter 1, p. 1{2. ISBN 978-1-78355-179-8.

KANTUE, P.; PEDRO, J. O. Real-time identi cation of faulty systems: Develop- ment of an aerial platform with emulated rotor faults. In: 2019 4th Conference on Control and Fault Tolerant Systems (SysTol). IEEE, 2019. Available at: .

KARPATHY, A. Layers used to build ConvNets. 2019. Available at: .

KERAS. Keras: The Python Deep Learning library. 2020. Accessed on: March 03rd, 2020. Available at: .

34

KOENIG, N.; HOWARD, A. Design and use paradigms for gazebo, an open- source multi-robot simulator. In: 2004 IEEE/RSJ International Conference on In- telligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566). IEEE, 2004. Available at: .

KOH, L. P.; WICH, S. A. Dawn of drone ecology: low-cost autonomous aerial vehicles for conservation. Mongabay.com, 2012. Available at: .

KOUBA, A. Services. 2019. Accessed on: December 19th, 2019. Available at: .

KRANTHI, S.; PRANATHI, K.; SRISAILA, A. Automatic number plate recogni- tion. In: International Journal of Advancements in Technology (IJoAT). Ambala, India: Ambala: Nupur Publishing House, 2011. (3, v. 2), p. 408{422.

KRAUSE, J. et al. 3d object representations for ne-grained categorization. In: 4th International IEEE Workshop on 3D Representation and Recognition (3dRR- 13). Sydney, Australia: [n.p.], 2013.

KUMAR, G.; BHATIA, P. K. Neural network based approach for recognition of text images. International Journal of Computer Applications, Foundation of Com- puter Science, v. 62, n. 14, p. 8{13, jan. 2013. Available at: .

LECUN, Y.; BENGIO, Y.; HINTON, G. Deep learning. Nature, v. 521, p. 436{44, 05 2015.

LEE, T.; LEOK, M.; MCCLAMROCH, N. H. Geometric tracking control of a quadrotor UAV on SE(3). In: 49th IEEE Conference on Decision and Control (CDC). IEEE, 2010. Available at:

LIN, T.-Y. et al. Microsoft coco: Common objects in context. In: European Conference on Computer Vision (ECCV). Zurich: [n.p.], 2014. Oral. Available at: .

LIU, G. et al. The calculation method of road travel time based on license plate recognition technology. In: Communications in Computer and Information Sci- ence. Springer Berlin Heidelberg, 2011. p. 385{389. Available at: .

35

LIU, W. SSD: Single Shot MultiBox Detector. 2018. Accessed on: January 14th, 2020. Available at: .

LIU, W. et al. Ssd: Single shot multibox detector. Lecture Notes in Computer Sci- ence, Springer International Publishing, p. 21{37, 2016. ISSN 1611-3349. Availa- ble at: .

LUUKKONEN, T. Modelling and control of quadcopter. Master Thesis (Inde- pendent research project in applied mathematics) | Department of Mathematics and Systems Analysis, Aalto University School of Science, Espoo, Finland, aug 2011.

MAO, W. et al. Indoor follow me drone. In: Proceedings of the 15th Annual Inter- national Conference on Mobile Systems, Applications, and Services - MobiSys ’17. ACM Press, 2017. Available at: .

MARTINEZ, A. Getting started with ros. In: Learning ROS for robotics pro- gramming: a practical, instructive, and comprehensive guide to introduce yourself to ROS, the top-notch, leading robotics framework. Birmingham, England: Packt Publishing, 2013. Chapter 1, p. 7{8. ISBN 978-1-78216-144-8.

MARTINS, W. M. et al. A computer vision based algorithm for obstacle avoid- ance. In: Advances in Intelligent Systems and Computing. Springer International Publishing, 2018. p. 569{575. Available at: .

MAVLINK. MAVLink Developer Guide. 2019. Accessed on: December 01st, 2019. Available at: .

MAVSDK. MAVSDK (develop). 2019. Accessed on: December 01st, 2019. Available at: .

MEIER, L. et al. PIXHAWK: A micro aerial vehicle design for autonomous ight using onboard computer vision. Autonomous Robots, Springer Science and Busi- ness Media LLC, v. 33, n. 1-2, p. 21{39, fev. 2012. Available at: .

MEYER, A. X-Plane. 2020. Accessed on: March 01st, 2020. Available at: .

MISHRA, N. et al. Shirorekha chopping integrated tesseract OCR engine for en- hanced hindi language recognition. International Journal of Computer Applica- 36

tions, Foundation of Computer Science, v. 39, n. 6, p. 19{23, feb 2012. Available at: .

MITHE, R.; INDALKAR, S.; DIVEKAR, N. Optical character recognition. In: In- ternational Journal of Recent Technology and Engineering (IJRTE). G18-19-20, Block-B, Tirupati Abhinav Homes, Damkheda, Bhopal (Madhya Pradesh)- 462037, India.: Blue Eyes Intelligence Engineering and Sciences Publication (BEIESP), 2013. (1, v. 2), p. 72{75.

MITSCH, S.; GHORBAL, K.; PLATZER, A. On provably safe obstacle avoid- ance for autonomous robotic ground vehicles. In: Robotics: Science and Systems IX. Robotics: Science and Systems Foundation, 2013. Available at: .

NGUYEN, K. D.; HA, C.; JANG, J. T. Development of a new hybrid drone and software-in-the-loop simulation using PX4 code. In: Intelligent Computing Theo- ries and Application. Springer International Publishing, 2018. p. 84{93. Available at: .

NOGUEIRA, L. Comparative Analysis Between Gazebo and V-REP Robotic Simulators. Master Thesis (Independent research project in applied mathematics) | School of Electrical and Computer Engineering, Campinas University, Campinas, S~ao Paulo, Brazil, dec 2014.

OLIVEIRA NA ESTRADA. Marginal Pinheiros alterac~oes no caminho para Castelo Branco. 2018. Accessed on: January 05th, 2020. Available at: .

OPALA, M. Top Machine Learning Frameworks Compared: Scikit-Learn, Dlib, MLib, Tensor ow, and More. 2018. Accessed on: March 04th, 2020. Available at: .

. Deep Learning Frameworks Comparison { Tensor ow, PyTorch, Keras, MXNet, The Microsoft Cognitive Toolkit, Ca e, Deeplearning4j, Chaine. 2019. Accessed on: March 03rd, 2020. Available at: .

OPELT, A. et al. Generic object recognition with boosting. IEEE Transactions on Pattern Analysis and Machine Intelligence, Institute of Electrical and Electronics Engineers (IEEE), v. 28, n. 3, p. 416{431, mar 2006. Available at: .

37

OPENALPR. OpenALPR Documentation. 2017. Accessed on: January 07th, 2020. Available at: .

O’SHEA, K.; NASH, R. An Introduction to Convolutional Neural Networks. 2015.

OZA, P.; PATEL, V. M. One-class convolutional neural network. IEEE Signal Processing Letters, Institute of Electrical and Electronics Engineers (IEEE), v. 26, n. 2, p. 277{281, feb 2019. Available at: .

OZONEDEV. Deep License Plate Recognition comes to Zoneminder. 2019. Ac- cessed on: January 10th, 2020. Available at: .

PAPAGEORGIOU, C.; POGGIO, T. A trainable system for object detection. In- ternational Journal of Computer Vision, Springer Science and Business Media LLC, v. 38, n. 1, p. 15{33, 2000. Available at: .

PAREKH, K. B. Graduation Monograph, A Computer Vision Application to Ac- curately Estimate Object Distance. 1600 Grand Avenue ,Saint Paul, MN 55105- 1899, USA: [n.p.], 2010.

PATEL, C.; PATEL, A.; PATEL, D. Optical character recognition by open source OCR tool tesseract: A case study. International Journal of Computer Applications, Foundation of Computer Science, v. 55, n. 10, p. 50{56, out. 2012. Available at: .

PINTO, L. G. M. et al. A ssd { ocr approach for real-time active car tracking on quadrotors. In: . 16th International Conference on Information Technology-New Generations (ITNG 2019). Springer International Publishing, 2019. p. 471{476. Available at: .

PIXHAWK. What is Pixhawk. 2019. Accessed on: December 16th, 2019. Availa- ble at: .

PX4. What Is PX4? 2019. Accessed on: December 03rd, 2019. Available at: .

. Simple multirotor simulator with MAVLink protocol support. 2020. Ac- cessed on: March 01st, 2020. Available at: .

38

PX4 DEV. MAVROS. 2019. Accessed on: December 19th, 2019. Available at: .

. Using DroneKit to communicate with PX4. 2019. Accessed on: Decem- ber 21st, 2019. Available at: .

PX4DOCS. Airframes Reference. 2019. Accessed on: December 03rd, 2019. Available at: .

. Multicopter PID Tuning Guide. 2019. Accessed on: December 03rd, 2019. Available at:

PYTORCH. Tensors and Dynamic neural networks in Python with strong GPU acceleration. 2020. Accessed on: March 03rd, 2020. Available at: .

QADRI, M. T.; ASIF, M. Automatic number plate recognition system for vehicle identi cation using optical character recognition. In: 2009 International Conference on Education Technology and Computer. IEEE, 2009. Available at: .

QGROUNDCONTROL. QGROUNDCONTROL: Intuitive and Powerful Ground Control Station for the MAVLink protocol. 2019. Accessed on: December 19th, 2019. Available at: .

. QGroundControl User Guide. 2019. Accessed on: December 06th, 2019. Available at: .

QUIGLEY, M. et al. Ros: an open-source robot operating system. In: . [n.a.: n.p.], 2009. v. 3.

RAMIREZ-ATENCIA, C.; CAMACHO, D. Extending QGroundControl for au- tomated mission planning of UAVs. Sensors, MDPI AG, v. 18, n. 7, p. 2339, jul. 2018. Available at: .

RAMPASEK, L.; GOLDENBERG, A. TensorFlow: Biology’s gateway to deep learning? Cell Systems, Elsevier BV, v. 2, n. 1, p. 12{14, jan. 2016. Available at: .

REAL, F. et al. Ual: an abstraction layer for unmanned aerial vehicles. In: Interna- tional Symposium on Aerial Robotics. [n.a.: n.p.], 2018. 39

REDMON, J. et al. You only look once: Uni ed, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2016. Available at: .

REDMON, J.; FARHADI, A. Yolo9000: Better, faster, stronger. 2017 IEEE Con- ference on Computer Vision and Pattern Recognition (CVPR), IEEE, Jul 2017. Available at: .

REMANAN, S. Beginner’s Guide to Object Detection Algorithms. 2019. Ac- cessed on: January 02nd, 2020. Available at: .

REN, S. et al. Faster r-cnn: Towards real-time object detection with region pro- posal networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, Institute of Electrical and Electronics Engineers (IEEE), v. 39, n. 6, p. 1137{1149, Jun 2017. ISSN 2160-9292. Available at: .

REZATOFIGHI, H. et al. Generalized intersection over union: A metric and a loss for bounding box regression. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, Jun 2019. Available at: .

ROGOWSKI, M. V. da S. LiPRIF: Aplicativo para identi cac~ao de permiss~ao de acesso de ve culos e condutores ao estacionamento do IFRS. Monography (Graduation Final Project) | Instituto Federal de Educac~ao, Ci^encia e Tecnolo- gia do Rio Grande do Sul (IFRS), Campus Porto Alegre, Av. Cel. Vicente, 281, Porto Alegre - RS - Brasil, dec 2018.

ROS WIKI. MAVROS. 2019. Accessed on: December 19th, 2019. Available at: .

SABATINO, F. Quadrotor control: modeling, nonlinear control design, and simu- lation. Master Thesis (MSc) | School of Electrical Engineering and Computer Sci- ence - KTH Royal Institute of Technology, Stockholm, Sweden, jun 2015.

SAGAR, A. 5 Techniques to Prevent Over tting in Neural Networks. 2019. Ac- cessed on: March 03rd, 2020. Available at: .

SAMBASIVARAO, K. Non-maximum Suppression (NMS). 2019. Accessed on: December 40

10th, 2019. Available at: .

SARFRAZ, M.; AHMED, M.; GHAZI, S. Saudi arabian license plate recognition system. In: 2003 International Conference on Geometric Modeling and Graphics, 2003. Proceedings. IEEE Comput. Soc, 2003. Available at: .

SAWANT, A. S.; CHOUGULE, D. Script independent text pre-processing and segmentation for OCR. In: 2015 International Conference on Electrical, Electron- ics, Signals, Communication and Optimization (EESCO). IEEE, 2015. Available at: .

SCHWARTZ, B. et al. Towards a graphical language for quadrotor missions. 2014.

SCIKIT-LEARN. scikit-learn: machine learning in Python. 2020. Accessed on: March 03rd, 2020. Available at: .

SHAFAIT, F.; KEYSERS, D.; BREUEL, T. M. E cient implementation of local adaptive thresholding techniques using integral images. In: YANIKOGLU, B. A.; BERKNER, K. (Ed.). Document Recognition and Retrieval XV. SPIE, 2008. Available at: .

SHAH, S. et al. Airsim: High- delity visual and physical simulation for autono- mous vehicles. In: Field and Service Robotics. [n.p.], 2017. Available at: .

SHOBHA, G.; RANGASWAMY, S. Machine learning. In: Handbook of Statis- tics. Elsevier, 2018. p. 197{228. Available at: .

SHUCK, T. J. Development of Autonomous Optimal Cooperative Control in Re- lay Rover Con gured Small Unmanned Aerial Systems. Master Thesis (MSc) | Graduate School of Engineering and Management, Air Force Institute of Technol- ogy, Air University, Wright-Patterson Air Force Base (WPAFB), Ohio, US, mar 2013.

SILVA, S. M.; JUNG, C. R. License plate detection and recognition in uncon- strained scenarios. In: Computer Vision { ECCV 2018. Springer International Publishing, 2018. p. 593{609. Available at: . 41

SIMONYAN, K.; ZISSERMAN, A. Very deep convolutional networks for large- scale image recognition. In: BENGIO, Y.; LECUN, Y. (Ed.). 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings. [n.p.], 2015. Available at: .

SINGH, R. et al. Optical character recognition (ocr) for printed devnagari script using arti cial neural network. In: International Journal of Computer Science & Communication (IJCSC). [n.a.: n.p.], 2010. (1, v. 1), p. 91{95.

SMITH, R. An overview of the tesseract OCR engine. In: Ninth International Conference on Document Analysis and Recognition (ICDAR 2007) Vol 2. IEEE, 2007. Available at: .

SMITH, R.; ANTONOVA, D.; LEE, D.-S. Adapting the tesseract open source OCR engine for multilingual OCR. In: Proceedings of the International Workshop on Multilingual OCR - MOCR ’09. ACM Press, 2009. Available at: .

SOBOTTKA, K. et al. Text extraction from colored book and journal covers. In: KISE DANIEL LOPRESTI, S. M. K. (Ed.). International Journal on Document Analysis and Recognition (IJDAR). Tiergartenstrasse 17 69121 - Heidelberg, Germany: Springer-Verlag GmbH Germany, part of Springer Nature, 2000. (4, v. 2), p. 163{176.

SONGER, S. A. AERIAL NETWORKING FOR THE IMPLEMENTATION OF COOPERATIVE CONTROL ON SMALL UNMANNED AERIAL SYSTEMS. Master Thesis (MSc) | Graduate School of Engineering and Management, Air Force Institute of Technology, Air University, Wright-Patterson Air Force Base (WPAFB), Ohio, US, mar 2013.

SOVIANY, P.; IONESCU, R. T. Frustratingly easy trade-o optimization between single-stage and two-stage deep object detectors. In: Lecture Notes in Computer Science. Springer International Publishing, 2019. p. 366{378. Available at: .

STRIMEL, G.; BARTHOLOMEW, S.; KIM, E. Engaging children in engineering design through the world of quadcopters. Children’s Technology and Engineering Journal, v. 21, p. 7{11, 05 2017.

42

STUTZ, D. Understanding convolutional neural networks. In: VISUAL COMPUTING INSTITUTE, RWTH AACHEN UNIVERSITY. Current Topics in Computer Vision and Machine Learning. [N.a.], 2015. (Seminar Report).

SZEGEDY, C. et al. Going deeper with convolutions. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, Jun 2015. Available at: .

. Scalable, High-Quality Object Detection. 2014.

TALABIS, M. R. M. et al. Analytics de ned. In: Information Security Analytics. Elsevier, 2015. p. 1{12. Available at: .

TALIN, T. LabelImg. 2015. Accessed on: January 13th, 2020. Available at: .

TAVARES, D. M.; CAURIN, G. A. P.; GONZAGA, A. Tesseract ocr: a case study for license plate recognition in brazil. In: . [n.a.: n.p.], 2010.

TENSORFLOW. Tensor ow detection model zoo. 2019. Accessed on: January 13th, 2020. Available at: .

. Why TensorFlow. 2019. Accessed on: December 22nd, 2019. Available at: .

. Get started with TensorBoard. 2020. Accessed on: January 13th, 2020. Available at: .

TOKUI, S. et al. Chainer: a next-generation open source framework for deep learning. In: Proceedings of Workshop on Machine Learning Systems (Learn- ingSys) in The Twenty-ninth Annual Conference on Neural Information Pro- cessing Systems (NIPS). [n.p.], 2015. Accessed on: December 22nd, 2019. Avail- able at: .

TOMPKIN, J. Deep Learning with TensorFlow: Introduction to Computer Vision. 2017.Accessed on: December 10th, 2019. Available at: .

VARGAS, A. M. P. C. A. C. G.; VASCONCELOS, C. N. Um estudo sobre redes neurais convolucionais e sua aplicac~ao em detecc~ao de pedestres. In: CAPPABIANCO, F. A. M. et al. (Ed.). Electronic Proceedings of the 29th Con- 43

ference on Graphics, Patterns and Images (SIBGRAPI’16). S~ao Jose dos Cam- pos, SP, Brazil: [n.p.], 2016. Accessed on: November 23rd, 2019. Available at: .

VENTHAN, T. Integrating sonar and IR sensor plugin to robot model in Gazebo with ROS. 2018. Accessed on: November 28th, 2019. Available at: .

WEERASINGHE, R. et al. NLP applications of Sinhala: TTS & OCR. In: Pro- ceedings of the Third International Joint Conference on Natural Language Pro- cessing: Volume-II. [n.p.], 2008. Accessed on: January 03rd, 2020. Available at: .

WOO, A.; BRoKER, H.-B. gnuplot Quick Reference. 2004. Accessed on: January 14th, 2020. Available at: .

YAN, Q.-Z.; WILLIAMS, J. M.; LI, J. Chassis control system development using simulation: Software in the loop, rapid prototyping, and hardware in the loop. In: SAE Technical Paper Series. SAE International, 2002. Available at: .

ZHANG, H. et al. An improved scene text extraction method using conditional random eld and optical character recognition. In: 2011 International Conference on Document Analysis and Recognition. IEEE, 2011. Available at: .