Master Thesis

Total Page:16

File Type:pdf, Size:1020Kb

Master Thesis Master thesis To obtain a Master of Science Degree in Informatics and Communication Systems from the Merseburg University of Applied Sciences Subject: Tunisian truck license plate recognition using an Android Application based on Machine Learning as a detection tool Author: Supervisor: Achraf Boussaada Prof.Dr.-Ing. Rüdiger Klein Matr.-Nr.: 23542 Prof.Dr. Uwe Schröter Table of contents Chapter 1: Introduction ................................................................................................................................. 1 1.1 General Introduction: ................................................................................................................................... 1 1.2 Problem formulation: ................................................................................................................................... 1 1.3 Objective of Study: ........................................................................................................................................ 4 Chapter 2: Analysis ........................................................................................................................................ 4 2.1 Methodological approaches: ........................................................................................................................ 4 2.1.1 Actual approach: ................................................................................................................................... 4 2.1.2 Image Processing with OCR: ................................................................................................................. 6 2.1.3 Chosen approach: ................................................................................................................................. 7 Chapter 3: Artificial Intelligence & Machine Learning ..................................................................................... 8 3.1 Introduction: ................................................................................................................................................. 8 3.2 Types: ........................................................................................................................................................... 8 3.2.1 Supervised learning: ............................................................................................................................. 8 3.2.2 Unsupervised Learning: ........................................................................................................................ 9 3.2.3 Semi-supervised Learning: .................................................................................................................. 10 3.2.4 Reinforcement Learning: .................................................................................................................... 10 3.3 Techniques: ................................................................................................................................................. 12 3.3.1 SVM: ................................................................................................................................................... 12 3.3.2 Random Forest: .................................................................................................................................. 15 3.3.2.1 Decision Tree: ............................................................................................................................. 15 3.3.2.2 Random Forest: ........................................................................................................................... 17 3.3.2.3 Feature Importance: ................................................................................................................... 18 3.3.3 Deep Learning: .................................................................................................................................... 19 3.3.3.1 Deep Learning and Neural Networks: ......................................................................................... 20 3.3.3.2 Perceptron: ................................................................................................................................. 25 3.3.3.3 Feed Forward: ............................................................................................................................. 25 3.3.3.4 Recurrent Neural Network: ........................................................................................................ 25 3.3.3.5 Deep Convolutional Network: .................................................................................................... 26 3.3.4 Convolutional Neural Network: .......................................................................................................... 33 3.3.4.1 The Convolution operation: ........................................................................................................ 33 3.3.4.2 Motivation: ................................................................................................................................. 35 3.3.4.3 Pooling: ....................................................................................................................................... 40 3.3.4.4 Normalization: ............................................................................................................................ 41 3.3.4.5 Random or unsupervised Features: ............................................................................................ 42 3.3.4.6 Regularization: ............................................................................................................................ 42 3.3.4.7 Probability conversion: ............................................................................................................... 43 3.3.5 Transfer Learning: ............................................................................................................................... 46 3.3.5.1 Applications of Transfer Learning: .............................................................................................. 49 3.4 Recapitulation: ........................................................................................................................................... 51 Chapter 4: Implementation .......................................................................................................................... 52 4.1 Software & Tools: ....................................................................................................................................... 52 4.2 Object Detection API:.................................................................................................................................. 54 4.3 Model Training: .......................................................................................................................................... 54 4.4 Mobile implementation: ............................................................................................................................. 58 4.5 Optimization for mobile usage: .................................................................................................................. 62 4.5.1 Minimum Device Requirement: ......................................................................................................... 62 4.5.2 Removing training-only nodes: ........................................................................................................... 62 4.5.3 Recompiling TensorFlow inference library: ........................................................................................ 64 4.5.4 Retrain with mobile data: ................................................................................................................... 65 4.5.5 Reduce model loading time or memory footprint & improve RAM usage: ........................................ 66 4.5.6 Reduce model size: ............................................................................................................................. 67 4.5.7 Exploring Quantized Calculations: ...................................................................................................... 67 4.6 Recapitulation: ........................................................................................................................................... 68 Chapter 5: Summary .................................................................................................................................... 69 5.1 Discussion: .................................................................................................................................................. 69 5.1.1 Results: ............................................................................................................................................... 69 5.1.2 Future Improvements: ........................................................................................................................ 72 5.1.2.1 TensorFlow Lite: .......................................................................................................................... 72 5.1.2.2 Pre-trained models: .................................................................................................................... 74 5.1.2.3 Training and serving with Cloud TPUs: ....................................................................................... 74 5.2 Conclusion: ................................................................................................................................................
Recommended publications
  • Open Source Used in Influx1.8 Influx 1.9
    Open Source Used In Influx1.8 Influx 1.9 Cisco Systems, Inc. www.cisco.com Cisco has more than 200 offices worldwide. Addresses, phone numbers, and fax numbers are listed on the Cisco website at www.cisco.com/go/offices. Text Part Number: 78EE117C99-1178791953 Open Source Used In Influx1.8 Influx 1.9 1 This document contains licenses and notices for open source software used in this product. With respect to the free/open source software listed in this document, if you have any questions or wish to receive a copy of any source code to which you may be entitled under the applicable free/open source license(s) (such as the GNU Lesser/General Public License), please contact us at [email protected]. In your requests please include the following reference number 78EE117C99-1178791953 Contents 1.1 golang-protobuf-extensions v1.0.1 1.1.1 Available under license 1.2 prometheus-client v0.2.0 1.2.1 Available under license 1.3 gopkg.in-asn1-ber v1.0.0-20170511165959-379148ca0225 1.3.1 Available under license 1.4 influxdata-raft-boltdb v0.0.0-20210323121340-465fcd3eb4d8 1.4.1 Available under license 1.5 fwd v1.1.1 1.5.1 Available under license 1.6 jaeger-client-go v2.23.0+incompatible 1.6.1 Available under license 1.7 golang-genproto v0.0.0-20210122163508-8081c04a3579 1.7.1 Available under license 1.8 influxdata-roaring v0.4.13-0.20180809181101-fc520f41fab6 1.8.1 Available under license 1.9 influxdata-flux v0.113.0 1.9.1 Available under license 1.10 apache-arrow-go-arrow v0.0.0-20200923215132-ac86123a3f01 1.10.1 Available under
    [Show full text]
  • Easybuild Documentation Release 20210907.0
    EasyBuild Documentation Release 20210907.0 Ghent University Tue, 07 Sep 2021 08:55:41 Contents 1 What is EasyBuild? 3 2 Concepts and terminology 5 2.1 EasyBuild framework..........................................5 2.2 Easyblocks................................................6 2.3 Toolchains................................................7 2.3.1 system toolchain.......................................7 2.3.2 dummy toolchain (DEPRECATED) ..............................7 2.3.3 Common toolchains.......................................7 2.4 Easyconfig files..............................................7 2.5 Extensions................................................8 3 Typical workflow example: building and installing WRF9 3.1 Searching for available easyconfigs files.................................9 3.2 Getting an overview of planned installations.............................. 10 3.3 Installing a software stack........................................ 11 4 Getting started 13 4.1 Installing EasyBuild........................................... 13 4.1.1 Requirements.......................................... 14 4.1.2 Using pip to Install EasyBuild................................. 14 4.1.3 Installing EasyBuild with EasyBuild.............................. 17 4.1.4 Dependencies.......................................... 19 4.1.5 Sources............................................. 21 4.1.6 In case of installation issues. .................................. 22 4.2 Configuring EasyBuild.......................................... 22 4.2.1 Supported configuration
    [Show full text]
  • Bringing Probabilistic Programming to Scientific Simulators at Scale
    Etalumis: Bringing Probabilistic Programming to Scientific Simulators at Scale Atılım Güneş Baydin, Lei Shao, Wahid Bhimji, Lukas Heinrich, Lawrence Meadows, Jialin Liu, Andreas Munk, Saeid Naderiparizi, Bradley Gram-Hansen, Gilles Louppe, Mingfei Ma, Xiaohui Zhao, Philip Torr, Victor Lee, Kyle Cranmer, Prabhat, and Frank Wood SC19, Denver, CO, United States 19 November 2019 Simulation and HPC Computational models and simulation are key to scientific advance at all scales Particle physics Nuclear physics Material design Drug discovery Weather Climate science Cosmology 2 Introducing a new way to use existing simulators Probabilistic programming Simulation Supercomputing (machine learning) 3 Simulators Parameters Outputs (data) Simulator 4 Simulators Parameters Outputs (data) Simulator Prediction: ● Simulate forward evolution of the system ● Generate samples of output 5 Simulators Parameters Outputs (data) Simulator Prediction: ● Simulate forward evolution of the system ● Generate samples of output 6 Simulators Parameters Outputs (data) Simulator Prediction: ● Simulate forward evolution of the system ● Generate samples of output WE NEED THE INVERSE! 7 Simulators Parameters Outputs (data) Simulator Prediction: ● Simulate forward evolution of the system ● Generate samples of output Inference: ● Find parameters that can produce (explain) observed data ● Inverse problem ● Often a manual process 8 Simulators Parameters Outputs (data) Inferred Simulator Observed data parameters Gene network Gene expression 9 Simulators Parameters Outputs (data)
    [Show full text]
  • Advances in Electron Microscopy with Deep Learning
    Advances in Electron Microscopy with Deep Learning by Jeffrey Mark Ede Thesis To be submitted to the University of Warwick for the degree of Doctor of Philosophy in Physics Department of Physics January 2021 Contents Contents i List of Abbreviations iii List of Figures viii List of Tables xvii Acknowledgments xix Declarations xx Research Training xxv Abstract xxvi Preface xxvii I Initial Motivation......................................... xxvii II Thesis Structure.......................................... xxvii III Connections............................................ xxix Chapter 1 Review: Deep Learning in Electron Microscopy1 1.1 Scientific Paper..........................................1 1.2 Reflection............................................. 100 Chapter 2 Warwick Electron Microscopy Datasets 101 2.1 Scientific Paper.......................................... 101 2.2 Amendments and Corrections................................... 133 2.3 Reflection............................................. 133 Chapter 3 Adaptive Learning Rate Clipping Stabilizes Learning 136 3.1 Scientific Paper.......................................... 136 3.2 Amendments and Corrections................................... 147 3.3 Reflection............................................. 147 Chapter 4 Partial Scanning Transmission Electron Microscopy with Deep Learning 149 4.1 Scientific Paper.......................................... 149 4.2 Amendments and Corrections................................... 176 4.3 Reflection............................................. 176
    [Show full text]
  • An Accuracy Examination of OCR Tools
    International Journal of Innovative Technology and Exploring Engineering (IJITEE) ISSN: 2278-3075, Volume-8, Issue-9S4, July 2019 An Accuracy Examination of OCR Tools Jayesh Majumdar, Richa Gupta texts, pen computing, developing technologies for assisting Abstract—In this research paper, the authors have aimed to do a the visually impaired, making electronic images searchable comparative study of optical character recognition using of hard copies, defeating or evaluating the robustness of different open source OCR tools. Optical character recognition CAPTCHA. (OCR) method has been used in extracting the text from images. OCR has various applications which include extracting text from any document or image or involves just for reading and processing the text available in digital form. The accuracy of OCR can be dependent on text segmentation and pre-processing algorithms. Sometimes it is difficult to retrieve text from the image because of different size, style, orientation, a complex background of image etc. From vehicle number plate the authors tried to extract vehicle number by using various OCR tools like Tesseract, GOCR, Ocrad and Tensor flow. The authors in this research paper have tried to diagnose the best possible method for optical character recognition and have provided with a comparative analysis of their accuracy. Keywords— OCR tools; Orcad; GOCR; Tensorflow; Tesseract; I. INTRODUCTION Optical character recognition is a method with which text in images of handwritten documents, scripts, passport documents, invoices, vehicle number plate, bank statements, Fig.1: Functioning of OCR [2] computerized receipts, business cards, mail, printouts of static-data, any appropriate documentation or any II. OCR PROCDURE AND PROCESSING computerized receipts, business cards, mail, printouts of To improve the probability of successful processing of an static-data, any appropriate documentation or any picture image, the input image is often ‘pre-processed’; it may be with text in it gets processed and the text in the picture is de-skewed or despeckled.
    [Show full text]
  • Curriculum Vitae
    Diogo Castro Curriculum Vitae Summary I’m a Software Engineer based in Belfast, United Kingdom. My professional journey began as a C# developer, but Functional Programming (FP) soon piqued my interest and led me to learn Scala, Haskell, PureScript and even a bit of Idris. I love building robust, reliable and maintainable applications. I also like teaching and I’m a big believer in "paying it forward". I’ve learned so much from many inspiring people, so I make it a point to share what I’ve learned with others. To that end, I’m currently in charge of training new team members in Scala and FP, do occasional presentations at work and aim to do more public speaking. I blog about FP and Haskell at https://diogocastro.com/blog. Experience Nov 2017-Present Principal Software Engineer, SpotX, Belfast, UK. Senior Software Engineer, SpotX, Belfast, UK. Developed RESTful web services in Scala, using the cats/cats-effect framework and Akka HTTP. Used Apache Kafka for publishing of events, and Prometheus/Grafana for monitor- ing. Worked on a service that aimed to augment Apache Druid, a timeseries database, with features such as access control, a safer and simpler query DSL, and automatic conversion of monetary metrics to multiple currencies. Authored a Scala library for calculating the delta of any two values of a given type using Shapeless, a library for generic programming. Taught a weekly internal Scala/FP course, with the goal of preparing our engineers to be productive in Scala whilst building an intuition of how to program with functions and equational reasoning.
    [Show full text]
  • KNIME Deep Learning Integration Installation Guide
    KNIME Deep Learning Integration Installation Guide KNIME AG, Zurich, Switzerland Version 3.7 (last updated on 2019-02-06) Table of Contents Introduction. 1 KNIME Deep Learning Integrations . 1 KNIME Keras Integration Installation. 2 Python Installation. 2 Installing the KNIME Keras Integration. 3 Extensions . 3 GPU Support. 4 KNIME TensorFlow Integration Installation . 4 Installation . 4 Advanced . 4 GPU Support. 4 KNIME Deeplearning4j Installation. 5 Installation . 5 GPU Support. 5 Known Issues . 5 KNIME Deep Learning Integration Installation Guide Introduction This document describes how to install the KNIME Deep Learning Integrations. These integrations bring deep learning capabilities to KNIME Analytics Platform, which allow you to read, create, edit, train, and execute deep neural networks within KNIME Analytics Platform. KNIME Deep Learning Integrations Three different deep learning libraries have been integrated: KNIME Keras Integration The KNIME Keras Integration utilizes the Keras deep learning framework to enable users to read, write, train, and execute Keras deep learning networks within KNIME. Furthermore, you can also build custom deep learning networks directly in KNIME via the Keras layer nodes. KNIME Tensor Flow Integration The KNIME TensorFlow Integration provides access to the powerful machine learning library TensorFlow* within KNIME. It enables you to read, write, train, and execute TensorFlow networks directly in KNIME. You can also convert your Keras networks to TensorFlow networks with this extension for even greater flexibility. * TensorFlow, the TensorFlow logo and any related marks are trademarks of Google Inc. KNIME Deeplearning4j Integration The KNIME Deeplearning4j Integration integrates the Deeplearning4j library into KNIME, which provides deep learning capabilities in Java. Within KNIME this means you can read, write, train, execute, and build Deeplearning4j networks.
    [Show full text]
  • Enforcing Abstract Immutability
    Enforcing Abstract Immutability by Jonathan Eyolfson A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Doctor of Philosophy in Electrical and Computer Engineering Waterloo, Ontario, Canada, 2018 © Jonathan Eyolfson 2018 Examining Committee Membership The following served on the Examining Committee for this thesis. The decision of the Examining Committee is by majority vote. External Examiner Ana Milanova Associate Professor Rensselaer Polytechnic Institute Supervisor Patrick Lam Associate Professor University of Waterloo Internal Member Lin Tan Associate Professor University of Waterloo Internal Member Werner Dietl Assistant Professor University of Waterloo Internal-external Member Gregor Richards Assistant Professor University of Waterloo ii I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. iii Abstract Researchers have recently proposed a number of systems for expressing, verifying, and inferring immutability declarations. These systems are often rigid, and do not support “abstract immutability”. An abstractly immutable object is an object o which is immutable from the point of view of any external methods. The C++ programming language is not rigid—it allows developers to express intent by adding immutability declarations to methods. Abstract immutability allows for performance improvements such as caching, even in the presence of writes to object fields. This dissertation presents a system to enforce abstract immutability. First, we explore abstract immutability in real-world systems. We found that developers often incorrectly use abstract immutability, perhaps because no programming language helps developers correctly implement abstract immutability.
    [Show full text]
  • CSI: Inferring Mobile ABR Video Adaptation Behavior Under HTTPS and QUIC
    CSI: Inferring Mobile ABR Video Adaptation Behavior under HTTPS and QUIC Shichang Xu Subhabrata Sen Z. Morley Mao University of Michigan AT&T Labs – Research University of Michigan Abstract Server Manifest Network Client Mobile video streaming services have widely adopted Adap- Chunks HTTP tive Bitrate (ABR) streaming to dynamically adapt the stream- Track ing quality to variable network conditions. A wide range of 720p 1 Buffer third-party entities such as network providers and testing 480p IP packets services need to understand such adaptation behavior for 360p 1 2 3 Index purposes such as QoE monitoring and network management. CSI The traditional approach involved conducting test runs and analyzing the HTTP-level information from the associated network traffic to understand the adaptation behavior under Figure 1. ABR streaming overview different network conditions. However, end-to-end traffic encryption protocols such as HTTPS and QUIC are being increasingly used by streaming services, hindering such tra- Rate (ABR) streaming (predominantly HLS [75] and DASH [31]) ditional traffic analysis approaches. has been widely adopted in industry for delivering satisfac- To address this, we develop CSI (Chunk Sequence Infer- tory Quality of Experience (QoE) over dynamic cellular net- encer), a general system that enables third-parties to conduct work conditions. The server encodes each video into multiple active measurements and infer mobile ABR video adapta- versions with different picture quality levels and encoding tion behavior based on packet size and timing information bitrates (with higher bitrates for higher-quality encodings) still available in the encrypted traffic. We perform exten- called tracks, and splits each track into shorter chunks, each sive evaluations and demonstrate that CSI achieves high representing a few seconds worth of playback content (Fig- inference accuracy for video encodings of popular streaming ure 1).
    [Show full text]
  • Comparison and Benchmarking of AI Models and Frameworks on Mobile Devices
    Comparison and Benchmarking of AI Models and Frameworks on Mobile Devices 1st Chunjie Luo 2nd Xiwen He 3nd Jianfeng Zhan Institute of Computing Technology Institute of Computing Technology Institute of Computing Technology Chinese Academy of Sciences Chinese Academy of Sciences Chinese Academy of Sciences BenchCouncil BenchCouncil BenchCouncil Beijing, China Beijing, China Beijing, China [email protected] [email protected] [email protected] 4nd Lei Wang 5nd Wanling Gao 6nd Jiahui Dai Institute of Computing Technology Institute of Computing Technology Beijing Academy of Frontier Chinese Academy of Sciences Chinese Academy of Sciences Sciences and Technology BenchCouncil BenchCouncil Beijing, China Beijing, China Beijing, China [email protected] wanglei [email protected] [email protected] Abstract—Due to increasing amounts of data and compute inference on the edge devices can 1) reduce the latency, 2) resources, deep learning achieves many successes in various protect the privacy, 3) consume the power [2]. domains. The application of deep learning on the mobile and em- To make the inference on the edge devices more efficient, bedded devices is taken more and more attentions, benchmarking and ranking the AI abilities of mobile and embedded devices the neural networks are designed more light-weight by using becomes an urgent problem to be solved. Considering the model simpler architecture, or by quantizing, pruning and compress- diversity and framework diversity, we propose a benchmark suite, ing the networks. Different networks present different trade- AIoTBench, which focuses on the evaluation of the inference offs between accuracy and computational complexity. These abilities of mobile and embedded devices.
    [Show full text]
  • Expense Tracking Mobile Application with Receipt Scanning Functionality Bachelor’S Thesis
    TALLINN UNIVERSITY OF TECHNOLOGY Faculty of Information Technology Department of Computer Science Chair of Network Software Expense tracking mobile application with receipt scanning functionality Bachelor’s thesis Student: Roman Kaskman Student code: 113089 IAPB Advisor: Roger Kerse Tallinn 2015 Author’s declaration I declare that this thesis is the result of my own research except as cited in the references. The thesis has not been accepted for any degree and is not concurrently submitted in candidature of any other degree. 25.05.2015 Roman Kaskman (date) (signature) Abstract The purpose of this thesis is to create a mobile application for expense tracking, with the main focus on functionality allowing to take pictures of receipts issued by Estonian enterprises, extract basic expense information from the captured receipt images and store extracted expenses information in authenticated user’s expense list. The main problems covered in this work are finding the best architectural and design solutions for the application from the perspective of performance, usability, security and further development as well as researching and implementing techniques to handle expense recognition from receipts in an efficient way. As a result of the thesis, a working implementation of expense tracking mobile application for Android appears. After functionality of expenses information extraction from receipt images passes the testing phase, conclusion regarding its reliability is made. Moreover, proposals for further improvements of the application’s functionality are also presented. The thesis is in English and contains 53 pages of text, 6 chapters and 14 figures. Annotatsioon Käesoleva bakalaureusetöö eesmärk on luua mobiilirakendus kasutaja kulude üle arvestuse pidamiseks ja dokumenteerimiseks.
    [Show full text]
  • See License Notices
    Open Source Components License Notices Last updated December 22, 2020 @nebular/eva-icons @ngx-translate/core @ngx-translate/http-loader libunwind dropbear 2017.75 popt MIT License Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice (including the next paragraph) shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. --------------------------------------------------------------------------------------------------------------------- @angular/animations @angular/common @angular/core @angular/forms @angular/platform-browser @angular/router (v.1_11-19-20) The MIT License Copyright (c) 2010-2020 Google LLC. http://angular.io/license Permission
    [Show full text]