Character Recognition in Natural Images Utilising Tensorflow

Total Page:16

File Type:pdf, Size:1020Kb

Character Recognition in Natural Images Utilising Tensorflow DEGREE PROJECT IN TECHNOLOGY, FIRST CYCLE, 15 CREDITS STOCKHOLM, SWEDEN 2017 Character Recognition in Natural Images Utilising TensorFlow ALEXANDER VIKLUND EMMA NIMSTAD KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF COMPUTER SCIENCE AND COMMUNICATION Character Recognition in Natural Images Utilising TensorFlow ALEXANDER VIKLUND EMMA NIMSTAD Degree project in Computer Science, DD143X Date: June 12, 2017 Supervisor: Kevin Smith Examiner: Örjan Ekeberg Swedish title: Teckenigenkänning i naturliga bilder med TensorFlow School of Computer Science and Communication Abstract Convolutional Neural Networks (CNNs) are commonly used for character recogni- tion. They achieve the lowest error rates for popular datasets such as SVHN and MNIST. Usage of CNN is lacking in research about character classification in nat- ural images regarding the whole English alphabet. This thesis conducts an experi- ment where TensorFlow is used to construct a CNN that is trained and tested on the Chars74K dataset, with 15 images per class for training and 15 images per class for testing. This is done with the aim of achieving a higher accuracy than the non-CNN approach by de Campos et al. [1], that achieved 55:26%. The thesis explores data augmentation techniques for expanding the small training set and evaluates the result of applying rotation, stretching, translation and noise- adding. The result of this is that all of these methods apart from adding noise gives a positive effect on the accuracy of the network. Furthermore, the experiment shows that with a three layered convolutional neural network it is possible to create a character classifier that is as good as de Campos et al.’s. It is believed that even better results can be achieved if more experiments would be conducted on the parameters of the network and the augmentation. ii Sammanfattning Det är vanligt att använda konvolutionära artificiella neuronnät (CNN) för bildigen- känning, då de ger de minsta felmarginalerna på kända datamängder som SVHN och MNIST. Dock saknas det forskning om användning av CNN för klassificering av bok- stäver i naturliga bilder när det gäller hela det engelska alfabetet. Detta arbete beskri- ver ett experiment där TensorFlow används för att bygga ett CNN som tränas och testas med bilder från Chars74K. 15 bilder per klass används för träning och 15 per klass för testning. Målet med detta är att uppnå högre noggrannhet än 55:26%, vilket är vad de Campos et al. [1] uppnådde med en metod utan artificiella neuronnät. I rapporten utforskas olika tekniker för att artificiellt utvidga den lilla datamäng- den, och resultatet av att applicera rotation, utdragning, translation och bruspåslag utvärderas. Resultatet av det är att alla dessa metoder utom bruspåslag ger en positiv effekt på nätverkets noggrannhet. Vidare visar experimentet att med ett CNN med tre lager går det att skapa en bokstavsklassificerare som är lika bra som de Campos et al.s klassificering. Om fler experiment skulle genomföras på nätverkets och utvidgningens parametrar är det tro- ligt att ännu bättre resultat kan uppnås. iii Contents Contents iv 1 Introduction 1 1.1 Problem Statement . .2 1.2 Scope and constraints . .2 1.3 Thesis outline . .2 2 Background 3 2.1 Text recognition . .3 2.1.1 Pre-processing . .3 2.1.2 Character recognition . .3 2.2 Neural networks . .4 2.2.1 Feedforward networks . .4 2.2.2 Convolutional neural networks . .4 2.2.3 Training . .5 2.2.4 Training neural networks with small datasets . .7 3 Method 8 3.1 Overview . .8 3.2 Chars74K dataset . .9 3.3 Image processing . .9 3.3.1 Pre-processing . .9 3.3.2 Dataset augmentation . 10 3.4 Neural network . 11 3.4.1 TensorFlow . 11 3.4.2 Input . 11 3.4.3 Convolutional layers . 12 3.4.4 Fully connected layers . 12 3.4.5 Cost function . 12 3.5 Training . 13 4 Result 14 4.1 Processed images . 14 4.1.1 Pre-processing . 14 4.1.2 Data augmentation . 16 4.2 Neural network accuracy . 16 iv CONTENTS CONTENTS 5 Discussion 17 5.1 Discussion on the dataset . 17 5.2 Discussion on image processing methods . 17 5.3 Discussion on neural network approach . 18 6 Conclusion 19 Bibliography 20 v Chapter 1 Introduction The problem of recognising text has arisen from the massive data collection performed with modern technology. One aspect of this is how text recognition is used in tech- niques for document digitisation, automated data entry from scanned form sheets and text separation from graphics, that is, tasks that can be performed with classic OCR (Optical Character Recognition). These tasks can be performed by OCR with certain limitations; it is required that the images are fronto-parallel and skew compen- sated [2]. Another aspect of text recognition is its usage for analysing images. The amount of data generated from an abundance of camera-equipped devices, such as smart- phones and wearables, is increasing [3], and it is desired that computers become bet- ter at analysing and indexing those images. Possibly, there is text in the images that could give important contextual information, and thus, the indexing would be im- proved if the text could be found and read. There is also a need for accurate text recognition in natural scenes within the field of automation, such as automatic read- ing of signs in driverless cars. This thesis is treating the problem of identifying already individually extracted characters from natural images. There are many different approaches to this problem, where the methods range from very traditional ones to using modern and more ad- vanced techniques. In general, the methods use some assumptions about the letters, such as them being monochrome with a significant contrast to their background. Yet in a picture of a natural setting, the light conditions and texture on materials will af- fect the colours and contrasts. This means that even if the letters originally are mono- chrome, that is not what they will look like in a digital picture, and that is one of the challenges that this problem poses. Character recognition is well suited for an artificial neural network approach, an approach that has had great success in recent research. Neural networks have for example achieved the lowest error rate of the SVHN (Street View House Number) dataset of 1:64% [4] and the lowest error rate of the MNIST dataset of 0:21% [5]. A tool that has simplified working with neural networks is TensorFlow, which was released by Google in 2015. The TensorFlow API includes functions for build- ing and training neural networks, as well as for drawing diagrams and visualising the data. 1 CHAPTER 1. INTRODUCTION 1.1. PROBLEM STATEMENT 1.1 Problem Statement One of the biggest datasets with characters in natural images is the Chars74K dataset [6]. It was created by de Campos et al., who describe how they created it in the re- port Character Recognition in Natural Images [1]. The report also provides a couple of baseline experiments for character classifica- tion, neither of which uses neural networks. The most accurate method is a Multiple Kernel Learning (MKL) algorithm that evaluates six histograms of different features. Their classification scheme is trained with 15 images per character class and then tested on 15 images per character class. The scheme results in an accuracy of 55:26%. This thesis aims to evaluate how a neural network as a classification scheme com- pares to the MKL algorithm executed by de Campos et al. To make it fair, the com- parison is made with the same training and test data, even though it is a small amount of data for training a neural network. The problem statement for this is: Is an artificial neural network method for classification of characters in natural images more accurate than the method tested by de Campos et al. [1]? 1.2 Scope and constraints The problem of identifying text in an image is a complex problem with many sub- problems. A full-fledged text recognition software would consist of algorithms to detect and extract the characters, categorise the individual characters and then use lexical and grammatical analysis to correct characters that may have been incorrectly categorised. This thesis is only concerned with the most central task of categorising characters already detected and extracted from images. No further lexical or grammatical analy- sis will be performed to improve the results. 1.3 Thesis outline This thesis begins with presenting the background of character recognition and neural networks. Focus is put on how convolutions and training of neural networks works. In the method chapter algorithms for processing the images and neural network teach- ing are described in detail, the Chars74K dataset is also presented. Then the achieved results are presented and compared to de Campos et al.’s results, to conclude how our algorithm performs on the given dataset. Ending the thesis is an analysis of the tested algorithm, to indicte its strengths and how it could be improved. 2 Chapter 2 Background In the background chapter, the techniques used in this thesis are described. First, the area of text recognition is outlined. Secondly, an introduction to neural networks is given, both their structure and how they are trained. Training optimisations and data augmentation methods are also described. 2.1 Text recognition The text recognition problem can according to Jung et al. [7] be divided into four sub- problems: 1. detection, 2. localisation, 3. extraction and enhancement, 4. recognition. The first three steps may be grouped as the pre-processing steps, the steps that are performed before the main task of recognising the characters. Jung et al. [7] mention that the terms for these steps are used interchangeably in many reports.
Recommended publications
  • Master Thesis
    Master thesis To obtain a Master of Science Degree in Informatics and Communication Systems from the Merseburg University of Applied Sciences Subject: Tunisian truck license plate recognition using an Android Application based on Machine Learning as a detection tool Author: Supervisor: Achraf Boussaada Prof.Dr.-Ing. Rüdiger Klein Matr.-Nr.: 23542 Prof.Dr. Uwe Schröter Table of contents Chapter 1: Introduction ................................................................................................................................. 1 1.1 General Introduction: ................................................................................................................................... 1 1.2 Problem formulation: ................................................................................................................................... 1 1.3 Objective of Study: ........................................................................................................................................ 4 Chapter 2: Analysis ........................................................................................................................................ 4 2.1 Methodological approaches: ........................................................................................................................ 4 2.1.1 Actual approach: ................................................................................................................................... 4 2.1.2 Image Processing with OCR: ................................................................................................................
    [Show full text]
  • An Accuracy Examination of OCR Tools
    International Journal of Innovative Technology and Exploring Engineering (IJITEE) ISSN: 2278-3075, Volume-8, Issue-9S4, July 2019 An Accuracy Examination of OCR Tools Jayesh Majumdar, Richa Gupta texts, pen computing, developing technologies for assisting Abstract—In this research paper, the authors have aimed to do a the visually impaired, making electronic images searchable comparative study of optical character recognition using of hard copies, defeating or evaluating the robustness of different open source OCR tools. Optical character recognition CAPTCHA. (OCR) method has been used in extracting the text from images. OCR has various applications which include extracting text from any document or image or involves just for reading and processing the text available in digital form. The accuracy of OCR can be dependent on text segmentation and pre-processing algorithms. Sometimes it is difficult to retrieve text from the image because of different size, style, orientation, a complex background of image etc. From vehicle number plate the authors tried to extract vehicle number by using various OCR tools like Tesseract, GOCR, Ocrad and Tensor flow. The authors in this research paper have tried to diagnose the best possible method for optical character recognition and have provided with a comparative analysis of their accuracy. Keywords— OCR tools; Orcad; GOCR; Tensorflow; Tesseract; I. INTRODUCTION Optical character recognition is a method with which text in images of handwritten documents, scripts, passport documents, invoices, vehicle number plate, bank statements, Fig.1: Functioning of OCR [2] computerized receipts, business cards, mail, printouts of static-data, any appropriate documentation or any II. OCR PROCDURE AND PROCESSING computerized receipts, business cards, mail, printouts of To improve the probability of successful processing of an static-data, any appropriate documentation or any picture image, the input image is often ‘pre-processed’; it may be with text in it gets processed and the text in the picture is de-skewed or despeckled.
    [Show full text]
  • Enforcing Abstract Immutability
    Enforcing Abstract Immutability by Jonathan Eyolfson A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Doctor of Philosophy in Electrical and Computer Engineering Waterloo, Ontario, Canada, 2018 © Jonathan Eyolfson 2018 Examining Committee Membership The following served on the Examining Committee for this thesis. The decision of the Examining Committee is by majority vote. External Examiner Ana Milanova Associate Professor Rensselaer Polytechnic Institute Supervisor Patrick Lam Associate Professor University of Waterloo Internal Member Lin Tan Associate Professor University of Waterloo Internal Member Werner Dietl Assistant Professor University of Waterloo Internal-external Member Gregor Richards Assistant Professor University of Waterloo ii I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. iii Abstract Researchers have recently proposed a number of systems for expressing, verifying, and inferring immutability declarations. These systems are often rigid, and do not support “abstract immutability”. An abstractly immutable object is an object o which is immutable from the point of view of any external methods. The C++ programming language is not rigid—it allows developers to express intent by adding immutability declarations to methods. Abstract immutability allows for performance improvements such as caching, even in the presence of writes to object fields. This dissertation presents a system to enforce abstract immutability. First, we explore abstract immutability in real-world systems. We found that developers often incorrectly use abstract immutability, perhaps because no programming language helps developers correctly implement abstract immutability.
    [Show full text]
  • CSI: Inferring Mobile ABR Video Adaptation Behavior Under HTTPS and QUIC
    CSI: Inferring Mobile ABR Video Adaptation Behavior under HTTPS and QUIC Shichang Xu Subhabrata Sen Z. Morley Mao University of Michigan AT&T Labs – Research University of Michigan Abstract Server Manifest Network Client Mobile video streaming services have widely adopted Adap- Chunks HTTP tive Bitrate (ABR) streaming to dynamically adapt the stream- Track ing quality to variable network conditions. A wide range of 720p 1 Buffer third-party entities such as network providers and testing 480p IP packets services need to understand such adaptation behavior for 360p 1 2 3 Index purposes such as QoE monitoring and network management. CSI The traditional approach involved conducting test runs and analyzing the HTTP-level information from the associated network traffic to understand the adaptation behavior under Figure 1. ABR streaming overview different network conditions. However, end-to-end traffic encryption protocols such as HTTPS and QUIC are being increasingly used by streaming services, hindering such tra- Rate (ABR) streaming (predominantly HLS [75] and DASH [31]) ditional traffic analysis approaches. has been widely adopted in industry for delivering satisfac- To address this, we develop CSI (Chunk Sequence Infer- tory Quality of Experience (QoE) over dynamic cellular net- encer), a general system that enables third-parties to conduct work conditions. The server encodes each video into multiple active measurements and infer mobile ABR video adapta- versions with different picture quality levels and encoding tion behavior based on packet size and timing information bitrates (with higher bitrates for higher-quality encodings) still available in the encrypted traffic. We perform exten- called tracks, and splits each track into shorter chunks, each sive evaluations and demonstrate that CSI achieves high representing a few seconds worth of playback content (Fig- inference accuracy for video encodings of popular streaming ure 1).
    [Show full text]
  • Expense Tracking Mobile Application with Receipt Scanning Functionality Bachelor’S Thesis
    TALLINN UNIVERSITY OF TECHNOLOGY Faculty of Information Technology Department of Computer Science Chair of Network Software Expense tracking mobile application with receipt scanning functionality Bachelor’s thesis Student: Roman Kaskman Student code: 113089 IAPB Advisor: Roger Kerse Tallinn 2015 Author’s declaration I declare that this thesis is the result of my own research except as cited in the references. The thesis has not been accepted for any degree and is not concurrently submitted in candidature of any other degree. 25.05.2015 Roman Kaskman (date) (signature) Abstract The purpose of this thesis is to create a mobile application for expense tracking, with the main focus on functionality allowing to take pictures of receipts issued by Estonian enterprises, extract basic expense information from the captured receipt images and store extracted expenses information in authenticated user’s expense list. The main problems covered in this work are finding the best architectural and design solutions for the application from the perspective of performance, usability, security and further development as well as researching and implementing techniques to handle expense recognition from receipts in an efficient way. As a result of the thesis, a working implementation of expense tracking mobile application for Android appears. After functionality of expenses information extraction from receipt images passes the testing phase, conclusion regarding its reliability is made. Moreover, proposals for further improvements of the application’s functionality are also presented. The thesis is in English and contains 53 pages of text, 6 chapters and 14 figures. Annotatsioon Käesoleva bakalaureusetöö eesmärk on luua mobiilirakendus kasutaja kulude üle arvestuse pidamiseks ja dokumenteerimiseks.
    [Show full text]
  • Wen-Chieh Wu Technical LEAD · SOFTWARE Architect · FULL STACK Engineer
    Wen-Chieh Wu TECHNiCAL LEAD · SOFTWARE ARCHiTECT · FULL STACK ENGiNEER [email protected] | jeromewu.github.io | jeromewu | wenchiehwu | jeromewus Summary Passionate problem solver, technology enthusiast, compassionate leader with 7+ years as software engineer and architect and 5+ years as technical lead in multidisciplinary environments. Excel in leadership, critical thinking, communication and software engineering & architecture . Strong skills in JavaScript (incl. React, Express), Golang, Docker, CI/CD pipelines and software architecture design. Major contributor of tesseract.js (24.8k+ stars) and ffmpeg.wasm (6.1k+ stars). Certified architecting, data engineering, big data and machine learning specializations in Google Cloud Platform. Work Experience ByteDance Singapore STAFF SOFTWARE ENGiNEER Aug. 2021 ‑ PRESENT INTELLLEX Singapore TECHNiCAL LEAD Apr. 2020 ‑ Jul. 2021 • Led a team of backend and frontend engineers to develop knowledge platform for law. • Led re‑architecture project which eliminates 10+ redundant services, boosts observability and flexibility of data analysis pipeline and reduce cost • Reduced 30% of AWS cost in 2 months • Succeed in supporting to pass company ISO 27001 annual audit without any NC. Delta Electronics Taipei City, Taiwan SOFTWARE ARCHiTECT & PRODUCT MANAGER May 2019 ‑ Mar. 2020 • Established a cross‑region project team with 10+ members to research and develop an edge computing solution in manufacturing industry. • Designed a microservice software architecture for edge computing solution to enable flexiblity and scalibility. • Adopted gitlab‑CI and Ansible to deploy microservices in Kubernetes cluster to minize deployment efforts. • Developed a tensorflow serving like interface library in Golang. SENiOR TECHNiCAL LEAD Feb. 2018 ‑ May 2019 (1 year 4 months) • Established a corss‑region team of engineers, designers and QA with 20+ members in charge of front‑end development.
    [Show full text]
  • Multilingual String Verification for Automotive Instrument Cluster Using Artificial Intelligence
    WHITE PAPER www.visteon.com Multilingual String Verification for Automotive Instrument Cluster Using Artificial Intelligence Multilingual string verification for automotive instrument cluster Using Artificial Intelligence Deepan Raj M a, Prabu A b Software Engineer at Visteon Technical & Services Centre, Chennai a Technical Profession at Visteon Technical & Services Centre, Chennai b Abstract: Validation of HMI contents in Driver information system, In-Vehicle Infotainment, and Centre media Display is a big challenge as it prone human manual testing errors. Automating the text verification for multi languages and horizontal/vertical scrolling text is difficult challenge as the intelligence needed to achieve is high. To mitigate this effect vision based machine/deep learning algorithms will be implemented in such a way that three important applications have been created which serves to be the mother code for all the features that involves text as it main component of automated testing. They are 1. Multilingual text verification at various illuminations, grade out conditions and various background gradients using Opencv and Tesseract-OCR. 2. Horizontal & Vertical scrolling text using image key point stitching. 3. Automatic region of interest generation for text regions in the message center display using inception v3 deep neural network architecture. With the power of artificial intelligence in validating text makes it easier to automate with greater accuracy and can also be implemented in other areas like User Setting Menu traversal verification, Clock verification and much can be automated with less time and higher accuracy. Keywords: Computer Vision, Machine Learning, Deep learning, Neural Networks, OCR, automating validation. 1. INTRODUCTION: In recent years there has been a drastic improvement in the field of automobile, all the analog and semi digital clusters are being changed to full digital cluster, and lot of driver assisting feature are being added as the days goes on.
    [Show full text]
  • Android Optical Character Recognition
    Imperial Journal of Interdisciplinary Research (IJIR) Vol-3, Issue-4, 2017 ISSN: 2454-1362, http://www.onlinejournal.in Android Optical Character Recognition 1 2 3 Shital Malhar , Manasi Gosavi , Pooja Lad 1,2,3 Dilkap Research Institute of Engineering and Management Studies, Neral. (Mumbai University) operating system which can run on every mobile Abstract - The next generation open operating device and not for their specific mobile devices systems are not on desktops or mainframes but on itself. This enables them to reach as many people as the small mobile devices people carry every day. possible. This application is useful for native The openness of these new environments leads to Tourists and Travellers who possess Android Smart new applications and markets and enables greater phones. The main objective of this application is to integration. Every day a Smartphone user may look help tourist to travel easily and freely without any for a new application dedicated for his need. difficulty. The proposed application also provides Android makes it easier for consumers to get and currency convertor, a real market value. use new content and applications on their Smart phones. The Proposed project presents an 1.1 Problem Definition extremely on-demand, fast and user friendly Android Application. This application is useful for The Project presents an extremely on-demand, fast native Tourists and Travellers who possess Android and user friendly Android Application. This Smart phones. One of the feature in application is it application is useful for native Tourists and enables Travellers and Tourists to easily capture Travelers who possess Android Smart phones.
    [Show full text]
  • 5 VI June 2017
    5 VI June 2017 www.ijraset.com Volume 5 Issue VI, June 2017 IC Value: 45.98 ISSN: 2321-9653 International Journal for Research in Applied Science & Engineering Technology (IJRASET) An Innovative Communication System For Deaf, Dumb and Blind People 1 2 3 4 5 Anish Kumar , Rakesh Raushan , Saurabh Aditya , Vishal Kumar Jaiswal , Mrs. Divyashree Y.V. (Asst. Prof) 1,2,3,4.5 Dept. of ECE, SJBIT, Bangalore, Karnataka, India Abstract: One of the most precious gifts to a human being is an ability to see, listen, speak and respond according to the situations. But there are some unfortunate ones who are deprived of this. Making a single compact device for people with Visual, Hearing and Vocal impairment is a tough job. Communication between deaf-dumb and normal person have been always a challenging task. This paper proposes an innovative communication system framework for deaf, dumb and blind people in a single compact device. We provide a technique for a blind person to read a text and it can be achieved by capturing an image through a camera which converts a text to speech (TTS). It provides a way for the deaf people to read a text by speech to text (STT) conversion technology. Also, it provides a technique for dumb people using text to voice conversion. The system is provided with four switches and each switch has a different function. The blind people can be able to read the words using by Tesseract OCR (Online Character Recognition), the dumb people can communicate their message through text which will be read out by espeak, the deaf people can be able to hear others speech from text.
    [Show full text]
  • C++ Const and Immutability: an Empirical Study of Writes-Through-Const
    C++ const and Immutability: An Empirical Study of Writes-Through-const Jon Eyolfson1 and Patrick Lam2 1 University of Waterloo Waterloo, ON, Canada rtifact Comple * A t * te n * A te * W is E s e P n [email protected] l C l o D O C o * * c u O e m s E u C 2 University of Waterloo e e n R E t v e o d t y * s E a a * l d u e a Waterloo ON, Canada t [email protected] Abstract The ability to specify immutability in a programming language is a powerful tool for developers, enabling them to better understand and more safely transform their code without fearing unin- tended changes to program state. The C++ programming language allows developers to specify a form of immutability using the const keyword. In this work, we characterize the meaning of the C++ const qualifier and present the ConstSanitizer tool, which dynamically verifies a stricter form of immutability than that defined in C++: it identifies const uses that are either not con- sistent with transitive immutability, that write to mutable fields, or that write to formerly-const objects whose const-ness has been cast away. We evaluate a set of 7 C++ benchmark programs to find writes-through-const, establish root causes for how they fail to respect our stricter definition of immutability, and assign attributes to each write (namely: synchronized, not visible, buffer/cache, delayed initialization, and incorrect). ConstSanitizer finds 17 archetypes for writes in these programs which do not respect our version of immutability.
    [Show full text]
  • Developing Mobile Application to Help Disabled People
    MOHAMMED PEOPLE WITH MACULAR WITH TO DEGENERATION MOHAMMEDSEE PEOPLE RAFIA KHALLEEFAH DEVELOPING MOBILE APPLICATION TO HELP DISABLED PEOPLE WITH MACULAR DEGENERATION HAMAD A THESIS SUBMITTED TO THE GRADUATE SCHOOL OF APPLIED SCIENCES OF NEAR EAST UNIVERSITY DEVELOPING MOBILE MOBILE HELPAPPLICATION DISABLED TO DEVELOPING By RAFIA KHALLEEFAH HAMAD MOHAMMED In Partial Fulfilment of the Requirements for the Degree of Master of Science 2018 in Computer Information Systems NEU NICOSIA, 2018. DEVELOPING MOBILE APPLICATION TO HELP DISABLED PEOPLE WITH MACULAR DEGENERATION A THESIS SUBMITTED TO THE GRADUATE SCHOOL OF APPLIED SCIENCES OF NEAR EAST UNIVERSITY By RAFIA KHALLEEFAH HAMAD MOHAMMED In Partial Fulfilment of the Requirements for the Degree of Master of Science in Computer Information Systems NICOSIA, 2018. Rafia Khalleefah Hamad MOHAMMED: DEVELOPING MOBILE APPLICATION TO HELP DISABLED PEOPLE WITH MACULAR DEGENERATION Approval of Director of Graduate School of Applied Sciences Prof. Dr. Nadire CAVUS We certify this thesis is satisfactory for the award of the degree of Masters of Science in Computer Information Systems Examining Committee in Charg I hereby declare that all information in this document has been obtained and presented in accordance with academic rules and ethical conduct. I also declare that, as required by these rules and conduct, I have fully cited and referenced all material and results that are not original to this work. Name, Last name:RAFIA MOHAMMED Signature: Date:3/12/2018 i ACKNOWLEDGEMENTS This thesis would not have been possible without the help, support and patience of my principal supervisor, my deepest gratitude goes toProf.Dr. Dogan Ibrahim, for his constant encouragement and guidance.
    [Show full text]
  • Optical Character Recognition U3A in Bath FOSS Group
    OCR – Optical Character Recognition U3A in Bath FOSS group Optical Character Recognition (OCR) By Andy Pepperdine Optical Character Recognition is the process whereby a picture of text is turned into a textual form that can be edited by a computer. It is a difficult job to do what we do so easily with our eyes, and there has been a lot of work done on the subject with mixed results. This paper is intended to describe what is available free of charge. There are some excellent proprietary programs issued with scanners and if you have a system that can run those, it may be the best option for you. This month we will be looking at the cases where that option is not available, either because the scanner did not provide a good one, or you are running an operating system that can execute it. Preliminaries Before I start, I must ask that you check that you have the relevant copyright protections to enable you to make the copy and use the results as you wish. If in doubt, get advice. I will be using a printed version of the introduction to the Charter of Fundamental Rights of the European Union purely as an example. Since I printed it out, it has moved site to http://www.europarl.europa.eu/charter/default_en.htm What will be covered? The process of OCR starts with a picture of the text, usually from a scanner. This might be as a JPEG picture, or a PDF form of a picture. PDF's that contain text can usually be manipulated by a PDF reader (e.g.
    [Show full text]