Machine Learning Is Going Mobile

Total Page:16

File Type:pdf, Size:1020Kb

Machine Learning Is Going Mobile Machine learning is going mobile By David Schatsky ACHINE learning—the process by which Signals Mcomputers can get better at perform- ing tasks through exposure to data, rather • Google has introduced language translation than through explicit programming—requires software, using small neural networks opti- massive computational power, the kind usu- mized for mobile phones, which can per- ally found in clusters of energy-guzzling, form well without an Internet connection.1 cloud-based computer servers outfitted with specialized processors. But an emerging trend • Lenovo announced a mobile phone that promises to bring the power of machine learn- uses multiple sensors, high-speed image ing to mobile devices that may lack or have processing hardware, and specialized only intermittent online connectivity. This will Google software to support capabili- give rise to machines that sense, perceive, learn ties such as indoor wayfinding, precision from, and respond to their environment and measuring, and augmented reality even their users, enabling the emergence of new when offline.2 product categories, reshaping how businesses engage with customers, and transforming how • NVIDIA, a maker of graphics processing work gets done across industries. technology, introduced an embeddable module for computer vision applications Machine learning is going mobile in devices such as drones and autono- systems run in the cloud on powerful serv- mous vehicles that the company says ers, processing data such as digitized voice or consumes one-tenth the power of a photos that users upload. competing offering.3 Until recently, a typical smartphone lacked the power to perform such tasks without con- • Qualcomm introduced a new processor and necting to the cloud, except in limited ways. software platform that support machine For instance, some mobile phone software can learning tasks such as image classification, recognize a single face—the owner’s—in order speech recognition, and anomaly detection to unlock the phone, or a small set of predeter- without a connection to a network.4 mined words such as “OK Google.” But offline support for increasingly powerful perception • Drone maker DJI recently introduced tasks is coming to mobile devices. a consumer-oriented drone that uses advanced computer vision hardware to Pushing machine learning enable it to follow a moving object while onto mobile devices automatically avoiding obstacles.5 Firms are starting to outfit smartphones, drones, and cars with chips based on new Machine learning is the designs that can run neural networks efficiently keystone cognitive technology while consuming 90 percent less power than 11 Emerging technologies rarely get as big a previous generations. Research efforts at MIT publicity boost as machine learning recently and IBM suggest that we will soon see more saw, when Google software defeated one of the chips on the market that excel at running neu- world’s top players of Go, one of the most com- ral networks at high speed, in small spaces and 12 plex board games ever created, in a best-of-five at low power. Because of this, mobile devices series of matches.6 The international headlines are becoming increasingly capable of perform- confirmed that machine learning—the pro- ing sophisticated feats that take advantage of cess by which fresh data can teach computers neural networks, such as computer vision and to better perform tasks—is one of the hottest speech recognition, once reserved for powerful domains within the field of artificial intel- servers running in the cloud. ligence, and that this cognitive technology is It is not only progress in hardware that is progressing rapidly.7 bringing machine learning to mobile devices. Neural networks—computer models Tech vendors are also finding ways to create designed to mimic aspects of the human compact neural networks capable of running brain’s structure and function, with elements tasks such as speech recognition and language representing neurons and their interconnec- translation on conventional mobile phones tions—are an increasingly popular way of with no connection to a server required. implementing machine learning. They are For instance, Google has introduced mobile particularly well suited for performing percep- language-translation software using small tual tasks such as computer vision and speech neural networks optimized for smartphones 13 recognition. Familiar examples of applications that can perform well even offline. And that employ neural networks for such tasks Google researchers recently published a paper include Google’s voice search,8 Facebook’s sys- describing an Internet-independent speech tem for tagging people in photos,9 and Google recognition system that performs well on a 14 Photos, which uses a neural network-based commercial mobile phone. image recognition system to automatically Mobile devices are acquiring the power to classify photos by their contents.10 All of these perform sophisticated perceptual tasks with- out dependence on connectivity to the cloud, 2 Machine learning is going mobile Perceptual interfaces bringing greater accuracy, reliability, and and interactivity responsiveness while strengthening user pri- In media and entertainment, we will likely vacy. This should greatly expand the number of see mobile devices—both general-purpose applications of perceptual computing coming ones such as mobile phones and special-pur- to market—and not only on mobile phones. pose ones such as augmented-reality head- Mobile machine learning and perceptual com- sets—offering ever more realistic and engaging puting will power a wide range of devices, from augmented and virtual reality for games and mobile sensors to phones, tablets, drones, cars, filmed entertainment. and new types of devices as yet unimagined, Ultralow power processors designed creating significant opportunities for business. for machine learning will likely help con- sumer and industrial devices and machines Many industries will see new understand and respond to the environ- and improved applications ment around them, and find their way into It’s impossible to enumerate all of the Internet-independent voice-controlled wear- applications we will see for mobile devices able devices, household appliances, and capable of performing sophisticated percep- industrial machinery. tual tasks involving vision, speech, or other Navigation and motion control sensory input. But they are likely to be found in every industry and have one or more of the Low-power chips with powerful com- following capabilities: puter vision support are bringing impressive capabilities to unmanned aerial vehicles, also • Analysis or diagnosis of sensory data known as drones, which already have applica- tions in many industries, from real estate and • Perceptual interfaces or interactivity construction to agriculture, energy, aero- space, and defense. Drone maker DJI recently • Navigation and motion control introduced a consumer-oriented aerial vehicle A few examples follow. able to follow a moving object while automati- cally avoiding obstacles.16 Analysis or diagnosis New, powerful mobile computer vision In health care, we envision a wide range of modules that use deep learning are helping diagnostic applications, including some aimed advanced driver assistance systems to “address at consumers. Imagine, for instance, a smart- the challenges of everyday driving, such as phone app that can diagnose skin conditions unexpected road debris, erratic drivers and 17 and insect bites by analyzing digital photos construction zones.” without transmitting the image data over Indoor navigation apps that use computer a network. vision to precisely locate a user, track her We imagine mobile architecture and motion, and guide her in interior spaces will design applications that use computer vision to find use in museums, train stations, airports, generate accurate 3D models of interior spaces malls, and retail stores, opening up new adver- quickly and easily. tising and commerce opportunities without the An ever more powerful and resilient need to deploy beacons or other connectivity- Internet of Things will include self-monitor- based approaches. ing industrial equipment that uses machine And on the horizon are applications not learning to predict maintenance needs and yet imagined, from wearable to pocketable to self-diagnose failures.15 portable, that can sense, analyze, and respond to sensory inputs including sound, video, 3 Machine learning is going mobile and biometrics—all enabled by low-power from household appliances to personal robots chips designed to support neural networks for to industrial equipment. machine learning. Marketing leaders should explore how a new generation of perceptive devices could Implications help cultivate closer and more responsive rela- Compact, efficient, low-power, high- tionships with customers. performance, mobile machine learning. New Operations executives should evaluate how products. New human-computer interfaces. such devices—including the evolving crop of Powerful new ways of engaging with and serv- augmented-reality tools for industry—could ing customers. The trend described here has help their people deliver an efficiency and implications for companies and professionals quality edge.18 across industries. Cyber risk professionals should explore Makers of mobile devices and mobile apps how mobile machine learning may present should begin to familiarize themselves with new ways of detecting and mitigating threats the potential of a new generation
Recommended publications
  • RV - RA - RM Realidades Virtuales, Aumentadas Y Mixtas Índice General
    RV - RA - RM Realidades virtuales, aumentadas y mixtas Índice general 1 Definiciones 1 1.1 Realidad virtual ............................................ 1 1.1.1 Virtualidad .......................................... 1 1.1.2 Relación realidad/irrealidad ................................. 1 1.1.3 Inmersión y navegación .................................... 2 1.1.4 Usos ............................................. 2 1.1.5 Productos ........................................... 3 1.1.6 Técnicas de realidad virtual ................................. 6 1.1.7 Problemas de la realidad virtual ............................... 7 1.1.8 Tecnoética realidad virtual .................................. 8 1.1.9 Véase también ........................................ 9 1.1.10 Referencias .......................................... 9 1.1.11 Enlaces externos ....................................... 10 1.1.12 Bibliografía .......................................... 10 1.2 Realidad aumentada .......................................... 10 1.2.1 Definiciones ......................................... 11 1.2.2 Cronología .......................................... 11 1.2.3 Tecnología .......................................... 12 1.2.4 Técnicas de visualización ................................... 13 1.2.5 Elementos de la realidad aumentada ............................. 13 1.2.6 Niveles ............................................ 14 1.2.7 Aplicaciones ......................................... 14 1.2.8 Aplicaciones futuras ..................................... 16 1.2.9 Literatura
    [Show full text]
  • Augmented Reality Applied Tolanguage Translation
    Ana Rita de Tróia Salvado Licenciado em Ciências da Engenharia Electrotécnica e de Computadores Augmented Reality Applied to Language Translation Dissertação para obtenção do Grau de Mestre em Engenharia Electrotécnica e de Computadores Orientador : Prof. Dr. José António Barata de Oliveira, Prof. Auxiliar, Universidade Nova de Lisboa Júri: Presidente: Doutor João Paulo Branquinho Pimentão, FCT/UNL Arguente: Doutor Tiago Oliveira Machado de Figueiredo Cardoso, FCT/UNL Vogal: Doutor José António Barata de Oliveira, FCT/UNL September, 2015 iii Augmented Reality Applied to Language Translation Copyright c Ana Rita de Tróia Salvado, Faculdade de Ciências e Tecnologia, Universi- dade Nova de Lisboa A Faculdade de Ciências e Tecnologia e a Universidade Nova de Lisboa têm o direito, perpétuo e sem limites geográficos, de arquivar e publicar esta dissertação através de ex- emplares impressos reproduzidos em papel ou de forma digital, ou por qualquer outro meio conhecido ou que venha a ser inventado, e de a divulgar através de repositórios científicos e de admitir a sua cópia e distribuição com objectivos educacionais ou de in- vestigação, não comerciais, desde que seja dado crédito ao autor e editor. iv To my beloved family... vi Acknowledgements "Coming together is a beginning; keeping together is progress; working together is success." - Henry Ford. Life can only be truly enjoyed when people get together to create and share moments and memories. Greatness can be easily achieved by working together and being sup- ported by others. For this reason, I would like to save a special place in this work to thank people who were there and supported me during all this learning process.
    [Show full text]
  • Echtzeitübersetzung Dank Augmented Reality Und Spracherkennung Eine Untersuchung Am Beispiel Der Google Translate-App
    ECHTZEITÜBERSETZUNG DANK AUGMENTED REALITY UND SPRACHERKENNUNG EINE UNTERSUCHUNG AM BEISPIEL DER GOOGLE TRANSLATE-APP Julia Herb, BA Matrikelnummer: 01216413 MASTERARBEIT eingereicht im Rahmen des Masterstudiums Translationswissenschaft Spezialisierung: Fachkommunikation an der Leopold-Franzens-Universität Innsbruck Philologisch-Kulturwissenschaftliche Fakultät Institut für Translationswissenschaft betreut von: ao. Univ.-Prof. Mag. Dr. Peter Sandrini Innsbruck, am 22.07.2019 1 Inhaltsverzeichnis Abstract ........................................................................................................................................................... 1 Abbildungsverzeichnis .................................................................................................................................... 2 1. Aktualität des Themas ............................................................................................................................. 4 2. Augmented Reality: die neue Form der Multimedialität ........................................................................ 7 a. Definition und Abgrenzung ................................................................................................................ 7 b. Anwendungsbereiche der AR-Technologie ...................................................................................... 11 3. Verbmobil: neue Dimensionen der maschinellen Übersetzung dank Spracherkennung ....................... 13 a. Besonderheiten eines Echtzeit-Übersetzungs-Projekts ....................................................................
    [Show full text]
  • Augmented Reality Based Word Translator Tabish Khan, Rishisingh Hora, Ashwin Bendre, Prof
    International Journal of Innovative Research in Computer Science & Technology (IJIRCST) ISSN: 2347-5552, Volume-2, Issue-2, March 2014 Augmented Reality Based Word Translator Tabish Khan, Rishisingh Hora, Ashwin Bendre, Prof. Sneha Tirth developed for iOS. The translation it provides is not always Abstract— People travel to different places not knowing 100 percent accurate. It has a hit rate of 90-95 percent. the language used in that region. Hence there is a need to translate these unknown words to recognizable text. This application is developed to help travelers who can get the translated text as an output of our application. Our application is able to recognize the text captured by a mobile phone camera, translate the text and display the translation result back onto the screen of the mobile phone in an augmented manner. Index Terms— Android, Augmented Reality, Optical Character Recognition, Word Translator. NTRODUCTION I. I The motivation of our application is to help tourist FIG 1: Output of word lens navigate in a foreign language environment. The The above figure depicts an example of the word lens application we developed enables the users to get text application which translates the words “Bienvenido Al translated as easy as a button click. The camera captures the Futuro” to the desired English language, the output being, text and returns the translated result in the real time. “Welcome to the future”. The system we developed includes automatic text detection, OCR and text translation. Although the current version of our application is limited to translation of a few languages, III. PROPOSED SYSTEM we can add a feature of text-to-speech, as an extension in Inspired by the existing application, we will use the same our application.
    [Show full text]
  • Final Thesis1
    Chapter 3. HISTORICAL BACKGROUND OF QR CODE & AR APPLICATION 3.1. What is a QR Code QR code (abbreviated from Quick Response Code) is the trademark for a type of matrix barcode (or two-dimensional bar code) A QR code is a matrix barcode readable by smart phones and mobile phones with cameras. They are sometimes referred to as 2d codes, 2d barcodes, or mobile codes. QR codes can hold much more information than a regular barcode. The information encoded in a QR code can be a URL, a phone number, an SMS message, a V-card, or any text. They are referred to as QR because they allow the contents to be decoded at high speed. QR codes were developed in 1994 by Denso-Wave, a Toyota subsidiary In Catalog records to offer patrons basic information about location of books and call number, users can scan the code and head to the stacks rather than writing or printing, this means, for example, a student who scans QR codes from a library OPAC in the middle of the night in his or her dorm room could head to the library stacks the next day, click open the app to refer to the call numbers of the titles scanned the night before, and quickly find the books. Also, expect to see QR code scanner/reader apps that allow users to tag, share, comment, collaborate, and more in the future. Taped to video/DVD cases, linking to mobile-friendly video trailers, Codes in the library stacks/end caps or magazine/journal areas that point to online electronic holdings of print materials or related subject guides, linking to library audio tours for orientations etc.
    [Show full text]
  • Arxiv:1611.08903V1 [Cs.LG] 27 Nov 2016 20 Percent Increases in Cash Collections [20]
    Should I use TensorFlow? An evaluation of TensorFlow and its potential to replace pure Python implementations in Machine Learning Martin Schrimpf1;2;3 1 Augsburg University 2 Technische Universit¨atM¨unchen 3 Ludwig-Maximilians-Universit¨atM¨unchen Seminar \Human-Machine Interaction and Machine Learning" Supervisor: Elisabeth Andr´e Advisor: Dominik Schiller Abstract. Google's Machine Learning framework TensorFlow was open- sourced in November 2015 [1] and has since built a growing community around it. TensorFlow is supposed to be flexible for research purposes while also allowing its models to be deployed productively [7]. This work is aimed towards people with experience in Machine Learning consider- ing whether they should use TensorFlow in their environment. Several aspects of the framework important for such a decision are examined, such as the heterogenity, extensibility and its computation graph. A pure Python implementation of linear classification is compared with an im- plementation utilizing TensorFlow. I also contrast TensorFlow to other popular frameworks with respect to modeling capability, deployment and performance and give a brief description of the current adaption of the framework. 1 Introduction The rapidly growing field of Machine Learning has been gaining more and more attention, both in academia and in businesses that have realized the added value. For instance, according to a McKinsey report, more than a dozen European banks switched from statistical-modeling approaches to Machine Learning tech- niques and, in some cases, increased their sales of new products by 10 percent and arXiv:1611.08903v1 [cs.LG] 27 Nov 2016 20 percent increases in cash collections [20]. Intelligent machines are used in a variety of domains, including writing news articles [5], finding promising recruits given their CV [9] and many more.
    [Show full text]
  • Applied Arts Department Cooks up a New Course
    2 THE NEW TRIER NEWS FEATURES FRIDAY, SEPTEMBER 13, 2013 Applied Arts Department cooks up a new course by Marissa Rogina didn’t have room for it.” Outside traditional recipes from these regions of culinary courses, Applied Arts and learn about how these cultures This year, New Trier’s Applied is also expanding other areas of and regions influence their cuisine. Arts Department added a new course its department. “You can see this Since this course is only offered to to its repertoire—International firsthand with our newer courses seniors it wraps up with a unit on Food. This class has been popular, such as Human Growth and Child cooking in college. featuring four class sections this Development, Project Lead the Cooking and baking are not the year. International Foods is different Way Engineering Courses, and our only things that students do to fill from the other culinary courses at Architectural course sequence,” their forty minutes in International New Trier—Creative Cuisine and continued Boumstein. Foods. They also spend time learning Gourmet Foods—because the foods Many students are confusing the about the various cultures and students cook are centered on what addition of International Foods with countries from which they cook up culture and influence is being studied Gourmet Foods, yet the two courses cuisine. Teachers initiate discussions in class. are actually very different. about these cultures and countries International Foods was Gourmet foods, offered only to that spark students to ask questions added because of the Applied Arts sophomores and juniors, focuses on and explore more about where the departments’ vision to offer more different types of gourmet cooking foods they make come from.
    [Show full text]
  • An Augmented Vision Tool with Real- Time Sensing Ability in Tagged Environments
    X-Vision: An Augmented Vision Tool with Real- Time Sensing Ability in Tagged Environments The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation Sun, Yongbin et al. "X-Vision: An Augmented Vision Tool with Real-Time Sensing Ability in Tagged Environments." 2018 IEEE International Conference on RFID Technology & Application, September 2018, Macau, China, Institute of Electrical and Electronics Engineers (IEEE), December 2018. © 2018 IEEE As Published http://dx.doi.org/10.1109/rfid-ta.2018.8552778 Publisher Institute of Electrical and Electronics Engineers (IEEE) Version Author's final manuscript Citable link https://hdl.handle.net/1721.1/123808 Terms of Use Creative Commons Attribution-Noncommercial-Share Alike Detailed Terms http://creativecommons.org/licenses/by-nc-sa/4.0/ X-Vision: An augmented vision tool with real-time sensing ability in tagged environments Yongbin Sun* , Sai Nithin R. Kantareddy* , Rahul Bhattacharyya , and Sanjay E. Sarma Auto-ID Labs, Department of Mechanical Engineering Massachusetts Institute of Technology Cambridge, USA fyb sun, nithin, rahul b, sesarmag at mit.edu Abstract—We present the concept of X-Vision, an enhanced Augmented Reality (AR)-based visualization tool, with the real- time sensing capability in a tagged environment. We envision that this type of a tool will enhance the user-environment interaction and improve the productivity in factories, smart- spaces, home & office environments, maintenance/facility rooms and operation theatres, etc. In this paper, we describe the design of this visualization system built upon combining the object’s pose information estimated by the depth camera and the object’s ID & physical attributes captured by the RFID tags.
    [Show full text]
  • X-Vision: an Augmented Vision Tool with Real-Time Sensing Ability in Tagged Environments
    X-Vision: An augmented vision tool with real-time sensing ability in tagged environments Yongbin Sun* , Sai Nithin R. Kantareddy* , Rahul Bhattacharyya , and Sanjay E. Sarma Auto-ID Labs, Department of Mechanical Engineering Massachusetts Institute of Technology Cambridge, USA fyb sun, nithin, rahul b, sesarmag at mit.edu Abstract—We present the concept of X-Vision, an enhanced Augmented Reality (AR)-based visualization tool, with the real- time sensing capability in a tagged environment. We envision that this type of a tool will enhance the user-environment interaction and improve the productivity in factories, smart- spaces, home & office environments, maintenance/facility rooms and operation theatres, etc. In this paper, we describe the design of this visualization system built upon combining the object’s pose information estimated by the depth camera and the object’s ID & physical attributes captured by the RFID tags. We built a Fig. 1. Left: A user wearing the system sees a cup with overlaid temperature information. Right: System components: an Intel RealSense D415 RGB-D physical prototype of the system demonstrating the projection of camera is attached on a HoloLens via a custom mount. 3D holograms of the objects encoded with sensed information like water-level and temperature of common office/household objects. The paper also discusses the quality metrics used to compare the pose estimation algorithms for robust reconstruction of the attributes can be wirelessly retrieved and mapped to a digital object’s 3D data. avatar. In this paper, we have designed a visualization framework I. INTRODUCTION called X-Vision, hoping to equip users with the ability to di- Superimposing information on to the real-world, the concept rectly see the physical attributes of surrounding objects (Figure commonly known to us as Augmented reality (AR), has been 1).
    [Show full text]
  • 2015-Nmc-Technology-Outlook-Brazilian-Universities-PT.Pdf
    Panorama Tecnológico NMC 2015 Universidades Brasileiras Uma Análise Regional do Horizon Project Resumo Executivo ........................................................................................................................................................................................ 3 Introdução .............................................................................................................................................................................................................. 3 Principais Tendências que Aceleram a Adoção de Tecnologia ........................................................ 6 Desafios Significativos que Impedem a Adoção de Tecnologia ................................................. 10 Desenvolvimentos Importantes em Tecnologia Educacional Período para Adoção: 1 Ano ou Menos § Computação na Nuvem (Cloud Computing) ...................................................................................... 12 § Publicação Eletrônica ..................................................................................................................................... 13 § Aprendizagem On-line ................................................................................................................................. 14 § Redes Sociais ..................................................................................................................................................... 15 Período para Adoção: 2 a 3 Anos § Traga Seu Próprio Dispositivo (Bring Your Own Device) ..............................................................
    [Show full text]
  • L'avenir DE L'innovation Chroniques Digitales D'un Tour Du Monde
    L'AVENIR DE L'INNOVATION Chroniques digitales d'un tour du monde Francis Pisani Publié par l’Observatoire Netexplo © Netexplo. 2016 Tous droits réservés ISBN : 978-2-9546672-2-5 Note de l’auteur : Les textes mentionnés en bleu font référence à des informations complémentaires disponibles sur le web. La version numérique de ce livre, consultable sur www.netexplo.org, permet un accès direct à ces informations. A Colette, ma mère et à Henri, mon beau-père qui m’ont donné l’envie de connaître d’autres terres. A Edgard, mon père qui m’a aidé à comprendre qu’il venait d’ailleurs, comme nous tous. A Fabien, Émilia, et Yara mes enfants nomades. A Catherine, qui me donne le goût de voyages différents. A ceux, qui maintiennent la tête hors des « boîtes » dans lesquelles il est si facile de trouver ses aises… et à ceux, qui courent le risque de se mettre en marge, pratiquent la désobéissance créatrice, poussent le nez dehors et s’inspirent de ce qu’ils rencontrent pour mieux faire bouger le schmilblick… PARTIR … Mes tours du monde de l'innovation sont d'abord la continuation naturelle des 15 ans que j'ai passés au Mexique (de 1980 à 1995) et des 15 ans qui ont suivi en Californie (1996-2010). Dans un cas comme dans l'autre, j'étais en quête de ce qui peut changer le monde. J'avais choisi Mexico pour suivre les bouleversements d'Amérique centrale lancés par l'insurrection du Nicaragua (quand les sandinistes portaient encore un plus de démocratie). J'ai déménagé à San Francisco pour couvrir l’essor des technologies de l'information, telles qu'on les conçoit et qu'on les crée autour de Silicon Valley.
    [Show full text]
  • Computer Vision News
    The magazine of the algorithm community December 2017 Challenge DCER & HPE - Emotions Recognition and Head-Pose Estimation Women in Computer Vision: Spotlight News Sarah Ostadabbas Project Management: Navigation through the early stages of the project by Ron Soferman Upcoming Events Image Processing: There is no bad OCR, there is OCR which wasn’t served well Review of Research Paper by ETH Zürich: DSLR-Quality Photos on Mobile Devices A publication by with Deep Convolutional Networks 2 Read This Month Computer Vision News Research by ETH Project Woman in Computer Vision DSLR-Quality Photos on Mobile There is No Bad OCR Sarah Ostadabbas 04 24 Applications Eyes of Things - Bonseyes 18 Upcoming Events with Oscar Deniz Meet us at… Challenge: DCER & HPE Emotions Recognition and Head-Pose Estimation 12 20 29 Project Management Spotlight News by Ron Soferman Workshop CoBCoM Winter School 16 23 30 Editorial Challenge - with Sylwia Hyniewska 03 by Ron Soferman 20 Emotions Recogn. and Head-Pose Est. Research Paper - by ETH Zürich Project Management Tip 04 DSLR-Quality Photos on Mobile Devices 23 Lecture by Ron Soferman Applications - with Oscar Deniz Women in Computer Vision 12 Eyes of Things - Bonseyes 24 Sarah Ostadabbas - ACLab Spotlight News Computer Vision Events 16 From Elsewhere on the Web 29 Upcoming events Dec - Feb Project - by RSIP Vision Workshop - CoBCoM 18 There is No Bad OCR, there is … 30 Computational brain connectivity mapping Welcome 3 Computer Vision News Dear reader, December is the right time to reflect on the past year and
    [Show full text]