Demystifying Artificial Intelligence What Business Leaders Need to Know About Cognitive Technologies

Total Page:16

File Type:pdf, Size:1020Kb

Demystifying Artificial Intelligence What Business Leaders Need to Know About Cognitive Technologies Demystifying artificial intelligence What business leaders need to know about cognitive technologies A Deloitte series on cognitive technologies Deloitte Consulting LLP’s Enterprise Science offering employs data science, cognitive technologies such as machine learning, and advanced algorithms to create high-value solutions for clients. Services include cognitive automation, which uses cognitive technologies such as natural language processing to automate knowledge-intensive processes; cognitive engagement, which applies machine learning and advanced analytics to make customer interactions dramatically more personalized, relevant, and profitable; and cognitive insight, which employs data science and machine learning to detect critical patterns, make high-quality predictions, and support business performance. For more information about the Enterprise Science offering, contact Plamen Petrov (ppetrov@deloitte. com) or Rajeev Ronanki ([email protected]). What business leaders need to know about cognitive technologies Contents Overview | 2 Artificial intelligence and cognitive technologies | 3 Cognitive technologies are already in wide use | 8 Why the impact of cognitive technologies is growing | 10 How can your organization apply cognitive technologies? | 13 Endnotes | 14 1 Demystifying artificial intelligence Overview N the last several years, interest in artificial positive change, but also risk significant Iintelligence (AI) has surged. Venture capital negative consequences as well, including investments in companies developing and mass unemployment.6 commercializing AI-related products and technology have exceeded $2 billion since • Silicon Valley entrepreneur Elon Musk is 2011.1 Technology companies have invested investing in AI “to keep an eye” on it.7 He billions more acquiring AI startups. Press cov- has said it is potentially “more dangerous erage of the topic has been breathless, fueled than nukes.”8 by the huge investments and by pundits asserting that computers are starting to kill jobs, will soon be smarter than people, and could A useful definition of artificial threaten the survival of human- kind. Consider the following: intelligence is the theory and • IBM has committed $1 billion development of computer to commercializing Watson, its cognitive computing platform.2 systems able to perform • Google has made major invest- tasks that normally require ments in AI in recent years, including acquiring eight robot- human intelligence. ics companies and a machine- learning company.3 • Facebook hired AI luminary Yann LeCun • Renowned theoretical physicist Stephen to create an AI laboratory with the goal of Hawking said that success in creating true 4 bringing major advances in the field. AI could mean the end of human history, “unless we learn how to avoid the risks.”9 • Researchers at the University of Oxford published a study estimating that 47 percent Amid all the hype, there is significant of total US employment is “at risk” due to commercial activity underway in the area of the automation of cognitive tasks.5 AI that is affecting or will likely soon affect organizations in every sector. Business leaders • The New York Times bestseller The Second should understand what AI really is and where Machine Age argued that digital technolo- it is heading. gies and AI are poised to bring enormous 2 What business leaders need to know about cognitive technologies Artificial intelligence and cognitive technologies HE first steps in demystifying AI are The history of artificial Tdefining the term, outlining its history, intelligence and describing some of the core technologies underlying it. AI is not a new idea. Indeed, the term itself dates from the 1950s. The history of the field is marked by “periods of hype and high expecta- 10 Defining artificial intelligence tions alternating with periods of setback and The field of AI suffers from both too few disappointment,” as a recent apt summation and too many definitions. Nils Nilsson, one of puts it.16 After articulating the bold goal of the founding researchers in the field, has writ- simulating human intelligence in the 1950s, ten that AI “may lack an agreed-upon defini- researchers developed a range of demonstra- tion. .”11 A well-respected AI textbook, now tion programs through the 1960s and into the in its third edition, offers eight definitions, and ‘70s that showed computers able to accom- declines to prefer one over the other.12 For us, a plish a number of tasks once thought to be useful definition of AI is the theory and devel- solely the domain of human endeavor, such as opment of computer systems able to perform proving theorems, solving calculus problems, tasks that normally require human intelligence. responding to commands by planning and Examples include tasks such as visual percep- performing physical actions—even imperson- tion, speech recognition, decision making ating a psychotherapist and composing music. under uncertainty, learning, and translation But simplistic algorithms, poor methods for between languages.13 Defining AI in terms of handling uncertainty (a surprisingly ubiqui- the tasks humans do, rather than how humans tous fact of life), and limitations on computing think, allows us to discuss its practical appli- power stymied attempts to tackle harder or cations today, well before science arrives at a definitive understand- ing of the neurological mechanisms of intelligence.14 It is worth noting that the set of tasks that normally require human intelligence is subject to change as computer systems able to perform those tasks are invented and then widely diffused. Thus, the meaning of “AI” evolves over time, a phenomenon known as the “AI effect,” concisely stated as “AI is whatever hasn’t been done yet.”15 3 Demystifying artificial intelligence more diverse problems. Amid disappointment Intel cofounder Gordon Moore, has benefited with a lack of continued progress, AI fell out of all forms of computing, including the types AI fashion by the mid-1970s. researchers use. Advanced system designs that In the early 1980s, Japan launched a might have worked in principle were in prac- program to develop an advanced computer tice off limits just a few years ago because they architecture that could advance the field of AI. required computer power that was cost-pro- Western anxiety about losing ground to Japan hibitive or just didn’t exist. Today, the power contributed to decisions to invest anew in AI. necessary to implement these designs is readily The 1980s saw the launch of commercial ven- available. A dramatic illustration: The current dors of AI technology products, some of which generation of microprocessors delivers 4 mil- had initial public offerings, such as Intellicorp, lion times the performance of the first single- Symbolics,17 and Teknowledge.18 By the end of chip microprocessor introduced in 1971.20 the 1980s, perhaps half of the Fortune 500 were Big data. Thanks in part to the Internet, developing or maintaining “expert systems,” social media, mobile devices, and low-cost an AI technology that models human exper- sensors, the volume of data in the world is tise with a knowledge base of facts and rules.19 increasing rapidly.21 Growing understanding of High hopes for the potential of expert systems the potential value of this data22 has led to the were eventually tempered as their limitations, development of new techniques for manag- including a glaring lack of common sense, the ing and analyzing very large data sets.23 Big difficulty of capturing experts’ tacit knowledge, data has been a boon to the development of and the cost and complexity of building and AI. The reason is that some AI techniques use maintaining large systems, became widely statistical models for reasoning probabilisti- recognized. AI ran out of steam again. cally about data such as images, text, or speech. In the 1990s, technical work on AI con- These models can be improved, or “trained,” by tinued with a lower profile. Techniques such exposing them to large sets of data, which are as neural networks and genetic algorithms now more readily available than ever.24 received fresh attention, in part because they The Internet and the cloud. Closely related avoided some of the limitations of expert to the big data phenomenon, the Internet systems and partly because new algorithms and cloud computing can be credited with made them more effective. The design of advances in AI for two reasons. First, they neural networks is inspired by the structure of make available vast amounts of data and infor- the brain. Genetic algorithms aim to “evolve” mation to any Internet-connected comput- solutions to problems by iteratively generating ing device. This has helped propel work on candidate solutions, culling the weakest, and AI approaches that require large data sets.25 introducing new solution variants by introduc- Second, they have provided a way for humans ing random mutations. to collaborate—sometimes explicitly and at other times implicitly—in helping to train AI Catalysts of progress systems. For example, some researchers have used cloud-based crowdsourcing services By the late 2000s, a number of factors like Mechanical Turk to enlist thousands of helped renew progress in AI, particularly in humans to describe digital images, enabling a few key technologies. We explain the fac- image classification algorithms to learn from tors most responsible for the recent progress these descriptions.26 Google’s language transla- below and then describe those technologies in tion project analyzes feedback and freely offers more detail. contributions from its users to improve the Moore’s Law. The relentless increase in quality of automated translation.27 computing power available at a given price and New algorithms. An algorithm is a routine size, sometimes known as Moore’s Law after process for solving a program or performing a 4 What business leaders need to know about cognitive technologies Figure 1. The field of artificial intelligence has produced a number of cognitive technologies Natural Computer Machine language vision learning processing Speech recognition Optimization Rules- Robotics Planning based systems & scheduling Graphic: Deloitte University Press | DUPress.com task.
Recommended publications
  • What's the Point: Semantic Segmentation with Point Supervision
    What's the Point: Semantic Segmentation with Point Supervision Amy Bearman1, Olga Russakovsky2, Vittorio Ferrari3, and Li Fei-Fei1 1 Stanford University fabearman,[email protected] 2 Carnegie Mellon University [email protected] 3 University of Edinburgh [email protected] Abstract. The semantic image segmentation task presents a trade-off between test time accuracy and training-time annotation cost. Detailed per-pixel annotations enable training accurate models but are very time- consuming to obtain; image-level class labels are an order of magnitude cheaper but result in less accurate models. We take a natural step from image-level annotation towards stronger supervision: we ask annotators to point to an object if one exists. We incorporate this point supervision along with a novel objectness potential in the training loss function of a CNN model. Experimental results on the PASCAL VOC 2012 benchmark reveal that the combined effect of point-level supervision and object- ness potential yields an improvement of 12:9% mIOU over image-level supervision. Further, we demonstrate that models trained with point- level supervision are more accurate than models trained with image-level, squiggle-level or full supervision given a fixed annotation budget. Keywords: semantic segmentation, weak supervision, data annotation 1 Introduction At the forefront of visual recognition is the question of how to effectively teach computers new concepts. Algorithms trained from carefully annotated data enjoy better performance than their weakly supervised counterparts (e.g., [1] vs. [2], [3] vs. [4], [5] vs. [6]), yet obtaining such data is very time-consuming [5, 7]. It is particularly difficult to collect training data for semantic segmentation, i.e., the task of assigning a class label to every pixel in the image.
    [Show full text]
  • Feature Mining: a Novel Training Strategy for Convolutional Neural
    Feature Mining: A Novel Training Strategy for Convolutional Neural Network Tianshu Xie1 Xuan Cheng1 Xiaomin Wang1 [email protected] [email protected] [email protected] Minghui Liu1 Jiali Deng1 Ming Liu1* [email protected] [email protected] [email protected] Abstract In this paper, we propose a novel training strategy for convo- lutional neural network(CNN) named Feature Mining, that aims to strengthen the network’s learning of the local feature. Through experiments, we find that semantic contained in different parts of the feature is different, while the network will inevitably lose the local information during feedforward propagation. In order to enhance the learning of local feature, Feature Mining divides the complete feature into two com- plementary parts and reuse these divided feature to make the network learn more local information, we call the two steps as feature segmentation and feature reusing. Feature Mining is a parameter-free method and has plug-and-play nature, and can be applied to any CNN models. Extensive experiments demonstrate the wide applicability, versatility, and compatibility of our method. 1 Introduction Figure 1. Illustration of feature segmentation used in our method. In each iteration, we use a random binary mask to Convolution neural network (CNN) has made significant divide the feature into two parts. progress on various computer vision tasks, 4.6, image classi- fication10 [ , 17, 23, 25], object detection [7, 9, 22], and seg- mentation [3, 19]. However, the large scale and tremendous parameters of CNN may incur overfitting and reduce gener- training strategy for CNN for strengthening the network’s alizations, that bring challenges to the network training.
    [Show full text]
  • Arxiv:2004.07999V4 [Cs.CV] 23 Jul 2021
    REVISE: A Tool for Measuring and Mitigating Bias in Visual Datasets Angelina Wang · Alexander Liu · Ryan Zhang · Anat Kleiman · Leslie Kim · Dora Zhao · Iroha Shirai · Arvind Narayanan · Olga Russakovsky Abstract Machine learning models are known to per- petuate and even amplify the biases present in the data. However, these data biases frequently do not become apparent until after the models are deployed. Our work tackles this issue and enables the preemptive analysis of large-scale datasets. REVISE (REvealing VIsual bi- aSEs) is a tool that assists in the investigation of a visual dataset, surfacing potential biases along three Fig. 1: Our tool takes in as input a visual dataset and its dimensions: (1) object-based, (2) person-based, and (3) annotations, and outputs metrics, seeking to produce geography-based. Object-based biases relate to the size, insights and possible actions. context, or diversity of the depicted objects. Person- based metrics focus on analyzing the portrayal of peo- ple within the dataset. Geography-based analyses con- sider the representation of different geographic loca- the visual world, representing a particular distribution tions. These three dimensions are deeply intertwined of visual data. Since then, researchers have noted the in how they interact to bias a dataset, and REVISE under-representation of object classes (Buda et al., 2017; sheds light on this; the responsibility then lies with Liu et al., 2009; Oksuz et al., 2019; Ouyang et al., 2016; the user to consider the cultural and historical con- Salakhutdinov et al., 2011; J. Yang et al., 2014), ob- text, and to determine which of the revealed biases ject contexts (Choi et al., 2012; Rosenfeld et al., 2018), may be problematic.
    [Show full text]
  • AI for All by Mr. S. Arjun
    AI for All Mr. S. Arjun 2nd Year, B.Tech CSE (Hons.), Lovely Professional University, Punjab [email protected] What is AI Many problems in the world that once seemed impossible for a computer to tackle without human intervention are solved today with Artificial Intelligence. We are witnessing the second major wave of AI, disrupting a plethora of unrelated fields such as health, ethics, politics, and economy. These intelligent systems prove that machines too can learn from experience, adapt, and make meaningful decisions. While the first wave was driven by rule-based systems where experts in performing a task handcrafted a set of rules for machines to follow, thus emulating intelligence, the second wave is driven by huge amounts of data, coupled with algorithms that enable machines to recognize patterns and learn from experience. Impact Though at a very early stage, AI has already made a profound impact on our lives. However, the nature of the impact it has made on different establishments is as unique as itself. On Enterprises Enterprises are seen to make the best out of the second wave of AI, primarily owing to the abundance of data they already collect every second. Deep Learning - a subset of Machine Learning, that allows recognizing and learning from complex patterns in data - feeds on huge magnitudes of data in the order of millions of samples. Large enterprises have the benefit of being capable of collecting such data in-house for their own systems, products, and services. Today, numerous such organizations that operate at scale are starting to use AI to automate workflows, streamline their operations, and optimize production.
    [Show full text]
  • Crowdsourcing in Computer Vision
    Foundations and Trends R in Computer Graphics and Vision Vol. 10, No. 3 (2014) 177–243 c 2016 A. Kovashka, O. Russakovsky, L. Fei-Fei and K. Grauman DOI: 10.1561/0600000073 Crowdsourcing in Computer Vision Adriana Kovashka Olga Russakovsky University of Pittsburgh Carnegie Mellon University [email protected] [email protected] Li Fei-Fei Kristen Grauman Stanford University University of Texas at Austin [email protected] [email protected] Contents 1 Introduction 178 2 What annotations to collect 181 2.1 Visual building blocks . 182 2.2 Actions and interactions . 191 2.3 Visual story-telling . 198 2.4 Annotating data at different levels . 204 3 How to collect annotations 205 3.1 Interfaces for crowdsourcing and task managers . 205 3.2 Labeling task design . 207 3.3 Evaluating and ensuring quality . 211 4 Which data to annotate 215 4.1 Active learning . 215 4.2 Interactive annotation . 221 5 Conclusions 226 References 228 ii Abstract Computer vision systems require large amounts of manually annotated data to properly learn challenging visual concepts. Crowdsourcing plat- forms offer an inexpensive method to capture human knowledge and un- derstanding, for a vast number of visual perception tasks. In this survey, we describe the types of annotations computer vision researchers have collected using crowdsourcing, and how they have ensured that this data is of high quality while annotation effort is minimized. We begin by discussing data collection on both classic (e.g., object recognition) and recent (e.g., visual story-telling) vision tasks. We then summarize key design decisions for creating effective data collection interfaces and workflows, and present strategies for intelligently selecting the most important data instances to annotate.
    [Show full text]
  • Cornernet-Lite- Efficient Keypoint Based Object Detection
    H. LAW, Y. TANG, O. RUSSAKOVSKY, J. DENG: CORNERNET-LITE 1 CornerNet-Lite: Efficient Keypoint-Based Object Detection Hei Law Department of Computer Science [email protected] Princeton University Yun Tang Princeton, NJ USA [email protected] Olga Russakovsky [email protected] Jia Deng [email protected] Abstract Keypoint-based methods are a relatively new paradigm in object detection, eliminat- ing the need for anchor boxes and offering a simplified detection framework. Keypoint- based CornerNet achieves state of the art accuracy among single-stage detectors. How- ever, this accuracy comes at high processing cost. In this work, we tackle the problem of efficient keypoint-based object detection and introduce CornerNet-Lite. CornerNet-Lite is a combination of two efficient variants of CornerNet: CornerNet-Saccade, which uses an attention mechanism to eliminate the need for exhaustively processing all pixels of the image, and CornerNet-Squeeze, which introduces a new compact backbone architecture. Together these two variants address the two critical use cases in efficient object detection: improving efficiency without sacrificing accuracy, and improving accuracy at real-time efficiency. CornerNet-Saccade is suitable for offline processing, improving the efficiency arXiv:1904.08900v2 [cs.CV] 16 Sep 2020 of CornerNet by 6.0x and the AP by 1.0% on COCO. CornerNet-Squeeze is suitable for real-time detection, improving both the efficiency and accuracy of the popular real-time detector YOLOv3 (34.4% AP at 30ms for CornerNet-Squeeze compared to 33.0% AP at 39ms for YOLOv3 on COCO). Together these contributions for the first time reveal the potential of keypoint-based detection to be useful for applications requiring processing efficiency.
    [Show full text]
  • Deep Learning with Spherical Data Min Sun National Tsing Hua University Vision Science Lab Self-Intro
    Deep Learning with Spherical Data Min Sun National Tsing Hua University Vision Science Lab http://aliensunmin.github.io/ Self-Intro 遠見專訪 https://www.gvm.com.tw/article.html?id=43046 Large-scale video analysis Video Title Generation Video QA Visual Forecasting ECCV 2016 AAAI 2017 ICCV 2017 ACCV 2016 Best Paper Award CVPR 2017 Car Accident Anticipation Risk Assessment Handcam Object Recognition via Handcam Grasp Assistant for BVI ECCV 2016 Under Submission 2018 ICCV 2017 Intention Anticipation 360 (Spherical) Vision Deep Pilot Saliency Prediction CVPR 2017 CVPR 2018 AAAI 2018 360 Grounding RoBotic Vision Learn from Demonstration Indoor Place Recognition & Navigation ICRA 2018 Robust Control Health Care Outline •Introduction to Deep Learning •Spherical data •Deep Learning for Spherical data Deep Learning (DL) • Data • GPU Computing • Talents Data: • 開始於 2007 @ Princeton • 初登場於 2009 @ CVPR • 照⽚停⽌搜集於 2010 Ø總共類別: 21841 Jia Deng Fei-Fei Li Ø總共圖⽚: 1千 4百萬 • ILSVR Challenge 從 2010到現今 Info from http://www.image-net.org/ 1K Image Classification Label = f(Image) Deep Learning 深度學習 Figure from Olga Russakovsky ECCV'14 workshop GPU: NVIDIA CUDA Tesla P100 With Over 20 TFLOPS Of FP16 Read more: http://wccftech.com/nvidia-pascal-gpu-gtc-2016/#ixzz456KT75Jf FaceBook Machine Learning Server Big Sur: 8 NVIDIA M40 https://techcrunch.com/gallery/a-look-inside-facebooks-data-center Talents: DNNresearch acquired By Google Geoffrey Hinton (right: Professor) Alex Krizhevsky (middle; PhD student), and Ilya Sutskever (left; Postdoc) Deep Learning: Convolutional Neural
    [Show full text]
  • WHEN SMART MACHINES ARE BIASED Olga Russakovsky Is Working to Change That
    SEXUAL-MISCONDUCT TRANSFER STUDENTS PRINCETON’S RECOMMENDATIONS ADMITTED DIGITAL ARCHIVES PRINCETON ALUMNI WEEKLY WHEN SMART MACHINES ARE BIASED Olga Russakovsky is working to change that JUNE 6, 2018 PAW.PRINCETON.EDU 00paw0606-coverFINALrev1.indd 1 5/22/18 10:23 AM That moment you finally get to go through the gates and see the next horizon. Communications of UNFORGETTABLE Office PRINCETON Your support makes it possible. This year’s Annual Giving campaign ends on June 30, 2018. To contribute by credit card, or for more information please call the gift line at 800-258-5421 (outside the US, 609-258-3373), or visit www.princeton.edu/ag. June 6, 2018 Volume 118, Number 14 An editorially independent magazine by alumni for alumni since 1900 PRESIDENT’S PAGE 2 INBOX 3 ON THE CAMPUS 5 Sexual-misconduct recommendations Transfer students Maya Lin visit Class Close-up: Video games deal with climate change A Day With ... dining hall coordinator Manuel Gomez Castaño ’20 Frank Stella ’58 exhibition at PUAM SPORTS: International rowers Championship for women’s crew The Big Three LIFE OF THE MIND 17 Author and professor Yiyun Li Rick Barton on peacemaking PRINCETONIANS 29 Mallika Ahluwalia ’05 Q&A: Yolanda Pierce ’94, first woman to lead Howard University divinity school Tiger Caucus in Congress CLASS NOTES 32 Maya Lin, page 7 Wikipedia Bias and Artificial Intelligence 20 Born Digital 24 ’20; MEMORIALS 49 Olga Russakovsky draws on personal With the decline of paper records, the staff at CLASSIFIEDS 54 experience and technical expertise to help Mudd Library is ensuring that digital records make artificial intelligence smarter.
    [Show full text]
  • Masters Thesis: Development of a Deep Learning Model for 3D Human Pose Estimation in Monocular Videos
    Development of a Deep Learning Model for 3D Human Pose Estimation in Monocular Videos Agnė Grinciūnaitė Master’s Degree Thesis VILNIUS GEDIMINAS TECHNICAL UNIVERSITY Faculty of Fundamental Sciences Department of Graphical Systems Agnė Grinciūnaitė Development of a Deep Learning Model for 3D Human Pose Estimation in Monocular Videos Master’s degree Thesis Information Technologies study programme, state code 621E14004 Multimedia Information Systems specialization Informatics Engineering study field Vilnius, 2016 The work in this thesis was supported by Vicar Vision. Their cooperation is hereby grate- fully acknowledged. Copyright © Department of Graphical Systems All rights reserved. VILNIUS GEDIMINAS TECHNICAL UNIVERSITY Faculty of Fundamental Sciences Department of Graphical Systems APPROVED BY Head of Department (Signature) (Name, Surname) (Date) Agnė Grinciūnaitė Development of a Deep Learning Model for 3D Human Pose Estimation in Monocular Videos Master’s degree Thesis Information Technologies study programme, state code 621E14004 Multimedia Information Systems specialization Informatics Engineering study field Supervisor (Title, Name, Surname) (Signature) (Date) Consultant (Title, Name, Surname) (Signature) (Date) Consultant (Title, Name, Surname) (Signature) (Date) Vilnius, 2016 Abstract There exists a visual system which can easily recognize and track human body position, movements and actions without any additional sensing. This system has the processor called brain and it is competent after being trained for some months. With a little bit more training it is also able to apply acquired skills for more complicated tasks such as understanding inter-personal attitudes, intentions and emotional states of the observed moving person. This system is called a human being and is so far the most inspirational piece of art for today’s artificial intelligence creators.
    [Show full text]
  • Computer Vision and Its Implications Based on the Paper Imagenet Classification with Deep Convolutional Neural Networks by Alex Krizhevsky, Ilya Sutskever, Geoffrey E
    Universität Leipzig Fakultät für Mathematik und Informatik Institut für Informatik Problem-seminar Deep-Learning Summary about Computer Vision and its implications based on the paper ImageNet Classification with Deep Convolutional Neural Networks by Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton Leipzig, WS 2017/18 vorgelegt von Christian Stur Betreuer: Ziad Sehili Contents 1.0. Introduction ..................................................................................................................... 2 1.1. Present significance ........................................................................................................ 2 1.2. Motivation of Computer Vision ...................................................................................... 2 2.0. Computer Vision ............................................................................................................. 3 2.1. Neural Networks ............................................................................................................. 3 2.2. Large Scale Visual Recognition Challenge (ILSVRC) .................................................. 4 2.2.1. Goal .......................................................................................................................... 4 2.2.2. Challenge Progression ............................................................................................. 4 2.2.3. The Data ................................................................................................................... 6 2.2.4. Issues
    [Show full text]
  • A Study of Face Obfuscation in Imagenet
    A Study of Face Obfuscation in ImageNet Kaiyu Yang 1 Jacqueline Yau 2 Li Fei-Fei 2 Jia Deng 1 Olga Russakovsky 1 Abstract devices taking photos (Butler et al., 2015; Dai et al., 2015). Face obfuscation (blurring, mosaicing, etc.) has Learning from the visual data has led to computer vision been shown to be effective for privacy protection; applications that promote the common good, e.g., better nevertheless, object recognition research typically traffic management (Malhi et al., 2011) and law enforce- assumes access to complete, unobfuscated images. ment (Sajjad et al., 2020). However, it also raises privacy In this paper, we explore the effects of face concerns, as images may capture sensitive information such obfuscation on the popular ImageNet challenge as faces, addresses, and credit cards (Orekondy et al., 2018). visual recognition benchmark. Most categories in Preventing unauthorized access to sensitive information in the ImageNet challenge are not people categories; private datasets has been extensively studied (Fredrikson however, many incidental people appear in the et al., 2015; Shokri et al., 2017). However, are publicly images, and their privacy is a concern. We first available datasets free of privacy concerns? Taking the annotate faces in the dataset. Then we demon- popular ImageNet dataset (Deng et al., 2009) as an example, strate that face blurring—a typical obfuscation there are only 3 people categories1 in the 1000 categories technique—has minimal impact on the accuracy of ImageNet Large Scale Visual Recognition Challenge of recognition models. Concretely, we benchmark (ILSVRC) (Russakovsky et al., 2015); nevertheless, the multiple deep neural networks on face-blurred dataset exposes many people co-occurring with other objects images and observe that the overall recognition in images (Prabhu & Birhane, 2021), e.g., people sitting accuracy drops only slightly (≤ 0:68%).
    [Show full text]
  • On Imagenet Roulette and Machine Vision
    Humans Categorise Humans: on ImageNet Roulette and Machine Vision Olga Goriunova Published in Donaufestival: Redefining Arts Catalogue, April 2020. In September 2019, artist Trevor Paglen and researcher Kate Crawford were getting through the final week of their ImageNet Roulette being publicly available on the web. ImageNet Roulette is part of their exhibition “Training Humans,” which ran between September 2019 and February 2020 at the Foundazione Prada. With this exhibition, collaborators Crawford and Paglen queried the collections of images and the processes used to train a wide range of machine learning algorithms to recognise and label images (known as “machine vision”).1 Crawford and Paglen want to draw attention to how computers see and categorize the world: seeing and recognizing is not neutral for humans (one need only think about “seeing” race or gender), and the same goes for machines. Machine vision, by now a large and successful part of AI, has only taken off in the last decade. For “machines” to recognize images, they need to be “trained” to do so. The first major step for this to happen is to have a freely available and pre-labelled very large image-data-set. The first and largest training dataset that Crawford and Paglen focused on and which they derived the title of their project from is ImageNet, which is only ten years old. ImageNet Roulette is a website or an app that allows one to take a selfie and run it through ImageNet. It uses an “open-source Caffe deep-learning framework … trained on the images and labels in the ‘person’ categories.”2 One is “seen” and “recognized” or labelled.
    [Show full text]