DISCRIMINATING SYSTEMS Gender, Race, and Power in AI
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Reddit Writting Prompts Always Late Until on Time
Reddit Writting Prompts Always Late Until On Time Candescent Shayne usually reimbursing some insatiety or ditch meditatively. Sloane fuzzes lanceolately. Wasteful Dwayne telescope some admonitions and refuel his opinionativeness so swith! Otas more individualistic than not believing their subordinates and frightening themes should ask someone close despite the time on reddit is necessary cookies are given a crown I started getting cancer as he is always on his phone and transfer go so whatever he. Sleep debt grows when people sleep accumulates over time. Story 2 Matt Kruk Story 3 Jorge Abeita Story 4 Anon Jan 30 2020 Dark. Are always late too late now stop the time, prompts can only showed people with private mediator to teasing really funny, reddit writting prompts always late until on time the service shall indicate the. A person infected with Ebola cannot spread that disease and they develop symptoms. Only time of late to always documented in texas real estate related subjects, reddit writting prompts always late until on time the truth waiting for this picture story based on? What Happened After My 13-Year-Old Son Joined the Alt-Right. Prepare Your COVID-19 Coronavirus Co-Parenting Plan. Does my children of any heating, reddit writting prompts always late until on time stop to prevent the commission may act as or a dialog and i never liked this. A Few Ideas for Dealing with civilian Work Cult of Pedagogy. The market focuses on user privacy prompt responses to customer audience and. It happened time and resign and was instrumental at stopping the. -
When Diving Deep Into Operationalizing the Ethical Development of Artificial Intelligence, One Immediately Runs Into the “Fractal Problem”
When diving deep into operationalizing the ethical development of artificial intelligence, one immediately runs into the “fractal problem”. This may also be called the “infinite onion” problem.1 That is, each problem within development that you pinpoint expands into a vast universe of new complex problems. It can be hard to make any measurable progress as you run in circles among different competing issues, but one of the paths forward is to pause at a specific point and detail what you see there. Here I list some of the complex points that I see at play in the firing of Dr. Timnit Gebru, and why it will remain forever after a really, really, really terrible decision. The Punchline. The firing of Dr. Timnit Gebru is not okay, and the way it was done is not okay. It appears to stem from the same lack of foresight that is at the core of modern technology,2 and so itself serves as an example of the problem. The firing seems to have been fueled by the same underpinnings of racism and sexism that our AI systems, when in the wrong hands, tend to soak up. How Dr. Gebru was fired is not okay, what was said about it is not okay, and the environment leading up to it was -- and is -- not okay. Every moment where Jeff Dean and Megan Kacholia do not take responsibility for their actions is another moment where the company as a whole stands by silently as if to intentionally send the horrifying message that Dr. Gebru deserves to be treated this way. -
Bias and Fairness in NLP
Bias and Fairness in NLP Margaret Mitchell Kai-Wei Chang Vicente Ordóñez Román Google Brain UCLA University of Virginia Vinodkumar Prabhakaran Google Brain Tutorial Outline ● Part 1: Cognitive Biases / Data Biases / Bias laundering ● Part 2: Bias in NLP and Mitigation Approaches ● Part 3: Building Fair and Robust Representations for Vision and Language ● Part 4: Conclusion and Discussion “Bias Laundering” Cognitive Biases, Data Biases, and ML Vinodkumar Prabhakaran Margaret Mitchell Google Brain Google Brain Andrew Emily Simone Parker Lucy Ben Elena Deb Timnit Gebru Zaldivar Denton Wu Barnes Vasserman Hutchinson Spitzer Raji Adrian Brian Dirk Josh Alex Blake Hee Jung Hartwig Blaise Benton Zhang Hovy Lovejoy Beutel Lemoine Ryu Adam Agüera y Arcas What’s in this tutorial ● Motivation for Fairness research in NLP ● How and why NLP models may be unfair ● Various types of NLP fairness issues and mitigation approaches ● What can/should we do? What’s NOT in this tutorial ● Definitive answers to fairness/ethical questions ● Prescriptive solutions to fix ML/NLP (un)fairness What do you see? What do you see? ● Bananas What do you see? ● Bananas ● Stickers What do you see? ● Bananas ● Stickers ● Dole Bananas What do you see? ● Bananas ● Stickers ● Dole Bananas ● Bananas at a store What do you see? ● Bananas ● Stickers ● Dole Bananas ● Bananas at a store ● Bananas on shelves What do you see? ● Bananas ● Stickers ● Dole Bananas ● Bananas at a store ● Bananas on shelves ● Bunches of bananas What do you see? ● Bananas ● Stickers ● Dole Bananas ● Bananas -
Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification∗
Proceedings of Machine Learning Research 81:1{15, 2018 Conference on Fairness, Accountability, and Transparency Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification∗ Joy Buolamwini [email protected] MIT Media Lab 75 Amherst St. Cambridge, MA 02139 Timnit Gebru [email protected] Microsoft Research 641 Avenue of the Americas, New York, NY 10011 Editors: Sorelle A. Friedler and Christo Wilson Abstract who is hired, fired, granted a loan, or how long Recent studies demonstrate that machine an individual spends in prison, decisions that learning algorithms can discriminate based have traditionally been performed by humans are on classes like race and gender. In this rapidly made by algorithms (O'Neil, 2017; Citron work, we present an approach to evaluate and Pasquale, 2014). Even AI-based technologies bias present in automated facial analysis al- that are not specifically trained to perform high- gorithms and datasets with respect to phe- stakes tasks (such as determining how long some- notypic subgroups. Using the dermatolo- one spends in prison) can be used in a pipeline gist approved Fitzpatrick Skin Type clas- that performs such tasks. For example, while sification system, we characterize the gen- face recognition software by itself should not be der and skin type distribution of two facial analysis benchmarks, IJB-A and Adience. trained to determine the fate of an individual in We find that these datasets are overwhelm- the criminal justice system, it is very likely that ingly composed of lighter-skinned subjects such software is used to identify suspects. Thus, (79:6% for IJB-A and 86:2% for Adience) an error in the output of a face recognition algo- and introduce a new facial analysis dataset rithm used as input for other tasks can have se- which is balanced by gender and skin type. -
Extinction Rebellion: Unabombers Without Bombs Page 5
EXTINCTION REBELLION: UNABOMBERS WITHOUT BOMBS PAGE 5 ALSO IN THIS ISSUE: The Great Ka-Pao! Silicon The Proteus Fund, Green Pension Valley’s Strike Against “Dark Money” 11 Hold-Up 20 Conservatives 37 Shapeshifter www.CapitalResearch.org Want to know more about the donors, foundations, nonprofits, activists, and others working to influence public policy? Visit: INFLUENCE WATCH.ORG Launched by Capital Research Center in August 2017, InfluenceWatch w i l l bring unprecedented transparency to the history, motives, and i n t e r conn ect ions o f all entities invo lve d in the advocacy m o v ement. Today, our growing website includes over 6,800 pages and over 1,200 full profiles, with more added each week. L ear n mor e a t In f l uen c e Wa t c h .o r g ISSUE 6, 2019 CONTENTS GREEN WATCH 3 Extinction Rebellion: COMMENTARY 5 Unabombers Without Bombs Donor Advised By Ken Braun “Dark Money”? By Hayden Ludwig LABOR WATCH The Great Green 11 Pension Hold-Up By Neil Meghami Capital Research is a monthly publication of the Capital Research Center (CRC), a nonpartisan education and research organization, classified by the IRS as a 501(c)(3) public charity. CRC is an independent, tax-exempt DECEPTION & MISDIRECTION institution governed by an independent Ka-Pao! Silicon Valley’s board of trustees. We rely on private 20 financial support from the general Strike Against Conservatives public—individuals, foundations, By David Hogberg and corporations—for our income. We accept no government funds and perform no contract work. -
Startup Essentials
DOWNLOADFREE Startup Essentials A Curated Collection of Chapters from the O'Reilly Business Library Startup Essentials A Curated Collection of Chapters from the O’Reilly Business Library If you’re considering a software or hardware startup venture, this collection provides valuable excerpts from four existing and forthcoming O’Reilly books on the subject. Sample chapters from Hello, Startup, Hot Seat, The Hardware Startup, and Managing Startups feature stories, case studies, and salient advice from companies and experts that have blazed trails from successful launch to market acceptance. These four books also take you through the many dilemmas and points of failure possible with product development, funding, and startup management along the way. You’ll dive into sample chapters from: Hello, Startup—the tutorial for building products, technologies, and teams, based on interviews with programmers from successful startups. Featured chapter: “Why Startups.” Hot Seat: The Startup CEO Guidebook—four-time founder/CEO Dan Shapiro tells stories of startups that have survived and thrived by the advice in this book. Featured chapter: “The Cofounder Dilemma.” The Hardware Startup—two-dozen case studies of real-world startups illustrate successes and failures at every stage of the process. Featured chapters: “Knowing Your Market” and “Branding.” Managing Startups—the best posts from many blogs on technology startup management, compiled by Harvard Business School Professor Tom Eisenmann. Featured chapter: “Why I Left Consulting and Joined a Startup.” Hello, Startup A Programmer’s Guide to Building Products, Technologies, and Teams Available in Early Release Chapter 1. Why Startups Hot Seat The Startup CEO Guidebook Available Now Chapter 2. -
AI Principles 2020 Progress Update
AI Principles 2020 Progress update AI Principles 2020 Progress update Table of contents Overview ........................................................................................................................................... 2 Culture, education, and participation ....................................................................4 Technical progress ................................................................................................................... 5 Internal processes ..................................................................................................................... 8 Community outreach and exchange .....................................................................12 Conclusion .....................................................................................................................................17 Appendix: Research publications and tools ....................................................18 Endnotes .........................................................................................................................................20 1 AI Principles 2020 Progress update Overview Google’s AI Principles were published in June 2018 as a charter to guide how we develop AI responsibly and the types of applications we will pursue. This report highlights recent progress in AI Principles implementation across Google, including technical tools, educational programs and governance processes. Of particular note in 2020, the AI Principles have supported our ongoing work to address -
Facial Recognition Is Improving — but the Bigger That Had Been Scanned
Feature KELVIN CHAN/AP/SHUTTERSTOCK KELVIN The Metropolitan Police in London used facial-recognition cameras to scan for wanted people in February. Fussey and Murray listed a number of ethical and privacy concerns with the dragnet, and questioned whether it was legal at all. And they queried the accuracy of the system, which is sold by Tokyo-based technology giant NEC. The software flagged 42 people over 6 trials BEATING that the researchers analysed; officers dis- missed 16 matches as ‘non-credible’ but rushed out to stop the others. They lost 4 people in the crowd, but still stopped 22: only 8 turned out to be correct matches. The police saw the issue differently. They BIOMETRIC BIAS said the system’s number of false positives was tiny, considering the many thousands of faces Facial recognition is improving — but the bigger that had been scanned. (They didn’t reply to issue is how it’s used. By Davide Castelvecchi Nature’s requests for comment for this article.) The accuracy of facial recognition has improved drastically since ‘deep learning’ tech- hen London’s Metropolitan extracted key features and compared them niques were introduced into the field about a Police tested real-time to those of suspects from a watch list. “If there decade ago. But whether that means it’s good facial-recognition technology is a match, it pulls an image from the live feed, enough to be used on lower-quality, ‘in the wild’ between 2016 and 2019, they together with the image from the watch list.” images is a hugely controversial issue. -
Distributed Blackness Published Just Last Year (2020, New York University Press), Dr
Transcript The Critical Technology Episode 4: Distributed Blackness Published just last year (2020, New York University Press), Dr. Andre Brock Jr’s Distributed Blackness: African American Cybercultures has already become a highly influential, widely cited, must read text for anyone interested in digital culture and technology. Through a theoretically rich and methodologically innovative analysis, Brock positions Blackness at the centre of Internet technoculture, its technological designs and infrastructures, its political economic underpinnings, its cultural significance and its emergent informational identities. Sara Grimes 0:00 Like many of you, as the fall 2020 semester was wrapping up, my Twitter feed was suddenly flooded with the news that Dr. Timnit Gebru had been fired from Google. A world renowned computer scientist, leading AI ethics researcher, and co-founder of Black in AI. Dr. Gebru had recently completed research raising numerous ethical concerns with the type of large scale AI language models used in many Google products. Her expulsion from Google highlights the discrimination and precarity faced by black people, especially black women, working in the tech industries and in society at large. On the one hand, Dr. Gebru’s treatment by Google execs made visible through screenshots and conversations posted on Twitter was met with an outpouring of public support from fellow Google employees, tech workers, academics and concerned citizens. The #ISupportTimnit was often paired with #BelieveBlackWomen, connecting this event with the systematic anti-black racism and sexism found across technoculture. On the other hand, since December, Dr. Gebru and her supporters have also been targeted by Twitter trolls, and subjected to a bombardment of racist and sexist vitriol online. -
Saving Face: Investigating the Ethical Concerns of Facial Recognition Auditing
Saving Face: Investigating the Ethical Concerns of Facial Recognition Auditing Inioluwa Deborah Raji Timnit Gebru Margaret Mitchell University of Toronto Google Research Google Research Canada United States United States [email protected] [email protected] [email protected] Joy Buolamwini Joonseok Lee Emily Denton MIT Media Lab. Google Research Google Research United States United States United States [email protected] [email protected] [email protected] ABSTRACT the positive uses of this technology includes claims of “understand- Although essential to revealing biased performance, well inten- ing users”, “monitoring or detecting human activity”, “indexing and tioned attempts at algorithmic auditing can have effects that may searching digital image libraries”, and verifying and identifying harm the very populations these measures are meant to protect. subjects “in security scenarios” [1, 8, 16, 33]. This concern is even more salient while auditing biometric sys- The reality of FPT deployments, however, reveals that many tems such as facial recognition, where the data is sensitive and of these uses are quite vulnerable to abuse, especially when used the technology is often used in ethically questionable manners. for surveillance and coupled with predatory data collection prac- We demonstrate a set of five ethical concerns in the particular case tices that, intentionally or unintentionally, discriminate against of auditing commercial facial processing technology, highlighting marginalized groups [2]. The stakes are particularly high when additional design considerations and ethical tensions the auditor we consider companies like Amazon and HireVue, who are selling needs to be aware of so as not exacerbate or complement the harms their services to police departments or using the technology to help propagated by the audited system. -
2019 Global Go to Think Tank Index Report
University of Pennsylvania ScholarlyCommons Think Tanks and Civil Societies Program TTCSP Global Go To Think Tank Index Reports (TTCSP) 6-18-2020 2019 Global Go To Think Tank Index Report James G. McGann University of Pennsylvania, [email protected] Follow this and additional works at: https://repository.upenn.edu/think_tanks Part of the International and Area Studies Commons McGann, James G., "2019 Global Go To Think Tank Index Report" (2020). TTCSP Global Go To Think Tank Index Reports. 17. https://repository.upenn.edu/think_tanks/17 2020 Copyright: All rights reserved. No part of this report may be reproduced or utilized in any form or by any means, electronic or mechanical, including photocopying, recording, or by an information storage or retrieval system, without written permission from the University of Pennsylvania, Think Tanks and Civil Societies Program. All requests, questions and comments should be sent to: James G. McGann, Ph.D. Senior Lecturer, International Studies Director Think Tanks and Civil Societies Program The Lauder Institute University of Pennsylvania Email: [email protected] This paper is posted at ScholarlyCommons. https://repository.upenn.edu/think_tanks/17 For more information, please contact [email protected]. 2019 Global Go To Think Tank Index Report Abstract The Think Tanks and Civil Societies Program (TTCSP) of the Lauder Institute at the University of Pennsylvania conducts research on the role policy institutes play in governments and civil societies around the world. Often referred to as the “think tanks’ think tank,” TTCSP examines the evolving role and character of public policy research organizations. Over the last 29 years, the TTCSP has developed and led a series of global initiatives that have helped bridge the gap between knowledge and policy in critical policy areas such as international peace and security, globalization and governance, international economics, environmental issues, information and society, poverty alleviation, and healthcare and global health. -
Organizing Big Tech by Sarah Jaffe Published by the Rosa Luxemburg Stiftung, New York Office, April 2021
Organizing Big Tech By Sarah Jaffe Published by the Rosa Luxemburg Stiftung, New York Office, April 2021 Executive Director: Andreas Günther Editor: Aaron Eisenberg and Maria Savel Address: 275 Madison Avenue, Suite 2114, New York, NY 10016 Email: [email protected] Phone: +1 (917) 409-1040 With support from the German Foreign Office. The Rosa Luxemburg Foundation is an internationally operating, progressive non-profit institution for civic education. In coop- eration with many organizations around the globe, it works on democratic and social participation, empowerment of disadvan- taged groups, alternatives for economic and social development, and peaceful conflict resolution. The New York Office serves two major tasks: to work around issues concerning the United Nations and to engage in dialogue with North American progressives in universities, unions, social movements, and politics. www.rosalux.nyc 2 Rosa Luxemburg Stiftung New York Office Organizing Big Tech By Sarah Jaffe A supporter of the RWDSU unionization effort takes a photo of the RWDSU union rep standing with other supporters outside the Amazon fulfillment ware- house at the center of a unionization drive on March 29, 2021 in Bessemer, Alabama. (Photo by Elijah Nouvelage/Getty Images) Introduction The “tech” sector occupies a central place in American capitalism in 2021, and for good reason. As Logic magazine editor Ben Tarnoff notes, “Tech is an oasis of profitability in an era of stagnation. For this reason, it also serves a valuable ideological function.”1 But what do companies like Google, Amazon, and Tesla have in common, really? One began as a search engine; another, an online bookstore; the third an electric car maker.