Deepfakes Generation and Detection: State-Of-The-Art, Open Challenges, Countermeasures, and Way Forward
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Artificial Intelligence in Health Care: the Hope, the Hype, the Promise, the Peril
Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril Michael Matheny, Sonoo Thadaney Israni, Mahnoor Ahmed, and Danielle Whicher, Editors WASHINGTON, DC NAM.EDU PREPUBLICATION COPY - Uncorrected Proofs NATIONAL ACADEMY OF MEDICINE • 500 Fifth Street, NW • WASHINGTON, DC 20001 NOTICE: This publication has undergone peer review according to procedures established by the National Academy of Medicine (NAM). Publication by the NAM worthy of public attention, but does not constitute endorsement of conclusions and recommendationssignifies that it is the by productthe NAM. of The a carefully views presented considered in processthis publication and is a contributionare those of individual contributors and do not represent formal consensus positions of the authors’ organizations; the NAM; or the National Academies of Sciences, Engineering, and Medicine. Library of Congress Cataloging-in-Publication Data to Come Copyright 2019 by the National Academy of Sciences. All rights reserved. Printed in the United States of America. Suggested citation: Matheny, M., S. Thadaney Israni, M. Ahmed, and D. Whicher, Editors. 2019. Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril. NAM Special Publication. Washington, DC: National Academy of Medicine. PREPUBLICATION COPY - Uncorrected Proofs “Knowing is not enough; we must apply. Willing is not enough; we must do.” --GOETHE PREPUBLICATION COPY - Uncorrected Proofs ABOUT THE NATIONAL ACADEMY OF MEDICINE The National Academy of Medicine is one of three Academies constituting the Nation- al Academies of Sciences, Engineering, and Medicine (the National Academies). The Na- tional Academies provide independent, objective analysis and advice to the nation and conduct other activities to solve complex problems and inform public policy decisions. -
Real Vs Fake Faces: Deepfakes and Face Morphing
Graduate Theses, Dissertations, and Problem Reports 2021 Real vs Fake Faces: DeepFakes and Face Morphing Jacob L. Dameron WVU, [email protected] Follow this and additional works at: https://researchrepository.wvu.edu/etd Part of the Signal Processing Commons Recommended Citation Dameron, Jacob L., "Real vs Fake Faces: DeepFakes and Face Morphing" (2021). Graduate Theses, Dissertations, and Problem Reports. 8059. https://researchrepository.wvu.edu/etd/8059 This Thesis is protected by copyright and/or related rights. It has been brought to you by the The Research Repository @ WVU with permission from the rights-holder(s). You are free to use this Thesis in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you must obtain permission from the rights-holder(s) directly, unless additional rights are indicated by a Creative Commons license in the record and/ or on the work itself. This Thesis has been accepted for inclusion in WVU Graduate Theses, Dissertations, and Problem Reports collection by an authorized administrator of The Research Repository @ WVU. For more information, please contact [email protected]. Real vs Fake Faces: DeepFakes and Face Morphing Jacob Dameron Thesis submitted to the Benjamin M. Statler College of Engineering and Mineral Resources at West Virginia University in partial fulfillment of the requirements for the degree of Master of Science in Electrical Engineering Xin Li, Ph.D., Chair Natalia Schmid, Ph.D. Matthew Valenti, Ph.D. Lane Department of Computer Science and Electrical Engineering Morgantown, West Virginia 2021 Keywords: DeepFakes, Face Morphing, Face Recognition, Facial Action Units, Generative Adversarial Networks, Image Processing, Classification. -
Synthetic Video Generation
Synthetic Video Generation Why seeing should not always be believing! Alex Adam Image source https://www.pocket-lint.com/apps/news/adobe/140252-30-famous-photoshopped-and-doctored-images-from-across-the-ages Image source https://www.pocket-lint.com/apps/news/adobe/140252-30-famous-photoshopped-and-doctored-images-from-across-the-ages Image source https://www.pocket-lint.com/apps/news/adobe/140252-30-famous-photoshopped-and-doctored-images-from-across-the-ages Image source https://www.pocket-lint.com/apps/news/adobe/140252-30-famous-photoshopped-and-doctored-images-from-across-the-ages Image Tampering Historically, manipulated Off the shelf software (e.g imagery has deceived Photoshop) exists to do people this now Has become standard in Public have become tabloids/social media somewhat numb to it - it’s no longer as impactful/shocking How does machine learning fit in? Advent of machine learning Video manipulation is now has made image also tractable with enough manipulation even easier data and compute Can make good synthetic Public are largely unaware of videos using a gaming this and the danger it poses! computer in a bedroom Part I: Faceswap ● In 2017, reddit (/u/deepfakes) posted Python code that uses machine learning to swap faces in images/video ● ‘Deepfake’ content flooded reddit, YouTube and adult websites ● Reddit since banned this content (but variants of the code are open source https://github.com/deepfakes/faceswap) ● Autoencoder concepts underlie most ‘Deepfake’ methods Faceswap Algorithm Image source https://medium.com/@jonathan_hui/how-deep-learning-fakes-videos-deepfakes-and-how-to-detect-it-c0b50fbf7cb9 Inference Image source https://medium.com/@jonathan_hui/how-deep-learning-fakes-videos-deepfakes-and-how-to-detect-it-c0b50fbf7cb9 ● Faceswap model is an autoencoder. -
Proposed Framework for Digital Video Authentication
PROPOSED FRAMEWORK FOR DIGITAL VIDEO AUTHENTICATION by GREGORY SCOTT WALES A.S., Community College of the Air Force, 1990 B.S., Champlain College, 2012 M.S., Champlain College, 2015 A thesis submitted to the Faculty of the Graduate School of the University of Colorado in partial fulfillment of the requirements for the degree of Master of Science Recording Arts 2019 © 2019 GREGORY SCOTT WALES ALL RIGHTS RESERVED ii This thesis for the Master of Science degree by Gregory Scott Wales has been approved for the Recording Arts Program by Catalin Grigoras, Chair Jeffrey M. Smith Marcus Rogers Date: May 18, 2019 iii Wales, Gregory Scott (M.S., Recording Arts Program) Proposed Framework for Digital Video Authentication Thesis directed by Associate Professor Catalin Grigoras. ABSTRACT One simply has to look at news outlets or social media to see our society video records events from births to deaths and everything in between. A trial court’s acceptance of videos supporting administrative hearings, civil litigation, and criminal cases is based on a foundation that the videos offered into evidence are authentic; however, technological advancements in video editing capabilities provide an easy method to edit digital videos. The proposed framework offers a structured approach to evaluate and incorporate methods, existing and new, that come from scientific research and publication. The thesis offers a quick overview of digital video file creation chain (including factors that influence the final file), general description of the digital video file container, and description of camera sensor noises. The thesis addresses the overall development and proposed use of the framework, previous research of analysis methods / techniques, testing of the methods / techniques, and an overview of the testing results. -
Exposing Deepfake Videos by Detecting Face Warping Artifacts
Exposing DeepFake Videos By Detecting Face Warping Artifacts Yuezun Li, Siwei Lyu Computer Science Department University at Albany, State University of New York, USA Abstract sibility to large-volume training data and high-throughput computing power, but more to the growth of machine learn- In this work, we describe a new deep learning based ing and computer vision techniques that eliminate the need method that can effectively distinguish AI-generated fake for manual editing steps. videos (referred to as DeepFake videos hereafter) from real In particular, a new vein of AI-based fake video gen- videos. Our method is based on the observations that cur- eration methods known as DeepFake has attracted a lot rent DeepFake algorithm can only generate images of lim- of attention recently. It takes as input a video of a spe- ited resolutions, which need to be further warped to match cific individual (’target’), and outputs another video with the original faces in the source video. Such transforms leave the target’s faces replaced with those of another individ- distinctive artifacts in the resulting DeepFake videos, and ual (’source’). The backbone of DeepFake are deep neu- we show that they can be effectively captured by convo- ral networks trained on face images to automatically map lutional neural networks (CNNs). Compared to previous the facial expressions of the source to the target. With methods which use a large amount of real and DeepFake proper post-processing, the resulting videos can achieve a generated images to train CNN classifier, our method does high level of realism. not need DeepFake generated images as negative training In this paper, we describe a new deep learning based examples since we target the artifacts in affine face warp- method that can effectively distinguish DeepFake videos ing as the distinctive feature to distinguish real and fake from the real ones. -
JCS Deepfake
Don’t Believe Your Eyes (Or Ears): The Weaponization of Artificial Intelligence, Machine Learning, and Deepfakes Joe Littell 1 Agenda ØIntroduction ØWhat is A.I.? ØWhat is a DeepFake? ØHow is a DeepFake created? ØVisual Manipulation ØAudio Manipulation ØForgery ØData Poisoning ØConclusion ØQuestions 2 Introduction Deniss Metsavas, an Estonian soldier convicted of spying for Russia’s military intelligence service after being framed for a rape in Russia. (Picture from Daniel Lombroso / The Atlantic) 3 What is A.I.? …and what is it not? General Artificial Intelligence (AI) • Machine (or Statistical) Learning (ML) is a subset of AI • ML works through the probability of a new event happening based on previously gained knowledge (Scalable pattern recognition) • ML can be supervised, leaning requiring human input into the data, or unsupervised, requiring no input to the raw data. 4 What is a Deepfake? • Deepfake is a mash up of the words for deep learning, meaning machine learning using a neural network, and fake images/video/audio. § Taken from a Reddit user name who utilized faceswap app for his own ‘productions.’ • Created by the use of two machine learning algorithms, Generative Adversarial Networks, and Auto-Encoders. • Became known for the use in underground pornography using celebrity faces in highly explicit videos. 5 How is a Deepfake created? • Deepfakes are generated using Generative Adversarial Networks, and Auto-Encoders. • These algorithms work through the uses of competing systems, where one creates a fake piece of data and the other is trained to determine if that datatype is fake or not • Think of it like a counterfeiter and a police officer. -
Arxiv:1907.05276V1 [Cs.CV] 6 Jul 2019 Media Online Are Rapidly Outpacing the Ability to Manually Detect and Respond to Such Media
Human detection of machine manipulated media Matthew Groh1, Ziv Epstein1, Nick Obradovich1,2, Manuel Cebrian∗1,2, and Iyad Rahwan∗1,2 1Media Laboratory, Massachusetts Institute of Technology, Cambridge, Massachusetts, USA 2Center for Humans & Machines, Max Planck Institute for Human Development, Berlin, Germany Abstract Recent advances in neural networks for content generation enable artificial intelligence (AI) models to generate high-quality media manipulations. Here we report on a randomized experiment designed to study the effect of exposure to media manipulations on over 15,000 individuals’ ability to discern machine-manipulated media. We engineer a neural network to plausibly and automatically remove objects from images, and we deploy this neural network online with a randomized experiment where participants can guess which image out of a pair of images has been manipulated. The system provides participants feedback on the accuracy of each guess. In the experiment, we randomize the order in which images are presented, allowing causal identification of the learning curve surrounding participants’ ability to detect fake content. We find sizable and robust evidence that individuals learn to detect fake content through exposure to manipulated media when provided iterative feedback on their detection attempts. Over a succession of only ten images, participants increase their rating accuracy by over ten percentage points. Our study provides initial evidence that human ability to detect fake, machine-generated content may increase alongside the prevalence of such media online. Introduction The recent emergence of artificial intelligence (AI) powered media manipulations has widespread societal implications for journalism and democracy [1], national security [2], and art [3]. AI models have the potential to scale misinformation to unprecedented levels by creating various forms of synthetic media [4]. -
Deepfakes 2020 the Tipping Point, Sentinel
SENTINEL DEEPFAKES 2020: THE TIPPING POINT The Current Threat Landscape, its Impact on the U.S 2020 Elections, and the Coming of AI-Generated Events at Scale. Sentinel - 2020 1 About Sentinel. Sentinel works with governments, international media outlets and defense agencies to help protect democracies from disinformation campaigns, synthetic media and information operations by developing a state-of-the-art AI detection platform. Headquartered in Tallinn, Estonia, the company was founded by ex-NATO AI and cybersecurity experts, and is backed by world-class investors including Jaan Tallinn (Co-Founder of Skype & early investor in DeepMind) and Taavet Hinrikus (Co-Founder of TransferWise). Our vision is to become the trust layer for the Internet by verifying the entire critical information supply chain and safeguard 1 billion people from information warfare. Acknowledgements We would like to thank our investors, partners, and advisors who have helped us throughout our journey and share our vision to build a trust layer for the internet. Special thanks to Mikk Vainik of Republic of Estonia’s Ministry of Economic Affairs and Communications, Elis Tootsman of Accelerate Estonia, and Dr. Adrian Venables of TalTech for your feedback and support as well as to Jaan Tallinn, Taavet Hinrikus, Ragnar Sass, United Angels VC, Martin Henk, and everyone else who has made this report possible. Johannes Tammekänd CEO & Co-Founder © 2020 Sentinel Contact: [email protected] Authors: Johannes Tammekänd, John Thomas, and Kristjan Peterson Cite: Deepfakes 2020: The Tipping Point, Johannes Tammekänd, John Thomas, and Kristjan Peterson, October 2020 Sentinel - 2020 2 Executive Summary. “There are but two powers in the world, the sword and the mind. -
Editorial Algorithms
March 26, 2021 Editorial Algorithms The next AI frontier in News and Information National Press Club Journalism Institute Housekeeping - You will receive slides after the presentation - Interactive polls throughout the session - Use QA tool on Zoom to ask questions - If you still have questions, reach out to Francesco directly [email protected] or connect on twitter @fpmarconi What will your learn during today’s session 1. How AI is being used and will be used by journalists and communicators, and what that means for how we work 2. The ethics of AI, particularly related to diversity, equity and representation 3. The role of AI in the spread of misinformation and ‘fake news’ 4. How journalists and communications professionals can avoid pitfalls while taking advantage of the possibilities AI provides to develop new ways of telling stories and connecting with audiences Poll time! What’s your overall feeling about AI and news? A. It’s the future of news and communications! B. I’m cautiously optimistic C. I’m cautiously pessimistic D. It’s bad news for journalism Journalism was designed to solve information scarcity Newspapers were established to provide increasingly literate audiences with information. Journalism was originally designed to solve the problem that information was not widely available. The news industry went through a process of standardization so it could reliably source and produce information more efficiently. This led to the development of very structured ways of writing such as the “inverted pyramid” method, to organizing workflows around beats and to plan coverage based on news cycles. A new reality: information abundance The flood of information that humanity is now exposed to presents a new challenge that cannot be solved exclusively with the traditional methods of journalism. -
CNN-Generated Images Are Surprisingly Easy to Spot... for Now
CNN-generated images are surprisingly easy to spot... for now Sheng-Yu Wang1 Oliver Wang2 Richard Zhang2 Andrew Owens1,3 Alexei A. Efros1 UC Berkeley1 Adobe Research2 University of Michigan3 synthetic real ProGAN [19] StyleGAN [20] BigGAN [7] CycleGAN [48] StarGAN [10] GauGAN [29] CRN [9] IMLE [23] SITD [8] Super-res. [13] Deepfakes [33] Figure 1: Are CNN-generated images hard to distinguish from real images? We show that a classifier trained to detect images generated by only one CNN (ProGAN, far left) can detect those generated by many other models (remaining columns). Our code and models are available at https://peterwang512.github.io/CNNDetection/. Abstract are fake [14]. This issue has started to play a significant role in global politics; in one case a video of the president of In this work we ask whether it is possible to create Gabon that was claimed by opposition to be fake was one a “universal” detector for telling apart real images from factor leading to a failed coup d’etat∗. Much of this con- these generated by a CNN, regardless of architecture or cern has been directed at specific manipulation techniques, dataset used. To test this, we collect a dataset consisting such as “deepfake”-style face replacement [2], and photo- of fake images generated by 11 different CNN-based im- realistic synthetic humans [20]. However, these methods age generator models, chosen to span the space of com- represent only two instances of a broader set of techniques: monly used architectures today (ProGAN, StyleGAN, Big- image synthesis via convolutional neural networks (CNNs). -
OC-Fakedect: Classifying Deepfakes Using One-Class Variational Autoencoder
OC-FakeDect: Classifying Deepfakes Using One-class Variational Autoencoder Hasam Khalid Simon S. Woo Computer Science and Engineering Department Computer Science and Engineering Department Sungkyunkwan University, South Korea Sungkyunkwan University, South Korea [email protected] [email protected] Abstract single facial image to create fake images or videos. One popular example is the Deepfakes of former U.S. President, An image forgery method called Deepfakes can cause Barack Obama, generated as part of a research [29] focusing security and privacy issues by changing the identity of on the synthesis of a high-quality video, featuring Barack a person in a photo through the replacement of his/her Obama speaking with accurate lip sync, composited into face with a computer-generated image or another person’s a target video clip. Therefore, the ability to easily forge face. Therefore, a new challenge of detecting Deepfakes videos raises serious security and privacy concerns: imag- arises to protect individuals from potential misuses. Many ine hackers that can use deepfakes to present a forged video researchers have proposed various binary-classification of an eminent person to send out false and potentially dan- based detection approaches to detect deepfakes. How- gerous messages to the public. Nowadays, fake news has ever, binary-classification based methods generally require become an issue as well, due to the spread of misleading in- a large amount of both real and fake face images for train- formation via traditional news media or online social media ing, and it is challenging to collect sufficient fake images and Deepfake videos can be combined to create arbitrary data in advance. -
Deepfaked Online Content Is Highly Effective in Manipulating People's
Deepfaked online content is highly effective in manipulating people’s attitudes and intentions Sean Hughes, Ohad Fried, Melissa Ferguson, Ciaran Hughes, Rian Hughes, Xinwei Yao, & Ian Hussey In recent times, disinformation has spread rapidly through social media and news sites, biasing our (moral) judgements of other people and groups. “Deepfakes”, a new type of AI-generated media, represent a powerful new tool for spreading disinformation online. Although Deepfaked images, videos, and audio may appear genuine, they are actually hyper-realistic fabrications that enable one to digitally control what another person says or does. Given the recent emergence of this technology, we set out to examine the psychological impact of Deepfaked online content on viewers. Across seven preregistered studies (N = 2558) we exposed participants to either genuine or Deepfaked content, and then measured its impact on their explicit (self-reported) and implicit (unintentional) attitudes as well as behavioral intentions. Results indicated that Deepfaked videos and audio have a strong psychological impact on the viewer, and are just as effective in biasing their attitudes and intentions as genuine content. Many people are unaware that Deepfaking is possible; find it difficult to detect when they are being exposed to it; and most importantly, neither awareness nor detection serves to protect people from its influence. All preregistrations, data and code available at osf.io/f6ajb. The proliferation of social media, dating apps, news and copy of a person that can be manipulated into doing or gossip sites, has brought with it the ability to learn saying anything (2). about a person’s moral character without ever having Deepfaking has quickly become a tool of to interact with them in real life.