Use of Ai in Online Content Moderation
Total Page:16
File Type:pdf, Size:1020Kb
USE OF AI IN ONLINE CONTENT MODERATION 2019 REPORT PRODUCED ON BEHALF OF Foreword from Ofcom As the UK’s converged communications regulator, Ofcom oversees broadband and mobile telecoms, TV, radio, video-on-demand services, post and the airwaves used by wireless devices. We have a statutory duty to promote media literacy, under which we carry out research into people’s use of – and attitudes towards – various communications services, including online services such as social media and video sharing platforms. This remit reflects the increasing convergence of the companies and markets that we regulate. In recent years a wide-ranging, global debate has emerged around the risks faced by internet users, with a specific focus on protecting users from harmful content. A key element of this debate has centred on the role and capabilities of automated approaches (driven by Artificial Intelligence and Machine Learning techniques) to enhance the effectiveness of online content moderation and offer users greater protection from potentially harmful material. These approaches may have implications for people’s future use of – and attitudes towards – online communications services, and may be applicable more broadly to developing new techniques for moderating and cataloguing content in the broadcast and audiovisual media industries, as well as back-office support functions in the telecoms, media and postal sectors. Ofcom has commissioned Cambridge Consultants to produce this report as a contribution to the evidence base on people’s use of and attitudes towards online services, which helps enable wider debate on the risks faced by internet users. Cambridge Consultants has retained editorial responsibility for this report throughout the process, so the opinions stated and highlighted policy implications should not necessarily be taken to represent the views of Ofcom. REPORT USE OF AI IN ONLINE CONTENT MODERATION The advances of AI in recent years will continue, driven by EXECUTIVE SUMMARY commercial applications and enabled by continued progress in algorithm development, the increasing availability of low- cost computational power and the widespread collection In the last two decades, online platforms that permit users to of data. There are, however, some inhibitors to making the interact and upload content for others to view have become most of the potential of AI, such as the lack of transparency integral to the lives of many people and have provided a benefit of some AI algorithms which are unable to fully explain the to society. However, there is growing awareness amongst the reasoning for their decisions. Society has not yet built up the public, businesses and policy makers of the potential damage same level of trust in AI systems as in humans when making caused by harmful online material. The user generated complex decisions. There are risks that bias is introduced into content (UGC) posted by users contributes to the richness AI systems by incorporating data which is unrepresentative or and variety of content on the internet but is not subject to by programming in the unconscious bias of human developers. the editorial controls associated with traditional media. This In addition, there is a shortage of staff suitably qualified in enables some users to post content which could harm others, developing and implementing AI systems. However, many particularly children or vulnerable people. Examples of this of these inhibitors are being addressed: the supply of AI- include content which is cruel and insensitive to others, which skilled engineers will increase and it is likely that society will promotes terrorism or depicts child abuse. gradually develop greater confidence in AI as it is increasingly seen to perform complex tasks well and it is adopted in more As the amount of UGC that platform users upload continues to aspects of our lives. Other issues will, however, continue to be accelerate,1 it has become impossible to identify and remove a problem for at least the short term and these are discussed harmful content using traditional human-led moderation in more detail in this report. approaches at the speed and scale necessary. Current approaches and challenges to online content moderation This paper examines the capabilities of artificial intelligence (AI) technologies in meeting the challenges of moderating Effective moderation of harmful online content is a challenging online content and how improvements are likely to enhance problem for many reasons. While many of these challenges those capabilities over approximately the next five years. affect both human and automated moderation systems, some are especially challenging for AI-based automation systems to Recent advances in AI and its potential future impact overcome. The term ‘AI’ is used in this report to refer to the capability There is a broad range of content which is potentially harmful, of a machine to exhibit human-like performance at a defined including but not limited to: child abuse material, violent and task, rather than refer to specific technical approaches, such extreme content, hate speech, graphic content, sexual content, as ‘machine learning’. Although AI has been through several cruel and insensitive material and spam content. Some harmful cycles of hype followed by disillusionment since its inception in content can be identified by analysing the content alone, but the 1950s, the current surge in investment and technological other content requires an understanding of the context around progress is likely to be sustained. The recent advances in AI have it to determine whether or not it is harmful. Interpreting been enabled by progress in new algorithms and the availability this context consistently is challenging for both human and of computational power and data, and AI capabilities are now automated systems because it requires a broader understanding delivering real commercial value across a range of sectors. of societal, cultural, historical and political factors. Some of these contextual considerations vary around the world The latest advances in AI have mainly been driven by machine due to differences in national laws and what societies deem learning, which enables a computer system to make decisions acceptable. Content moderation processes must therefore be and predict outcomes without being explicitly programmed to contextually aware and culturally-specific to be effective. perform these tasks. This approach requires a set of data to train the system or a training environment in which the system Online content may appear in numerous different formats can experiment. The most significant breakthrough of machine which are more difficult to analyse and moderate, such as video learning in recent times is the development of ‘deep neural content (which requires image analysis over multiple frames to networks’ which enable ‘deep learning’. These neural networks be combined with audio analysis) and memes (which require enable systems to recognise features in complex data inputs a combination of text and image analysis with contextual and such as human speech, images and text. For many applications, cultural understanding). Deepfakes, which are created using the performance of these systems in delivering the specific task machine learning to generate fake but convincing images, for which they have been trained now compares favourably with video, audio or text, have the potential to be extremely harmful humans, but AI still does make errors. and are difficult to detect by human or AI methods. 04 USE OF AI IN ONLINE CONTENT MODERATION REPORT In addition, content may be posted as a live video stream or live whether content is harmful or not. Clarity in platforms’ text chat which must be analysed and moderated in real time. community standards is therefore essential to enabling the This is more challenging because the level of harmfulness can development and refinement of AI systems to enforce the escalate quickly and only the previous and current elements of standards consistently. the content are available for consideration. Overall, it is not possible to fully automate effective content Over time, the language and format of online content evolve moderation. For the foreseeable future the input of human rapidly and some users will attempt to subvert moderation moderators will continue to be required to review highly systems such as by adjusting the words and phrases they use. contextual, nuanced content. Human input is expensive and Moderation systems must therefore adapt to keep pace with difficult to scale as the volume of content uploaded increases. these changes. It also requires individuals to view harmful content in order to moderate it, which can cause them significant psychological Online platforms may moderate the third-party content they distress. However, AI-based content moderation systems can host to reduce the risk of exposing their users to harmful reduce the need for human moderation and reduce the impact material and to mitigate the reputational risk to the organisation. on them of viewing harmful content. However, removing content which is not universally agreed to be harmful can also result in reputational damage and Many organisations follow an online content moderation undermine users’ freedom of expression. Facebook, for workflow which uses moderation at one or both of two points example, has been criticised for removing an image of a statue in the workflow: of Neptune in Bologna, Italy for being sexually explicit and the iconic photograph