UNCORRECTED TRANSCRIPT of ORAL EVIDENCE to Be Published As HC 729-Iii
Total Page:16
File Type:pdf, Size:1020Kb
UNCORRECTED TRANSCRIPT OF ORAL EVIDENCE To be published as HC 729-iii HOUSE OF COMMONS ORAL EVIDENCE TAKEN BEFORE THE CULTURE, MEDIA AND SPORT COMMITTEE ONLINE SAFETY TUESDAY 19 NOVEMBER 2013 SIMON MILNER and SINÉAD MCSWEENEY TONY CLOSE and CLAUDIO POLLACK RT HON DAMIAN GREEN, MR EDWARD VAIZEY and CLAIRE PERRY Evidence heard in Public Questions 125 - 222 USE OF THE TRANSCRIPT 1. This is an uncorrected transcript of evidence taken in public and reported to the House. The transcript has been placed on the internet on the authority of the Committee, and copies have been made available by the Vote Office for the use of Members and others. 2. Any public use of, or reference to, the contents should make clear that neither witnesses nor Members have had the opportunity to correct the record. The transcript is not yet an approved formal record of these proceedings. 3. Members who receive this for the purpose of correcting questions addressed by them to witnesses are asked to send corrections to the Committee Assistant. 4. Prospective witnesses may receive this in preparation for any written or oral evidence they may in due course give to the Committee. 1 Oral Evidence Taken before the Culture, Media and Sport Committee on Tuesday 19 November 2013 Members present: Mr John Whittingdale (Chair) Mr Ben Bradshaw Angie Bray Conor Burns Tracey Crouch Paul Farrelly Jim Sheridan Mr Gerry Sutcliffe ________________ Examination of Witnesses Witnesses: Simon Milner, Policy Director, UK and Ireland, Facebook, and Sinéad McSweeney, Director, Public Policy, EMEA, Twitter International Company, gave evidence. Q125 Chair: This is the further session of the Committee’s inquiry into online safety and we have three panels this morning and I would like to welcome our first witnesses, Simon Milner, the Policy Director of Facebook, and Sinéad McSweeney, the Director of Public Policy at Twitter. Q126 Tracey Crouch: Good morning. I will start by asking if you could outline what it is that you already do to prevent harmful or illegal material online. Sinéad McSweeney: In terms of illegal content we have a zero tolerance for any child sexual exploitation material on the platform. So when that is reported to us, we review those accounts. They are suspended and they are, in addition, reported to the National Center for Missing and Exploited Children in the US and they then have channels of communication with the relevant law enforcement agencies, who can take action both on the offenders but also ensure that if there is a child at risk, that they can be identified, located and rescued as well. We also have relationships with law enforcement and police in different countries. So we have good working relationships with the police here in the UK not only so that they are aware of our processes and procedures in order to request information of us and work with us, but also so that we would participate in some education and awareness raising with them about how the platform works and how to contact us when they need to, whether that is in an emergency or when they are investigating crime. Q127 Tracey Crouch: What about harmful content, before I move to Simon? Sinéad McSweeney: Yes. Sorry, did you say Simon? Tracey Crouch: No, no. Sinéad McSweeney: Okay. In terms of content generally, clearly, there are rules as to the type of content that is allowed on the platform. So when content is reported to us that is in breach of our rules, we take action on those accounts. In addition, because there is such a diverse range of discussions and contact that take place on the platform, we also have built 2 into the product the ability for users to label the media that they may be uploading as sensitive, and the default setting for every user is that they do not see sensitive media without a warning. Q128 Tracey Crouch: You do not have a report abuse button though, do you? Sinéad McSweeney: We do, we do. Tracey Crouch: Okay. Sinéad McSweeney: We have it within each tweet because Twitter is built around the small bursts of information that are tweets. But right within each tweet there is a report button and once the user hits that, it takes them to the range of possible issues that they may be reporting, whether that is abusive behaviour, impersonation or illegal content. Tracey Crouch: Okay. Simon? Simon Milner: From our own perspective, we also have a zero tolerance approach towards both illegal and harmful material. On the illegal side, as well as having clear policies about what people can and cannot use our platform for, we also use technology. So there was much talk yesterday in terms of the announcements made at the PM Summit around technology. We use one of the pieces of technology that people are talking about, photo DNA, a piece of technology created by Microsoft, and we use that to scan every photo uploaded to Facebook. That are some 350 million photos a day against a database of known images of child exploitation imagery, and we will not allow those to be uploaded and will take action against those accounts, including notifying the authorities about them. Then we work closely with CEOP in respect of their access to that data via Netmec in order to track down the offenders. When it comes to harmful material, that is pretty much very hard to spot by technology so that is where our reporting processes come in, and we too have extensive reporting processes via both our Help Centre, which anybody can access—they do not have to have a Facebook account for that—or from individual pieces of content. They can report it to us or indeed report it to somebody else who can help that user take action so, particularly, in the areas, for instance, of bullying and harassment, we do not tolerate that. Those reporting processes are incredibly useful to someone either who themselves is feeling a risk of harm or if somebody sees somebody else on the platform who they think is at risk, then they can use those reporting processes. Q129 Tracey Crouch: This week, as I am sure you are aware, is anti-bullying week and there has been a great deal of discussion about cyber bullying over the past few months, especially as a consequence of some rather tragic events. What are you doing to work with organisations to try to raise awareness about cyber bullying in particular? Simon Milner: So perhaps if I take that one first. We work with a number of anti- bullying organisations. Indeed, yesterday, I was at an event in London with 450 schoolchildren and their teachers with the Diana Award. The Diana Award has an anti- bullying work stream and they have a fantastic champion there in a young man called Alex Holmes. He does great work around the UK, working with schoolchildren, enabling them to become anti-bullying champions so we support that work. We help to fund what they do and I was there yesterday to talk mainly to teachers to help them understand how to deal with these issues on Facebook, because it is something that we absolutely recognise they need help with. But we also provide lots of help via other people who reach out to schools. So Childnet are another great charity for doing this, because we recognise bullying typically does not just happen on one platform and will happen across a range of different media as well as in the school. Therefore, working with people who can not only talk to the schools about Facebook but also about Twitter, about Snapchat, about Ask.fm and other services that young people are using is much more effective than having Facebook necessarily coming to the schools. 3 Sinéad McSweeney: I think we would have a similar approach in terms of partnering with the people who are talking to children all the time in this space, so the South West Grid for Learning, the UK Safer Internet Centre, participating not just in the Safer Internet Day or the anti-bullying weeks but on an ongoing basis, highlighting the resources that are available both within the platform and more generally in terms of breaking down the advice that we have to ensure it is suitable for each different category of people who need to know and understand what is happening. So, for example, in our Safety Centre we would have tips for teachers, tips for parents, tips for teens, recognising that the language and the context may need to be slightly different for each audience. That is material that we have highlighted again, as I say, to organisations who are working in this space on a daily basis, whether it is Parent Zone or BeatBullying or whatever, so that they can link to our material and also so that we can link to their material so that there is a constant shared learning. Also, as Simon says, recognising that the context for the behaviour that you may be seeing on one platform may be behaviour and activity that is occurring in other platforms and also in the offline world, it also about encouraging how to deal with online conflict, encouraging empathy and also trying to educate young people that the consequences of their online behaviour are the same as offline. If you say something online, it has an impact on somebody in the real world although for young people it is their real world, so sometimes we have to be careful even in terms of that kind of language.