Front Cover

Transcription

Front Cover
House of Commons
Culture, Media and Sport
Committee
Online safety
Sixth Report of Session 2013–14
Volume I
Volume I: Report, together with formal
minutes, oral and written evidence
Additional written evidence is contained in
Volume II, available on the Committee website
at www.parliament.uk/cmscom
Ordered by the House of Commons
13 March 2014
HC 729
Published on 19 March 2014
by authority of the House of Commons
London: The Stationery Office Limited
£20.00
The Culture, Media and Sport Committee
The Culture, Media and Sport Committee is appointed by the House of Commons to
examine the expenditure, administration and policy of the Department for Culture,
Media and Sport and its associated public bodies.
Current membership
Mr John Whittingdale MP (Conservative, Maldon) (Chair)
Mr Ben Bradshaw MP (Labour, Exeter)
Angie Bray MP (Conservative, Ealing Central and Acton)
Conor Burns MP (Conservative, Bournemouth West)
Tracey Crouch MP (Conservative, Chatham and Aylesford)
Philip Davies MP (Conservative, Shipley)
Paul Farrelly MP (Labour, Newcastle-under-Lyme)
Mr John Leech MP (Liberal Democrat, Manchester, Withington)
Steve Rotheram MP (Labour, Liverpool, Walton)
Jim Sheridan MP (Labour, Paisley and Renfrewshire North)
Mr Gerry Sutcliffe MP (Labour, Bradford South)
The following members were also a member of the committee during the parliament:
David Cairns MP (Labour, Inverclyde)
Dr Thérèse Coffey MP (Conservative, Suffolk Coastal)
Damian Collins MP (Conservative, Folkestone and Hythe)
Alan Keen MP (Labour Co-operative, Feltham and Heston)
Louise Mensch MP (Conservative, Corby)
Mr Adrian Sanders MP (Liberal Democrat, Torbay)
Mr Tom Watson MP (Labour, West Bromwich East)
Powers
The committee is one of the departmental select committees, the powers of which
are set out in House of Commons Standing Orders, principally in SO No 152. These
are available on the internet via www.parliament.uk.
Publication
The Reports and evidence of the Committee are published by The Stationery Office
by Order of the House. All publications of the Committee (including press notices)
are on the internet at www.parliament.uk/cmscom. A list of Reports of the
Committee in the present Parliament is at the back of this volume.
The Reports of the Committee, the formal minutes relating to that report, oral
evidence taken and some of the written evidence are available in a printed volume.
Additional written evidence is published on the internet only.
Committee staff
The current staff of the Committee are Elizabeth Flood (Clerk), Grahame Danby
(Second Clerk), Kevin Candy (Inquiry Manager), Emily Gregory (Senior Committee
Assistant), Keely Bishop (Committee Assistant) and Jessica Bridges-Palmer (Media
Officer).
Contacts
All correspondence should be addressed to the Clerk of the Culture, Media and
Sport Committee, House of Commons, 7 Millbank, London SW1P 3JA. The telephone
number for general enquiries is 020 7219 6188; the Committee’s email address is
[email protected]
Online safety
1
Contents
Report
Page
Summary
3
1
Introduction
6
2
Child abuse images
8
Nature and scale
8
3
4
The law
Enforcem ent
Surveillance
9
10
12
Internet service providers
Internet Watch Foundation
Deterrence
14
15
16
Other material
17
Adult content
19
Nature and scale
19
The law
Enforcem ent
Age verification
20
21
24
Site blocking
Filters
Media literacy and education
26
27
31
Social media
34
Nature and scale
34
The law
Enforcem ent
Reporting
35
36
39
Advice and support
Anonymity
40
42
Formal Minutes
48
Witnesses
49
List of printed written evidence
50
List of additional written evidence
50
List of Reports from the Committee during the current Parliament
52
Online safety
3
Summary
The internet has revolutionised communications and information sharing. It provides an
ever increasingly important platform for creativity and economic growth. Online social
media services are providing new ways of interacting and keeping in touch. Online
communications enable expressions of human behaviour both positive and negative;
sometimes downright criminal. Our inquiry has focused on three disparate aspects of
online content and behaviour, all of which are of widespread concern: illegal content,
especially images of child abuse; harmful adult content being made freely available to
children; bullying and harassment on social media.
Both the publication and possession of child abuse images are rightly illegal. While these
offences are bad enough, it must not be forgotten that such images represent crime scenes,
often of the most horrific kind. There is a clear need to ensure that the police have
adequate resources to track down and arrest online paedophiles in sufficient numbers to
act as a meaningful deterrent to others. If necessary, additional funding should be
provided to recruit and train a sufficiently large number of police officers adequate to the
task.
The Child Exploitation and Online Protection (CEOP) Command, now part of the new
National Crime Agency, has a well-deserved reputation as a lead body in tackling child
abuse. It has been increasingly effective not least because it is not solely a criminal justice
organisation: its education and social care work has also been very important in increasing
public understanding of the problem of child abuse and in offering means of countering
abusers. All three elements of its mission – education, social care and criminal justice –
need to be actively pursued and publicised.
The Internet Watch Foundation, too, has played a crucial role in removing and blocking
child abuse images online. We very much welcome their new commitment to embark on
proactive searching for online child abuse images. The sooner these can be found and
removed, the better. However, we are concerned that the additional staff resources being
allocated to this task could prove woefully insufficient to achieve substantial progress
towards what must be an important intermediate goal: the eradication of child abuse
images from the open internet. Tracing paedophiles who share images on peer-to-peer
networks and the so-called hidden internet continues to challenge both the police and the
internet service providers; it is a challenge that, by working together, they must overcome.
Legal adult pornography is widely consumed. This includes explicit “hard core” material
that attracts an R18 certificate from the British Board of Film Classification. Parents and
carers clearly have a key role, not to mention interest, in preventing harmful material of
this kind becoming available to children. However, they should have access to more
information and help where and when they need it. In the off-line world, it is the
newsagent, not the parent, who voluntarily places some adult magazines on a top shelf out
of reach of children. It is the local authority, not the parent, which administers the licensing
of sex shops selling R18 pornography to which children may not be admitted. Some level of
analogous protection ought to be provided in relation to online material. At the moment,
4
Online safety
little is.
Legal adult sites could restrict access by children in a number of ways. In general a robust
age verification process should be in place; as part of this, sites could use a combination of
the following: requiring payment by a credit card linked to an adult; shielding the content
behind a warning page; attaching metadata to the website to make it easier for filters to
operate and for search engines not to return the material when operating in a safe search
mode.
Filters may not be failsafe, but they continue to improve and are an important way of
protecting children from harmful content. We very much welcome the introduction of
whole home filtering solutions that prompt account holders with a choice to apply them.
The main internet service providers should have contacted all their customers by the end
of the year to offer this valuable service. We want to see all other ISPs following suit.
Publishing adult pornography in a way that makes it readily available to children is likely to
be an offence under the Obscene Publications Act 1959. We do not believe the police
should be deterred from bringing to book publishers of adult pornography who make little
attempt to shield children from their product. While acknowledging that the enforcement
of obscenity legislation is fraught with difficulty, not least in the context of the internet, we
believe there is scope for greater enforcement in this area to provide some deterrent effect.
There may also be scope for blocking particularly harmful adult websites that make no
serious attempt to hinder access by children.
As part of its existing media literacy duties, Ofcom has an important role in monitoring
internet content and advising the public on online safety. However, we are anxious to
avoid suggesting a significant extension of formal content regulation of the internet.
Among the unintended consequences this could have would be a stifling of the free flow of
ideas that lies at the heart of internet communication. Rather, more needs to be done to
signpost the advice and educational resources available to both parents and teachers. This
is all the more pressing given the growing use of social media and its misuse by some –
both adults and children. Today, one in five 12–16 year-olds think being bullied online is
part of life.
Social media providers should offer a range of prominently displayed options for, and
routes to, reporting harmful content and communications. They should act on these
reports expeditiously, keeping the complainant and—where appropriate—the subject of
the complaints informed of outcomes and actions. Given that Facebook and Twitter are
aware of the extent to which their services are accessed by younger children, thanks to age
verification processes that are at best flimsy, we expect them to pay greater attention to
factoring this into the services provided, the content allowed and the access to both. The
same applies to other social media companies in a similar position.
Some of the worst online bullies and trolls are being brought to book in the courts. Much
of the abuse and bullying that takes place online is covered by existing laws, but these need
to be clarified with guidance updated for the online space. Young people especially are
distinguishing less and less between their lives on the internet and in the real world.
Bullying that takes place in the playground can merge seamlessly with bullying on smart
phones and tablets. Sometimes this ends with the tragedy of teenage suicide. It is just one
Online safety
reminder that staying safe off-line includes staying safe online too.
5
6
Online safety
1 Introduction
1. The internet is changing the way we communicate and modifying the way we behave. A
wealth of information and a platform for myriad interactions, it is fostering creativity and
economic growth on an unprecedented scale. Sadly, inevitably, the darker side of human
nature finds expression too: in the commissioning, distribution and viewing of illegal
images of child abuse; in adult material unshielded from children’s eyes; in threatening and
abusive messages via social media. In the belief, sometimes mistaken, that the internet
offers anonymity, trolls and bullies hound adults and children alike, often with tragic
consequences. The internet can amplify the pack mentality of the unthinking.
2. With these disparate concerns in mind, we decided to launch an inquiry into online
safety, taking care in our terms of reference to keep distinct three very separate aspects:
•
How best to protect minors from accessing adult content;
•
Filtering out (i.e. blocking and removing) extremist material, including images of child
abuse and material intended to promote terrorism or other acts of violence;
•
Preventing abusive or threatening comments on social media.
3. We took oral evidence from the following: John Carr, Secretary, Children’s Charities’
Coalition on Internet Safety; Claire Lilley, Senior Analyst, NSPCC; Anthony Smythe,
Managing Director, BeatBullying; Susie Hargreaves, Chief Executive, Internet Watch
Foundation; Peter Davies, Director, Child Exploitation and Online Protection (CEOP)
Command, National Crime Agency; Stella Creasy MP; Nicholas Lansman, Secretary
General, Internet Services Providers’ Association; Dido Harding, Chief Executive Officer,
TalkTalk Group; Hamish Macleod, Chair, Mobile Broadband Group; Jim Gamble,
Independent Chair, City and Hackney Safeguarding Children Board; Simon Milner, Policy
Director, UK and Ireland, Facebook; Sinéad McSweeney, Director, Public Policy, EMEA,
Twitter International Company; Tony Close, Director of Content Standards, Licensing and
Enforcement, Ofcom; Claudio Pollack, Group Director, Content Consumer and External
Affairs Group, Ofcom; Rt Hon Damian Green MP, Minister of State for Policing, Criminal
Justice and Victims; Edward Vaizey MP, Parliamentary Under-Secretary of State for
Culture, Communications and Creative Industries; Claire Perry MP, Special Adviser to the
Prime Minister on Preventing the Commercialisation and Sexualisation of Childhood.
4. Many other organisations and individuals submitted written evidence, among them
charities representing children and young people. Nevertheless, we felt it important to
engage directly with young people themselves. We were therefore grateful to the following
charities which, together with parents, organised a visit to the House of Commons by eight
young people ranging from ages 16 to 25: Childline Services, CARE,1 Childnet
International, Porn Scars, Safermedia and Brook. The young people each gave us
compelling and, at times, harrowing accounts of the harmful experiences they had endured
as a result of online bullying and access to age-inappropriate material, particularly adult
1
Christian Action Research and Education
Online safety
7
pornography. We are very grateful to them for speaking to us and, indeed, to all who
provided evidence to our inquiry.
8
Online safety
2 Child abuse images
Nature and scale
5. Both the publication and possession of child abuse images are illegal. This is one form of
censorship that commands near universal support. Not only do such images represent
actual crime scenes, but they are likely to be used by online paedophiles to encourage and
legitimise their criminal activities. The law in the UK recognises this by proscribing a
range of images of different types, whether photographic or not.
6. Estimating the scale of child abuse images on the internet is problematical given the
secretive nature of much online activity by paedophiles. The term “child pornography”,
while still in common use, not least in legal texts, is now largely abjured by professionals
working in child protection. In no way can images of this kind by likened to consensual
activities depicted in legal adult pornography. Paedophiles ought to be denied any comfort
when seeking to affirm their practices and preferences. Accordingly, we shall use the terms
“child abuse images” or “child sexual abuse images” to refer to the material covered in this
section of our report. Claire Lilley of the NSPCC told us:
We need to educate people that these are not just images; that by looking at these
images, they are committing an offence; that a child is being re-victimised every
single time an image is looked at; and that they are potentially affecting their own
sensitivity around the issue and escalating their own ability to go down that route
and end up abusing in the physical sense themselves. There is quite a lot of research
now about the crossover between non-contact and contact offending and CEOP
would put it at about 50%. There are a couple of studies that put it at between 40%
and 55%, which is a very high level of crossover. It is not just looking at an image; it is
much more dangerous than that.2
Peter Davies of CEOP underlined this point: “anybody who possesses indecent images of
children is a criminal, but is also somebody who might present an additional risk to
children, as if that were not enough.”3
7. As the NSPCC notes, child abuse images are a visual record of the sexual abuse of a
child. They can include pseudophotographs, animations, drawings, tracings, videos and
films which are being streamed live. In the UK images are graded on a 1–5 scale. Level 5
images involve sadism or bestiality, Level 4 will portray a child engaged in penetrative
sexual activity and so on to Level 1, where the images will depict erotic posing with no
visible sexual activity.4 In 2012 the NSPCC issued FOI requests to every local police force in
England and Wales asking them to state how many child abuse images they had seized in
arrests made in the two years ending April 2012. The five police forces (none of which had
a large metropolitan base)5 that replied had seized over 26 million such images. The
2
Q 29
3
Q 34
4
Ev 70
5
Q5
Online safety
9
Children’s Charities’ Coalition on Internet Safety told us that, on one calculation, that
would imply that over 300 million illegal images may have been seized by all forces over the
same period.6 Many of the images seen by the Internet Watch Foundation have been
recycled, though one or two new images—each representing a new victim—are seen
weekly.7
The law
8. Section 1 of the Protection of Children Act 1978 makes it a criminal offence to take,
permit to be taken or to make, distribute, show, advertise or possess for distribution any
indecent photograph or pseudo-photograph of a child under the age of 18. Simple
possession by individuals is proscribed by the Criminal Justice Act 1988. The 1978 Act
defines a pseudo-photograph as “an image, whether made by computer-graphics or
otherwise howsoever, which appears to be a photograph.” Pseudophotographs were
brought within the ambit of the Act by dint of the Criminal Justice and Public Order Act
1994. The Criminal Justice and Immigration Act 2008 extended the definition of an
indecent photograph to include a tracing or other image derived from a photograph or
pseudophotograph. Part 2, Chapter 2 of the Coroners and Justice Act 2009 extended the
law proscribing the possession of child pornography to include non-photographic images
such as cartoons, drawings and computer-generated images.
9. Peter Davies of CEOP told us: “it is quite remarkable to me how far the criminal
legislation, for example, around indecent images of children, which was, I believe,
conceived and passed before the internet was around, has still pretty much stood up to the
new world of child abuse online, and I do not think what we need is a basic root-andbranch piece of work.”8 He added that “the UK has just about the best suite of child
protection legislation that there is, and that we are more of an example to others than we
are in need of catching up.” 9 At a recent discussion meeting of the Digital Policy Alliance,
the Deputy Director of CEOP Command, Andy Baker, referred in general terms to child
protection legislation more widely; he suggested it would be better to “tidy it up”.10 The fact
that the Communications Act 2003 was drafted with no mention of the internet also
should be addressed. We believe that the Government should, in due course, consolidate
the law around child abuse images into a single Act of Parliament with a view to
providing even greater clarity for the purposes of law enforcement and deterrence.
10. Clearly the fight against child abuse and child abuse content on the internet is an
international one. Peter Davies referred to two conventions—the Budapest Convention
and the Lanzarote Convention—which together aim to provide a legal framework for
protecting children online. The Budapest Convention on Cybercrime11 and the Lanzarote
Convention on the Protection of Children against Sexual Exploitation and Sexual Abuse 12
6
Ev 66
7
Qq 34-35, 40
8
Q 48
9
Q 48
10
Recent Developments in Child Internet Safety, Digital Policy Alliance Discussion Meeting, 22 January 2014
11
http://conventions.coe.int/Treaty/Commun/ChercheSig.asp?NT=185&CM=8&DF=&CL=ENG
12
http://conventions.coe.int/Treaty/Commun/ChercheSig.asp?NT=201&CM=1&DF=&CL=ENG
10
Online safety
are both Conventions of the Council of Europe. The United Kingdom is a signatory to
both, though it has ratified only the former. Last July, the Parliamentary Under-Secretary
of State, Home Office (Lord Taylor of Holbeach) was unable to say when the Government
expected to ratify the Lanzarote Convention; he said: “Until the work to assess the practical
arrangements is complete, it will not be possible to confirm the timescales for ratification.
That work remains ongoing.”13 Given the worldwide nature of online crime, we
recommend that the Government press for wider international adoption of both the
Budapest and Lanzarote Conventions. The Government should ratify the Lanzarote
Convention as soon as practicable.
Enforcement
11. Peter Davies of CEOP referred to an assessment, published in June 2013,14 which gives
an estimate of around 50,000 people in the UK who commit child sexual abuse offences “at
least to the level of possessing indecent images of children.”15 He went on to tell us that the
highest volume of offenders are using peer-to-peer networks rather than the open internet.
“Far fewer” offenders are using the hidden internet, “also known as Tor or The Onion
Router or something similar.”16
12. The proliferation of material which is illegal in all circumstances poses a particular
threat and a range of challenges, not least to law enforcement. John Carr of the Children’s
Charities’ Coalition on Internet Safety referred to the challenge posed by the scale of online
child sexual abuse: “I think it is a profound shock to realise that so many people are
interested in that type of material. There is an appetite for it and people go and get it, and
they circulate it. They are getting away with it because we do not have the capacity to stamp
it out.” 17 Peter Davies of CEOP referred to a need for “a plan that goes beyond
enforcement”.18 That is not to diminish the vital role that law enforcement must continue
to play. The former chief executive of CEOP, Jim Gamble, emphasised the need to deter
paedophiles by applying the law as it currently stands.19 Citing figures given by Peter
Davies, Jim Gamble said:
On the 50,000 Peter talks about, those are peer-to-peer sites. Those are hard-core
paedophiles who do not stumble across anything on Google. They nest in these
places on the internet where they can share secretly. You have to infiltrate that. You
have to infiltrate it. You have to identify who they are and follow them offline and
arrest them.20
13. He went on to propose the recruitment, by each police force, of up to ten special
constables to work online. Their job would be to seek out online paedophiles with the aim
13
HL Deb, 24 July 2013, col 197WA
14
http://ceop.police.uk/Documents/ceopdocs/CEOP_TACSEA2013_240613%20FINAL.pdf
15
Q 40
16
Q 43
17
Q 27
18
Q 59
19
Qq 114, 116
20
Q 114
Online safety
11
of achieving arrests in sufficient numbers to have a significant deterrent effect. 21 Mr
Gamble puts a cost of £1.4 million on setting up such a scheme which would include the
recruitment of 12 detective sergeants and 12 trainers and coordinators. 22 He told us: “If
you did that, you would have 520 people online at any given time who could manifestly
cover 10, 20, 30 or 40 chat-room environments, where you would have people out there
protecting our children. Not rogue vigilantes, but properly vetted, properly trained and
accredited and properly supervised officers online. To me, that is what we should be
thinking about so we attack the root cause, which is people.”23
14. CEOP (Child Exploitation and Online Protection) is now one of the four commands
within the National Crime Agency, which commenced on 7 October 2013. The CEOP
centre has existed since 2006 and was previously affiliated to the Serious Organised Crime
Agency. Peter Davies told us: “Our mission, whether as a centre or as a command, as we
are now, is to protect children from sexual exploitation and sexual abuse. When we were
founded, there was heavy emphasis on online protection. Our remit does not limit us to
online activity, but what we try to do is to become a national hub for intelligence, a national
contact point and a source of expertise and specialist support on any aspect of child sexual
exploitation that would benefit from that approach.” Mr Davies also highlighted the
contributions being made by online child protection units in UK police forces. 24
15. In its evidence, the Home Office praised the Child Exploitation and Online Protection
Command of the National Crime Agency for both its operational work and educational
programmes. In 2012/13 CEOP received an average of 1,600 reports per month of abuse
from the public and the internet industry, a 14% increase on the previous year. “CEOP
maintains close relationships with a wide range of stakeholders from law enforcement,
industry and the educational sectors nationally and internationally.”25 In 2012/13, CEOP
protected 790 children from sexual abuse, an increase of 85% on the previous year, and its
work led to the arrest of 192 suspects.26
16. According to the Home Office, the CEOP budget “has effectively been protected in
cash terms since 2011/12” and “there are now more people working in CEOP than at any
time in its history”.27 The Home Office also states that becoming part of the NCA will
bring advantages to CEOP (such as access to greater capacity and support from other parts
of the NCA) and that all NCA officers will receive mandatory training on safeguarding
children. The Minister of State for Policing, Criminal Justice and Victims, Damian Green,
described becoming an arm of the National Crime Agency as a “game-changer”28 for
CEOP because of access to increased resources. Peter Davies told us that incorporation of
CEOP into the National Crime Agency was “absolutely a good thing.”29 His predecessor,
21
Q 113
22
Ibid.
23
Ibid.
24
Q 45
25
Ev 105
26
CEOP Annual Review 2012-2013 & Centre Plan 2013-2014
27
Ev 105
28
Q 204
29
Q 50
12
Online safety
Jim Gamble, clearly disagrees. He described subsuming CEOP into the National Crime
Agency as “a recipe for disaster”.30 Mr Gamble added:
To take a child protection agency and put it into a national crime agency where the
brand has now been diluted, where they will continue to have protection under
Freedom of Information, which is fundamentally wrong, how can you have a child
protection entity that is not subject to Freedom of Information? You have a child
protection entity that is answerable through the NCA to one political person, the
Home Secretary. Where is the credible reassurance around that? Where are the
lessons of serious case review around professional independence and challenge? The
fact of the matter is we had planned the three pillars to be education, social care, and
criminal justice. We now have moved to one pillar, which is criminal justice. 31
17. We recommend that the Government examines whether adequate resources are
being deployed to track down online paedophiles in sufficient numbers to act as a
meaningful deterrent to others. If not, additional funding should be provided to
recruit and train a sufficiently large number of police officers adequate to the task.
18. CEOP has been increasingly effective not least because it is not solely a criminal
justice organisation: its education and social care work has also been very important in
increasing public understanding of the problem of child abuse and in offering means of
countering abusers. We therefore recommend that CEOP continues to publish an
annual review which includes an assessment of its ongoing contribution to all three
elements of its mission—education, social care and criminal justice.
Surveillance
19. Tracing paedophiles online clearly involves allowing the police to deploy a range of
surveillance methods. The Regulation of Investigatory Powers Act 2000 aims to provide a
legal regime for both interception of communications and access to communications data
that is consistent with human rights legislation—specifically having regard to privacy and
proportionality. Peter Davies argued for further changes to the relevant legislation:
It is my firm operational view, and I have articulated it previously, that if there is one
piece of legislation that would most help us tackle online child abuse, it would be the
provision of a clear, consistent and properly enforced regime for retaining and
accessing communications data, because we are regularly in a situation where we are
unable to convert, for example, an IP address into a name and address for lack of that
precise thing.32
20. One potential area of operational difficulty was illustrated in evidence from the End
Violence Against Women Coalition; their evidence cites CEOP as having warned that live
streaming of child abuse through Skype and other platforms is emerging as a growing
30
Q 124
31
Q 119
32
Q 48
Online safety
13
method of abusers sharing child abuse images. 33 A government briefing on proposals on
the investigation of crime in cyberspace, prepared for the Queen’s Speech in 2013, noted:
When communicating over the Internet, people are allocated an Internet Protocol
(IP) address. However, these addresses are generally shared between a number of
people. In order to know who has actually sent an email or made a Skype call, the
police need to know who used a certain IP address at a given point in time. Without
this, if a suspect used the internet to communicate instead of making a phone call, it
may not be possible for the police to identify them.34
21. A draft Communications Data Bill was announced in the Queen’s Speech in 2012 and
it was published on 14 June 2012. It was scrutinised by a Joint Committee of both Houses
of Parliament and was also considered by the Joint Committee on Human Rights (JCHR)
and the Intelligence and Security Committee (ISC). The draft Bill would have extended
powers to obtain communications data covering messages sent on social media, webmail,
voice calls over the internet and gaming in addition to emails and phone calls. The data
could have included the time, duration, originator and recipient of a communication and
the location of the device from which it is made. However, it would not have included the
actual content of messages. The Joint Committee on the draft Bill published its report on
11 December 2012, concluding that the Bill’s scope should be significantly narrowed, while
recognising that more needed to be done to provide law enforcement and other agencies
with access to data they cannot currently obtain. The Bill would have sat alongside the
existing Data Retention (EC Directive) Regulations 2009. 35
22. An illustration of both the capability and concerns of internet service providers was
given us by Dido Harding of TalkTalk:
I feel very strongly the weight of responsibility as an internet service provider in that
our customers place their faith in the fact that it is their data, not our data, to choose
what to do with and, therefore, we need a clear legal framework on what we store
and what we do not store.
[...]
We do not keep browsing history of where our customers browse every day of the
week— that is their data, not ours—unless or until there was a change in legislation
that required us to.
[...]
It is very important that the ISPs do not feel like it is a free-for-all to use that data.
Our customers do not expect that of us. We have been thinking through, if it is
entirely criminal, how we get to a place where we can do that proactively with
CEOP.36
33
Ev w85-w88
34
HM Government, The Queen’s Speech 2013, 8 May 2013.
35
SI 2009/859
36
Qq 93,95
14
Online safety
23. The South West Grid for Learning told us that they have been working with the
Internet Watch Foundation and South West police forces since 2006 to flag up any
attempted access to websites containing child abuse images (CAI). This pilot project has
been managed by CEOP with Home Office approval. Its objective is to flag intelligence to
police forces if anyone in a school (specifically) attempts to access a website containing
child abuse images. This intelligence prompts the police to then undertake their normal
investigation routes and has resulted in a number of school staff being identified and
removed as a direct result of what South West Grid for Learning terms a “simple alerting
process.”37 They added: “We believe we should use technology better to identify those
accessing CAI (and extremist material) and we are involved in a project with Plymouth
University and IWF to extend and refine our existing alerting capability.”38 We welcome
the increasing use of alerting tools to identify individuals who seek out child abuse and
other illegal material online provided these tools are deployed in ways that do not
unduly compromise the privacy of the law-abiding majority.
Internet service providers
24. The scope for further practical action against illegal content is better understood by
recognising the different types of providers and how they fit within the current regulatory
framework. The Internet Services Providers’ Association identifies four main categories of
internet companies:
•
Access providers—considered to be “mere conduits” under the e-commerce
regulations39 (regulation 17).
•
Hosting providers—these include social networks; while these do not have editorial
control over content uploaded by users, they may have active or passive moderating
policies. Under Regulation 19 of the e-Commerce Regulations they are not liable for the
content they host as long as they do not have actual knowledge of unlawful activity or
information. “However, upon obtaining such knowledge, hosting providers become
liable if they do not act expeditiously to remove or to disable access to the
information.”40
•
Websites where operators have editorial control—content might include a news article
and user-generated content like comments on the article.
•
Search engines— considered as "caches" under Regulation 18 of the e-Commerce
Regulations; search engines “act expeditiously to remove or to disable access to any
information if they are made aware of that the fact that this information may be
illegal.”41
37
Ev w49
38
Ev w49
39
The Electronic Commerce (EC Directive) Regulations SI 2002/2013
40
Ev 79
41
Ev 79-80
Online safety
15
Internet Watch Foundation
25. The UK internet industry was responsible for founding the Internet Watch
Foundation, a membership organisation that serves as the UK hotline where the public can
report child sexual abuse content, criminally obscene adult content and non-photographic
images of child sexual abuse. In 2012 alone, the IWF processed 39,211 reports and assisted
with the removal of 9,696 URLs containing potentially criminal child sexual abuse content.
A URL can be as specific as a single image or could refer to an entire website containing
potentially thousands of child sexual abuse images or videos. The majority of victims (81%)
appeared to be 10 years old or younger (with 4% 2 years old or under) and 53% of the
images and videos depicted sexual activity between adults and children, including rape and
sexual torture.42
26. When child sexual abuse content is found to be hosted in the UK, the IWF will inform
CEOP. After confirmation from CEOP that action can be taken, the IWF will notify the
hosting provider who will remove the content from its servers, typically within 60 minutes
after receiving the notification from the IWF. This process is commonly referred to as
‘Notice and Takedown’. The IWF can also act against criminally obscene adult content
and non-photographic child sexual abuse content hosted in the UK. 43
27. When child sexual abuse content is found to be hosted outside the UK (accounting for
99% of known content), the IWF will inform its counterpart hotline in the hosting country
through INHOPE, the international association of hotlines, or link in directly with local
law enforcement. As other countries take significantly longer to remove child sexual abuse
content—50% of the content about which the IWF passes on details internationally is still
available after 10 days—the IWF adds the links (URLs) to the content to its URL list (or
‘blocking list’). IWF members can use this list to voluntarily block access to these URLs to
protect their customers from stumbling upon the images and videos. The Home Office
told us that such blocking arrangements apply to about 98.6% of domestic broadband
lines.44 Susie Hargreaves of the Internet Watch Foundation told us: “The most effective
way to remove content is to remove it at source. It is our view that blocking will only stop
inadvertent access and will not stop the determined.”45
28. On 18 June 2013, the Secretary of State for Culture, Media and Sport, Maria Miller,
hosted a summit on tackling child sexual abuse material on the internet as well as
protecting children from harmful or inappropriate content online. Participants included
internet service providers, search engines, mobile operators and social media companies.
Following the summit, the Secretary of State announced that the Internet Watch
Foundation would work with CEOP to actively seek out illegal images of child abuse on
the internet. The Internet Watch Foundation told us that, following a donation by Google
and additional funding by other members, they will be able to increase their number of
42
Ev 78
43
Ev 77
44
Ev 104
45
Q 34
16
Online safety
analysts from 4.5 (FTE46) to 11.5 (FTE) and start proactively searching for child sexual
abuse content as requested by the Government.47
29. We very much welcome the commitment by the Internet Watch Foundation to
embark on proactive searching for online child abuse images. The sooner these can be
found and removed, the better. However, we are concerned that seven additional staff
might prove woefully insufficient to achieve substantial progress towards what must be
an important intermediate goal: the eradication of child abuse images from the open
internet.
Deterrence
30. The IWF told us that, in addition to “Notice and Takedown” and the URL list, they also
compile a keyword list of terms that specifically refer to child sexual abuse content:
This list is used, for instance, by search engines to prevent people from finding
images and videos of child sexual abuse content. The keywords are very specific—or
very specific combinations of words—that carry no meaning besides the specific
reference to child sexual abuse content. This means the keywords will not prevent
access to legitimate websites such as, academic research papers into the area of child
sexual abuse or websites aimed to help or inform people in relation to child sexual
abuse.48
31. The Government has asked search engine providers to go further in restricting access
to child abuse images. According to the Home Office, they are being asked to develop
effective deterrence measures, to ensure child abuse images are not returned in search
results, and to prevent any search results being returned “when specific search terms are
used that have been identified by CEOP as being unambiguously aimed at accessing illegal
child sexual abuse images.”49 If progress is not forthcoming, the Government will consider
introducing legislation to ensure search engines comply.50
32. The IWF is also working with its members to introduce “splash pages”—these are
warning messages that appear if a user attempts to access a webpage that has been removed
for hosting illegal child abuse images. According to the IWF, “they deliver a hard-hitting
deterrence message to users seeking to access child abuse images.”51 Greater use of splash
pages and warning messages “to deter a certain class of person with a low level, opportunist
or early interest in child abuse images” is one of a number of tactics put to us by the
Children’s Charities’ Coalition on Internet Safety.52 Jim Gamble reminded us that splash
screens would not result in more hard-core paedophiles being arrested: “it is a diversion of
attention and resource that does not work. We tried it.”53 The Home Office pointed out to
46
Full time equivalent
47
Ev 78
48
Ev 77
49
Ev 104
50
Ev 105
51
Ev 104
52
Ev 67
53
Ev 116
Online safety
17
us that the objective of further work by search engines and greater use of splash pages “is to
make it more difficult for unsophisticated users to find a route from open searching to
more sophisticated offending environments, make it more difficult for inquisitive nonoffenders to access indecent images of children, and make it less likely that members of the
public could inadvertently come across such images.”54
33. Search engines and other internet service providers have a vital role in ensuring that
access to online child abuse images is prevented and deterred. We expect the
Government to monitor closely their degree of commitment and success and to
consider the introduction of legislation should they fall short of reasonable
expectations.
Other material
34. Much of the evidence we took on illegal content was in relation to child abuse images.
However, in the United Kingdom at least, certain categories of extreme adult pornography
are illegal both to publish and possess. Pornographic material that explicitly and
realistically depicts a variety of non-consensual and injurious practices was outlawed (in
England and Wales) by the Criminal Justice and Immigration Act 2008. Similar provisions
also appear in the Criminal Justice and Licensing (Scotland) Act 2010, with the notable
addition of obscene pornographic images which realistically depict rape or other nonconsensual penetrative sexual activity. The Criminal Justice and Courts Bill, currently
before Parliament, would extend to England and Wales the definition of the offence of
possession of extreme pornographic images to include rape. We welcome the
Government’s decision to include pornographic depictions of rape in the definition of
extreme pornography. It has been illegal to publish such images for many years;
outlawing their possession is long overdue.
35. Evidence we received from the Home Office also considered another area within the
inquiry’s terms of reference: tackling material intended to promote terrorism or other acts
of violence online. A specialist police unit, the Counter Terrorism Referral Unit (CTIRU)
proactively seeks and takes down UK-hosted material that breaches the Terrorism Act
2006. UK law has limited application in relation to the significant amount of material
hosted overseas; such material “is filtered from parts of the public estate”55 (the
Government has prioritised schools and some libraries). The filtering can currently be
circumvented by users changing their desktop settings; the Government is considering
how they can “further restrict access to illegal terrorist material (potentially at the network
level), further aligning with the IWF’s approach.”56
36. The evidence notes inconsistencies in approach among internet companies. “Whilst
engaging with industry to ensure that their own acceptable use policies are being applied
rigorously, we are also considering the Home Affairs Select Committee recommendation
of establishing a code of conduct for internet companies, distinct from their own terms and
54
Ev 105
55
Ev 106
56
Ev 106
18
Online safety
conditions, to improve the response to terrorist material (e.g. including ‘terrorism’ as a
category under unacceptable use).”57
37. There is clearly a need to obtain wider international consensus and cooperation in
relation to combating criminally obscene adult material and terrorist material and we
urge the Government to use all the influences it can bring to bear to bring this about
within a transparent, legal framework.
57
Ev 107
Online safety
19
3 Adult content
Nature and scale
38. While child abuse images attract worldwide opprobrium, there exists less consensus—
both nationally and internationally—on what other kinds of material adults might properly
access. Tensions can, and do, arise between the proscription of obscene material and
freedom of expression. In the United Kingdom, it is perfectly legal for adults to possess
images of explicit sexual activity of a kind that attracts an R18 certificate issued by the
British Board of Film Classification. The BBFC told us that that their guidelines “are the
result of extensive public consultation with over 10,000 people across the UK being
consulted during the most recent Guidelines consultation in 2013. Research demonstrates
that the public agrees with the BBFC’s classification decisions most of the time.”58
39. The BBFC’s written evidence provides a clear reminder of the qualitative difference in
the nature and accessibility of pornographic material between the off-line and online
worlds:
The BBFC removes any material from pornographic works which is potentially
harmful or otherwise illegal. As well as policing the border between legal and illegal
pornography, the BBFC polices the border between the strongest, hardcore
pornography, and the less strong, softcore pornography. The BBFC classifies
hardcore pornography as R18, which means that it may only be supplied through
licensed sex shops, as an extra precaution against underage viewing. However, the
risk of children accessing even the strongest legal pornography is far greater online.
In addition, there are fewer effective controls on the distribution online of
pornography which the BBFC would not classify at any category.59
40. Judging from the evidence we received, pornography was the category of adult content
that caused most concern. This could be an indication of its particular prevalence on the
internet. The Authority for Television on Demand (ATVOD) told us that five of the top 50
websites most commonly accessed from the UK are “tube” websites offering unrestricted
access to hardcore pornography videos. ATVOD also cited figures which suggest that
those five sites were (collectively) visited over 214 million times by UK internet users
during January 2013.60 John Carr of the Children’s Charities’ Coalition on Internet Safety
told us that “right now today within the UK, there is nothing there routinely that restricts
access to the most bizarre, the most violent and the most graphic types of pornography—
anybody can get it.”61 Pornographic material, much of it illegal, is but “two clicks of a
mouse”62 away.
58
Ev w14
59
Ev w15
60
Ev w135
61
Q7
62
Q7
20
Online safety
41. The NSPCC’s Childline service provides one indication of the harm done to young
people accessing pornography. According to the NSPCC:
During 2011–12, there were 641 counselling sessions where the young person
specifically mentioned being exposed to sexually indecent images. While these
incidents will not exclusively relate to online content, a large proportion of this
sexually explicit material will have been accessed through internet enabled devices.
Young people often told ChildLine that they felt guilty and disgusted about what
they had seen and were extremely worried about getting into trouble for accessing
these sites. ChildLine has also seen a growing trend of young people talking about
being addicted to online pornography.63
42. In January, we held a meeting at the House of Commons with eight young people,
some of whom had been harmed by exposure to adult pornography. One young man told
us how he had first encountered pornography at the age of eight; viewing pornography had
subsequently become a habit which distorted his picture of loving relationships. Another
participant told us how, as a teenager, she had been drawn accidentally into viewing
pornography from information in a fashion magazine; just one encounter had made her
feel ashamed and had affected her relationship with her father. Some girls told us how
boyfriends sometimes expected them to behave like “porn stars” and that the exchange of
sexually explicit material on mobile phones could lead to bullying.
The law
43. Online activity is subject to general offline legislation such as the Obscene Publications
Act 1959 and the Human Rights Act 1998. Publication of obscene material, including child
abuse images and extreme adult pornography, is illegal under the Obscene Publications
Act 1959 (which extends to England and Wales). An important point is that the definition
of obscene depends partly on the person who sees the material. “Legal” adult pornography
that has an R18 certificate, issued by the British Board of Film Classification, would likely
be classed as obscene if it was published in a way in which children could readily access it.
Both the Children’s Charities’ Coalition on Internet Safety and the Authority for Television
on Demand (ATVOD) cited case law (in particular, R v Perrin) in support of this assertion.
The test of obscenity in section 1 of the Act leaves little room for doubt in our minds:
For the purposes of this Act an article shall be deemed to be obscene if its effect or
(where the article comprises two or more distinct items) the effect of any one of its
items is, if taken as a whole, such as to tend to deprave and corrupt persons who are
likely, having regard to all relevant circumstances, to read, see or hear the matter
contained or embodied in it.64
44. The internet, where everyone with a connected computer is potentially a publisher, is
largely a free-for-all—though some audiovisual content is becoming subject to broadcaststyle regulation. The British Board of Film Classification engages extensively with the
public in reaching decisions as to what standards are acceptable. The R18 classification is
63
Ev 69
64
Section 1, Obscene Publications Act 1959
Online safety
21
given to restricted videos depicting explicit consensual sexual activity which, by definition,
excludes so-called “rape porn” and other illegal activities. R18 videos are only allowed to
be sold in licensed sex shops (to which only adults are admitted); they may not be supplied
by mail order. BBFC certification provides more than a useful yardstick as to content that
adults may legitimately choose to access:
The Government has recognised the dangers of extreme pornography and in 2008
made possession of extreme pornography an offence under the Criminal Justice and
Immigration Act. A BBFC classification is a defence against a prosecution under this
Act therefore purchasing a legally classified work is a protection against inadvertently
possessing extreme pornographic material. The BBFC regularly assists Local
Government trading standards officers in ensuring that pornographic material has
been classified by the BBFC and restricted for sale to licensed sex shops. However,
these methods of enforcement are not available online.65
45. Transposing accepted standards into the online context represents a challenge for
national institutions, not least because the internet is in many ways an international space.
Enforcement
46. Evidence from the Authority for Television on Demand (ATVOD) extensively explores
the potential use of the Obscene Publications Act. ATVOD draws attention to the current
Crown Prosecution Service guidance for prosecutors on the interpretation of the Obscene
Publications Act, which states that:
where children are likely to access material of a degree of sexual explicitness
equivalent to what is available to those aged 18 and above in a licensed sex shop, that
material may be considered to be obscene and subject to prosecution. This applies to
material which is not behind a suitable payment barrier or other accepted means of
age verification, for example, material on the front page of pornography websites and
non-commercial, user-generated material which is likely to be accessed by children
and meets the threshold. see R v Perrin, [2002] EWCA Crim 747. 66
47. ATVOD told us they could find no recent example of a prosecution being launched
under the above guidance. John Carr of the Children’s Charities’ Coalition on Internet
Safety told us:
That law is being honoured in the breach rather than in the observance. It is
principally because most of the publishers are based overseas and the British police
have not sought to extradite them or go after them, and that is a great pity. In our
evidence, we have made a number of practical suggestions about how we might try to
get at least companies that are based in Britain or operate from here to try to observe
that particular law. That is to say, “If you are going to publish porn, okay, that is your
65
Ev w15
66
Ev w137
22
Online safety
business, but please take concrete steps to make sure kids cannot get easy access to
it.”67
48. The Parliamentary Under-Secretary of State for Culture, Communications and
Creative Industries, Edward Vaizey said: “Obviously nothing should stop us doing the
right thing, in terms of prosecutions or clarifying the law. Nevertheless we do have to be
aware that a lot of these sites do not provide suitable identification as to who owns them,
and again ATVOD is suggesting that we clamp down on those sites by denying them their
financial support.”68 Jim Gamble did not see access to legal adult pornography primarily as
a law enforcement issue. Alluding to the “active choice” filtering solution, he told us: “I will
deal with the inappropriate material first because it is the easy one. I think inappropriate
material is a parental decision for those individuals who have duty of care of the young
people to make and I think active choice is absolutely right. If parents and others are
prompted to make a decision, I do not think you can do more than that. You are not going
to go out into their homes and look after their children for them.”69
49. ATVOD told us that they are working with the UK payments industry to design a
process which would enable payments from the UK to be prevented to foreign websites
which allow children to view hardcore pornography. However, there needed to be greater
clarity over the legality of providing unrestricted access to hardcore pornography.
50. ATVOD has an enforcement role in connection with a limited number of televisionlike services. It was designated by Ofcom in March 2010 to regulate the editorial content of
UK video on demand services. Its duties and powers derive from the EU Audiovisual
Media Services Directive, which was implemented in the UK via amendments to section
368 of the Communications Act 2003 (“the Act”). Under the Act, UK services which meet
the statutory definition of an on demand programme service (“ODPS”) must comply with
a limited number of statutory requirements which have been incorporated by ATVOD in a
set of Rules. ATVOD’s primary role in relation to protection of minors is founded in
Article 12 of the Directive, which forms ATVOD’s Rule 11: “If an on-demand programme
service contains material which might seriously impair the physical, mental or moral
development of persons under the age of eighteen, the material must be made available in a
manner which secures that such persons will not normally see or hear it.”70 ATVOD’s
evidence anticipates legislation in this area:
The DCMS strategy paper (“Connectivity, Content and Consumers”) published in
July 2013 sets out Government’s intention to legislate to make clear that R18equivalent material on ODPS must be provided only in a manner which ensures that
under 18s do not normally see or hear it. ATVOD considers that such legislation
would provide clarity for consumers and industry and better ensure deployment of
effective safeguards. Such legislation would also remove the possibility of ATVOD’s
consumer protection measures being undermined by a legal challenge.71
67
Q6
68
Q 222
69
Q 113
70
Ev w133
71
Ev w135-w136
Online safety
23
51. We believe that the existing obscenity laws already proscribe the publication of
adult material in ways that make it readily available to children. However, we are
concerned that no prosecutions have been brought despite the proliferation of
pornography sites which make no attempt to restrict access by children. We welcome
the Government’s declared intention to legislate to clarify the law in this area. However,
in the meantime, we urge the prosecuting authorities to use the existing law to crack
down on the worst offenders in order to put pressure on all suppliers of hardcore
pornography to make greater efforts to ensure that such material is accessible only by
adults.
52. A major difficulty lies in the fact that ATVOD and Ofcom can only regulate services
based in the United Kingdom. Furthermore, the requirements of the Audiovisual Media
Services Directive are interpreted differently in other Member States. The Dutch regulator,
for example, takes the view that hardcore pornography does not seriously impair the
development of under 18s and such services operating from the Netherlands are not
required to ensure that under 18s cannot normally see or hear them.72 In any case, the
majority of online hardcore pornography services available to children in the UK operate
from outside the European Union (most commonly from the USA).
53. The Government should seek agreement with other European Union Member
States to ban on demand programme services that make pornography readily available
to children. We further urge the Government to engage with other international
partners, particularly the USA, with the aim of securing a similar outcome more widely.
54. Evidence from Ofcom includes a reference to a research report, ‘Protecting audiences
in a converged world’.73 This research looked at public attitudes within the context of
convergence, in order to understand the public’s expectations for protection and how
content should be regulated in the future: viewers have high expectations of content
regulation on broadcast television, and associated video on demand and catch-up services,
less so for internet content accessed through devices such as PCs and laptops. Quite how
public expectations will develop as smart TVs and other manifestations of media
convergence become more commonplace remains to be seen.
55. An independent parliamentary inquiry into online child protection (April 2012),
chaired by Claire Perry, suggested that the Government should consider a new regulatory
structure for online content, with one regulator given a lead role in the oversight and
monitoring of internet content and in improving the dissemination of existing internet
safety education materials. Ofcom already has a role in relation to internet services, though
this is largely confined to promoting media literacy and performing research. Ofcom told
us: “We regulate television channels delivered over the internet and notified ODPS when
they are established in the UK; but we have no statutory powers to regulate any other
online content.”74 We believe that, as part of its existing media literacy duties, Ofcom
has an important role in monitoring internet content and advising the public on online
safety. However, we are anxious to avoid suggesting a significant extension of formal
72
Ev w134
73
http://stakeholders.ofcom.org.uk/binaries/research/tv-research/946687/Protecting-audiences.pdf
74
Ev 95
24
Online safety
content regulation of the internet. Among the unintended consequences this could
have would be a stifling of the free flow of ideas that lies at the heart of internet
communication.
Age verification
56. Providers of adult content can prevent children from accessing inappropriate and
harmful material by putting in place systems that require evidence of age. In this regard,
the mobile network operators have stolen a march over the internet service providers. A
majority of children have a mobile and an increasing proportion of them go online using a
mobile phone or smart phone. The mobile operators’ work in online safety is underpinned
by a Code of Practice that was first published in January 2004, ‘The UK code of practice for
the self-regulation of new forms of content on mobile’. The second edition of the code was
published in 2009 and the third (and current) edition in July 2013. The Code was the first
of its kind and was used as the boiler plate for similar codes introduced by mobile
operators throughout the EU.75
57. The Code covers a broad range of topics: commercial and internet content, illegal
content, malicious communications, spam communications and customer education. A
distinction is made between commercial and internet content. The Mobile Broadband
Group told us: “The mobile operators’ respective responsibilities for commercial content—
where they have contractual agreements in place with content providers—as against
general content on the Internet are different.”76 Any commercial content with an 18 rating
(determined by the British Board of Film Classification) is placed behind access controls
and subject to “a robust age verification process”77 (acceptable methods of which are set
out in the Code).
58. Age verification is clearly more challenging when accessing content does not involve a
direct financial transaction and where the users have an expectation of some degree of
anonymity. There is a stark contrast between the requirements on the online gambling
industry and those on other providers of online services to adults. As the Remote
Gambling Association highlighted:
The Gambling Act 2005 allowed for a wider range of advertising of gambling
products in Great Britain. To be able to advertise a gambling operator has to hold an
operating licence issued by the Gambling Commission, or an equivalent licence
issued by an EU gambling regulator or an overseas regulator which issues licences
with equivalent standards to the UK regulator. These licences require that before bets
can be settled the customer is over 18 and has had his or her identity verified.
As far as we are aware, no other adult service providers are required by law to ensure
that their customers are over the age of 18. This puts the regulated online gambling
industry in a different position to other e-commerce sectors. Because there are
75
Ev 86
76
Ev 86
77
Ev 86
Online safety
25
mandatory safeguards in place, but especially where children are concerned we
believe that the principles at least should be applied equally.78
59. Online gambling of necessity involves a financial transaction which makes age
verification relatively easy. There are, however, difficulties in relation to the use of other
services or consumption of content which do not necessarily involve direct payment.
British Naturism told us: “Age verification is used, for example, by gambling websites
where the possession of a valid credit card forms both the financial security and implicit
verification of the person’s right to gamble. But in the general case of free access to
unsuitable websites, it is unclear to us what mechanism could be devised that verifies the
age of the individual who has made initial access, but does not block the unverifiable access
by, say, another family member or friend to whom control of the device is passed.”79
60. The Christian social policy charity, CARE, acknowledged that many adult websites
require robust age verification to access 18+ content. “However there are many more
websites that provide such content for free without robust age verification. Moreover, the
business model of these websites can be driven by their click through rate as it relates to
advertising. This means that the more clicks a website receives, the more money they make,
disincentivising the owners of these websites from applying age verification.”80 The
Authority for Television on Demand (ATVOD) provided us with further details on the
business models of hardcore pornography services, most of which are operated from
overseas: “The most frequently accessed services use a variation on the You Tube business
model (and are consequently commonly referred to as “tube sites”). Such tube sites offer
significant quantities of unrestricted free hardcore porn videos as a shop window in order
to attract large number of viewers whose visits are monetised in a number of ways: by upselling to a premium version of the free service (offering a wider choice, longer videos,
better picture quality, etc); by driving traffic to other paid (pornographic) services operated
by the provider of the tube site; by charging on a ‘click through’ basis to affiliates whose
content is featured on a ‘try before you buy’ basis on the tube site; and by selling advertising
space (eg for ‘contact’ services or penis enlargement treatments).”81
61. Among the measures recommended by CARE is financial transaction blocking of adult
websites that do not put in place “robust” age verification procedures. ATVOD provided
us with the following examples of suitable age verification methods:
78
•
Confirmation of credit card ownership or other form of payment where mandatory
proof that the holder is 18 or over is required prior to issue.
•
A reputable personal digital identity management service which uses checks on an
independent and reliable database, such as the electoral roll.
•
Other comparable proof of account ownership which effectively verifies age82
Ev w7
79
Ev w120
80
Ev w129
81
Ev w134
82
Ev w133
26
Online safety
62. The Mobile Broadband Group argues that providers of age restricted services should
themselves be putting in place their own processes to protect minors. We agree. Providers
of adult content on the internet should take all reasonable steps to prevent children
under 18 from accessing inappropriate and harmful content. Such systems may
include, but will not necessarily be restricted to, processes to verify the age of users.
63. The Children’s Charities’ Coalition on Internet Safety suggested to us that more could
be done in relation to age verification. Their written evidence includes a suggestion that
refers to case law supporting the view that publishing adult material in ways that makes it
accessible by children is in breach of obscenity legislation: “Nominet should make
compliance with R v Perrin a condition of operating a .uk domain name e.g. if a site is to
publish pornography the operator must give a binding undertaking to put an effective age
verification process in place”.83 Nominet describes itself as “the trusted guardian of the .uk
domain name space, Nominet is responsible for the stability and security of one of the
largest internet registries in the world, with more than 10 million registered domain
names.”84 We have no reason to suppose that Nominet has either the resources or
inclination to police the internet. Age verification, while ideal, is not the only way of
preventing children from accessing unsuitable content. However, we believe that no
.uk site should offer unimpeded access to adult pornography to children. This should
be made a condition of registration.
Site blocking
64. The site blocking approach enabled by the Internet Watch Foundation, with the
necessary cooperation of ISPs, is one instrument—albeit a blunt one—aimed at preventing
access (by anyone) to online material. It has so far been applied mainly to images of child
abuse. Extending this approach to other material, including some adult sites, would face
challenges both of scale and cost. BCS 85, the Chartered Institute for IT, told us:
There has been much resistance to the Internet Watch Foundation’s widening its
remit to the other material in the Select Committee’s question, and BCS does not
believe that this is the way forward.
Some people say that more should be done, and imply, without saying so, that
content-based filtering should be used, so that more such material could be blocked.
This would require a major change in society’s attitude to censorship, as well as
primary legislation to enact fundamental changes to the Regulation of Investigatory
Powers Act. BCS does not believe that this is either feasible or desirable. 86
65. BT, a major ISP, has also expressed concerns:
In the absence of clear primary legislation from Parliament, or an EU-wide legislative
instrument, BT does not wish to police the internet beyond preventing access to
illegal material. To do so would set an unfortunate precedent in which an ISP would
83
Ev 65-66
84
www.nominet.org.uk/uk-domain-names
85
Formerly known as the British Computer Society
86
Ev w35
Online safety
27
become the arbiter of taste and decency in relation to online content. It is not for an
ISP to be placed in such a position.
Legal opinion informs us that filtering any internet material on home broadband or
public wi-fi may be illegal under RIPA 2000 and that this is so even if the purpose for
filtering is child protection, and even if the internet user has chosen to set up filters.
BT has raised with government the potential conflict between network level content
filtering and the Regulation of Investigatory Powers Act (RIPA 2000). We would
expect to receive clarity that our employees and or those of our wi-fi site partners
would not face a criminal prosecution under RIPA 2000 by offering filtering
activities to our wi-fi site partners for blocking unsuitable content from reaching
vulnerable individuals.87
66. The Internet Watch Foundation describes blocking of child abuse sites as “a short-term
disruption tactic which can help protect internet users from stumbling across these images,
whilst processes to have them removed are instigated.”88 Site blocking is highly unlikely
to be a suitable approach for adult pornography or violent material much of which is
legal (at least if it is unavailable to minors) and which is prevalent on the internet.
However, blocking should be considered as a last resort for particularly harmful adult
websites that make no serious attempt to hinder access by children.
Filters
67. An independent parliamentary inquiry into online child protection (April 2012),
chaired by Claire Perry, noted that “while parents should be responsible for monitoring
their children’s internet safety, in practice this is not happening”.89 The report went on to
recommend that the Government “should launch a formal consultation on the
introduction of an Opt-In content filtering system for all internet accounts in the UK” as
well as seeking “backstop legal powers to intervene should the ISPs fail to implement an
appropriate solution”.90 Following a subsequent Department for Education consultation,
the Government stopped short of proposing a default-on or opt-in filtering system, partly
on the grounds of practicality, the danger of blocking legitimate sites and the inability of
such systems to cut off other types of harmful material such as grooming and bullying.
The Perry report came on the back of a number of other studies that have looked at how
best to protect children, for example, Reg Bailey’s Letting Children be Children (June 2011)
and Tanya Byron’s Safer children in a digital world (March 2008). The latter report led to
the setting up of the UK Council on Child Internet Safety in September 2008. Our
predecessor Committee also published its report, Harmful content on the Internet and in
video games, in July 2008.
68. In its evidence, the Department for Culture, Media and Sport notes that 91% of
children live in households with internet access and that a greater proportion of children
aged 12–15 own smartphones than adults. The Government “understands that, first and
87
Ev w57-w58
88
www.iwf.org.uk/members/member-policies/url-list
89
Independent parliamentary inquiry into online child protection: findings and recommendations, April 2012, p5
90
Ibid, p8
28
Online safety
foremost, responsibility for keeping children safe online falls to parents and guardians;
however, Government is acting to ensure that parents have the tools and information they
need to be able to do so.”91 In particular, the Government has been working through the
UK Council for Child Internet Safety (UKCCIS) “to pursue a voluntary approach to child
internet safety and has called on industry to make the right tools available to allow parents
to protect children online.”92
69. In their submission, Intellect rehearsed the roles the technology industries have been
playing in the development of parental control tools such as filters. They also refer to a
“continual process to innovate and update these tools”. 93 Intellect added:
Thus it is clear that an irreversible momentum has developed across the industrial
ecosystem providing internet access to continually develop technology tools in
response to the fast evolving internet environment. It is this application of
technology innovation which will ensure the diverse set of tools needed to support a
safe online environment—not regulation which in contrast could freeze
innovation.94
70. In a speech to the NSPCC on 22 July 2013,95 the Prime Minister announced a range of
measures to tackle the “corroding” impact of online pornography on childhood. Some of
these would prevent children from being able to access (legal) pornography while other
measures would target child abuse images and the activities of paedophiles.
71. On access to pornography, Mr Cameron said that by the end of the year, familyfriendly filters would automatically be selected for all new broadband customers (unless the
account holder chose otherwise). Once installed, the filters would cover any device
connected to the customer’s internet account and only the account holder, who must be an
adult, would be able to change the filters. Internet service providers would be given until
the end of 2014 to contact existing customers and present them with an “unavoidable
decision” about whether or not to install family friendly content filters.
72. In March 2012, TalkTalk had become the first internet service provider to introduce a
genuine “unavoidable choice” for new customers when they signed up to TalkTalk
broadband, as per the recommendation of the Bailey Review. TalkTalk told us that
customers are asked to make a ‘yes’ or ‘no’ decision as to whether they want to filter access
to content that might be inappropriate for under 18s on their broadband connection or
not.96 TalkTalk then applies this to their internet connection as soon as it is live, and no
further action is required by the customer. The customer is also alerted by email and/or
text if any of the so-called Homesafe settings are changed – safeguards such as this aim to
ensure children aren’t changing settings without their parents’ knowledge.
91
Ev 108
92
Ev 108
93
Ev w140
94
Ibid.
95
https://www.gov.uk/government/speeches/the-internet-and-pornography-prime-minister-calls-for-action
96
Ev 83
Online safety
29
73. The Internet Service Providers’ Association told us: “The main consumer facing ISPs
are moving to a system where new and existing customers are presented with an
unavoidable choice of whether to apply filters or not. These filters cover the whole home ...
Some smaller consumer-facing providers are considering solutions that offer family
friendly filters but can be deployed on smaller scale and at lower costs. ISPA is currently
discussing this issue with its members.”97 Claire Perry told us that the top four ISPs are
introducing “unavoidable choice” filtering solutions. She said: “we would like the others to
commit to the same thing. They will ask you in different ways. Some might ask you when
you go to query your bill online, some may interrupt your browsing session, which is a first
for the UK to do that. This is the commitment that Ofcom will be monitoring. Every
household will be contacted and asked whether or not they would like to switch on the
filters, and the box, “Yes, please” will be pre-ticked.”98
74. We welcome the introduction of whole home filtering solutions that prompt
account holders with a choice to apply them. We encourage all internet service
providers to offer their customers this valuable service. Ofcom should monitor the
implementation of this filtering and report back on its level of success and adoption.
75. While greater use of filters is welcome, they should not be seen as a panacea. ATVOD
told us: “The usefulness of parental control software depends not only on its uptake but
also on its effectiveness. This is especially important lest parents who use parental control
software are lulled into a false sense of security about the extent to which their children
have been protected when using the internet.”99 ATVOD further cites EU Commission
research which suggests that the filters themselves when set to block “adult” content suffer
from relatively high rates of over-blocking (accidentally blocking non-adult sites) and
under-blocking (failure to block adult sites). Although the efficacy of parental controls may
have improved since that research was conducted in 2011, ATVOD told us it is clear that
both “over-blocking” and “under-blocking” still occur.100 The Internet Service Providers’
Association told us that filtering does have limitations and that over-blocking and underblocking of content “is inevitable”.101
76. When we held a meeting with eight young people in January, we heard varying views
on filtering. Some called for stronger filtering to prevent access to harmful material online,
particularly pornography. We were told of boys circumventing the filters in place in school
to access age-inappropriate content. However, others expressed concern that if filters were
too strong or inappropriately applied, young people could be prevented from accessing
websites offering advice on sexual health and online safety.
77. TalkTalk described its ‘Homesafe’ system as a “whole home parental controls system
that allows parents to protect every device connected to the home broadband and control
the types of websites their family is able to visit.”102 Homesafe has three features:
97
Ev 80
98
Q 207
99
Ev w137
100 Ev w138
101 Ev 81
102 Ev 82
30
Online safety
•
Kids Safe—parental controls that allow the account holder to block content they
don’t want to be accessed on their connection. There are nine different categories,
and customers can also choose to block other specific websites.
•
Virus Alerts—an alert system that blocks access to web pages infected with
malware103 and phishing sites.104
•
Homework Time—this allows parents to block social networking and online games
sites—common sources of distraction for children from homework—during a
specified time of day.
The Association for UK Interactive Entertainment describes as “deeply troubling” the
inclusion of games in the TalkTalk list. They told us of the “potential for this to do
significant collateral damage to the UK games industry.”105 We value the UK games
industry and the many educational and recreational benefits it provides to children. As
filtering technologies continue to develop, as they should, we trust parents will be
empowered to provide the supervision they want of what games their children play and
when.
78. In connection with filters applied by mobile network operators, the BBFC has a role in
calibration: providing advice to mobile operators on where to set their Internet filters. The
Mobile Broadband Group told us that processes also exist to remedy inadvertent or
unfounded over-blocking. The Group also told us: “The BBFC framework is binary—18 or
unrestricted. This is because 18 is the only age at which it is currently practical to
implement convenient, ubiquitous and robust on-line age verification. Stricter filters are
available in the market for parents that may want a narrower range of content for younger
users but these fall outside the Code.”106 At a recent seminar, Adam Kinsley, Director of
Policy, BSkyB, gave details of that ISP’s more granular filtering solution which has
additional age categories analogous to those used for cinema exhibition. 107
79. The Mobile Broadband Group reminded us that mobile operators have had network
filtering in place for nearly ten years. They added: “Great pressure has also recently been
put on the domestic ISPs and public wi-fi operators to do the same—and this is happening.
However, all these efforts would be complemented with the availability of better user
controls at operating system and device level. The UK, through UKCCIS and other
channels, should continue to examine closely what the manufacturers and operating
system providers are offering in the area of child safety and challenge them to be as equally
committed as the network providers.”108 We agree that the availability and performance
of filtering solutions must be closely monitored, both for efficacy and the avoidance of
over-blocking. It should also be easy for websites inadvertently blocked to report the
fact and for corrective action to be taken.
103 Malicious software
104 Phishing sites aim to obtain personal data by deception
105 Ev w92
106 Ev 87
107 Westminster eForum Keynote Seminar, Childhood and the internet – safety, education and regulation, 29 January
2014
108 Ev 88
Online safety
31
80. A recent report by Ofcom notes: “The provision of accurate content labels or metadata
by content providers would help filtering systems to categorise content correctly. However,
only a tiny proportion of websites are labelled in a way that allows easy categorisation for
the purposes of filtering.”109 Websites that provide adult content should signal the fact
clearly to enable filters better to take effect. A failure on the part of the operators of
such sites to do so should be a factor in determining what measures should be taken
against them.
81. One of the arguments given against filtering is the ease with which it can be
circumvented. According to Ofcom research, most parents report that they know enough
to keep their child safe online, but around half of parents continue to feel that their child
knows more about the internet than they do, including 14% of parents of children aged 3–
4. Ofcom also acknowledges: “In some cases, children will be able to bypass filters, either
by altering the filtering settings or by using tools to conceal the sites they are visiting from
the filtering software. The main mechanisms by which filters may be bypassed are through
the use of a VPN (virtual private network), which encrypts all internet traffic, and the use
of proxy sites.”110 Ofcom research has also established that 18% of children aged 12–15
know how to disable online filters or controls, but considerably fewer (6%) have done this
in the past year. Filters are clearly a useful tool to protect children online. Ofcom should
continue to monitor their effectiveness and the degree to which they can be
circumvented.
Media literacy and education
82. Filtering systems will in general fail to capture text and picture messages sent directly
between individuals. Andy Phippen, Professor of Social Responsibility in Information
Technology, Plymouth University, told us about some of the discussions he has had with
young people themselves:
Protection from access is an interesting concept. How can we protect them from
content they wish to access (which is certainly something I would observe from
talking to boys far more than girls)? This, again, was reflected in discussions recently
with a very mature group of 14–16 year old boys in a local school—one boy, who was
discussing the recent policy discussions around “opt-in” and filtering in the home,
made a very clear statement: “You will not prevent teenage boys from accessing
pornography”. He did not state this to be rebellious or controversial, he was stating it
from his observations of his peers. They access and share pornography and have
many ways of doing so.111
83. Comments such as this serve only to highlight the importance of media literacy and
education. The difficulty in entirely preventing access to age-inappropriate material
emphasises the importance of e-safety in the school curriculum and the availability of
advice to parents and carers. Such advice should include how to report harmful material.
Evidence we received from the DCMS did highlight the role of education in online
109 Ofcom Report on Internet safety measures: Strategies of parental protection for children online, 15 January 2014
110 Ibid.
111 Ev w111
32
Online safety
safety.112 From September 2014, the national curriculum will extend e-safety teaching to
pupils aged between 5 and 10 (11-16 years olds are already covered). The DCMS also
referred to several general educational tools and programmes: Think U Know (from
CEOP); the UK’s Safer Internet Centre (which has a hotline for internet safety
information); Get Safe Online (providing advice at the initiative of government, law
enforcement, businesses and the public sector); online resources, including Know IT All,
from the Childnet charity; the South West Grid for Learning; and ParentPort. The last of
these is a complaints portal that directs individuals to the relevant media regulator, or to
sources of advice for content that has no regulator responsible. Edward Vaizey told us:
“Ofcom was behind setting up ParentPort because it felt you needed a one-stop shop for
parents to go to get all the advice they needed.”113 We welcome the introduction of
ParentPort but believe Ofcom should seek to promote and improve it further. For
example, more use could be made of it to collect data on complaints concerning
children’s access to adult material.
84. We further recommend that Ofcom regularly reports on children’s access to agerestricted material, particularly adult pornography and the effectiveness of filters and
age verification measures. Ofcom is well-placed to fulfil this role given the work it does
on its Children and Parents: Media Use and Attitudes Report.
85. Childnet International commented on the need for further work on the education
front:
There is a need for ongoing educational and awareness work in this area ... As the
UK Safer Internet Centre, Childnet (along with the Internet Watch Foundation and
South West Grid for Learning) will be running Safer Internet Day 2014 which will
take place on 11th February. The theme of Safer Internet Day is “Let’s create a better
internet together”. This positive call to action provides all stakeholders with the
opportunity to reach out and positively work to empower internet users in the UK.
We are hoping a proposed industry-led awareness campaign, led mainly by the 4 big
ISPs, can combine with our work and help make Safer Internet Day 2014 even bigger
than last SID 2013, which reached 10% of the population, and led to 40% changing
their online behaviour as a result of the campaign.114
Safer Internet Day 2014 was subsequently celebrated by 106 countries, and early
indications are that it was a great success: over 25 million people were reached by the
“SID2014” Twitter hashtag alone.115
86. In their evidence, the sexual health charities FPA116 and Brook included the following:
“Sex and Relationships Education (SRE) guidance pre-dates recent and emerging issues on
technology and safeguarding, with no reference to addressing on-line safety, “sexting” or
pornography in SRE. Brook and FPA recommend that the Government should update SRE
112 Ev 109
113 Q 206
114 Ev w90
115 https://infogr.am/sid-2014-overview?src=web
116 Family Planning Association
Online safety
33
guidance for the modern era.”117 The young people we met in January were unanimous
that schools should be required to offer sex and relationships education. As one young
person put it, teachers are a sounder source of professional information on sex than friends
or the internet. The young people said providing them with the knowledge, tools and
confidence to navigate potential online dangers would ultimately be more beneficial than
technical measures. The NSPCC recently told us that the Government has committed to
emailing every school SRE advice developed by Brook, the sex education forum, and
others.118 We note comments on the state of, and access to, sex and relationships
education. We are aware this is a politically contested subject but believe the
Government should take into account the views of the young people who gave evidence
to us of the value and importance of good quality mandatory sex and relationship
education as policy develops. In the mean time, teachers have many opportunities to
use their professional judgement in advising children both on online safety and on
respect for each other. We believe there is scope for providing teachers with clearer
signposting of the advice and educational resources that are already available.
117 Ev w100
118 Ev 112
34
Online safety
4 Social media
Nature and scale
87. From chat rooms to Facebook, from Snapchat to Twitter, social media platforms play
to the human desire to keep in touch. Online social media provide new ways of interacting
and for modified ways of behaving. Everyone with a connected computer is now a
potential publisher and some people publish with too little regard for the consequences—
for others as well as themselves.
88. The most recent research 119 from the NSPCC shows that 28% of young people who
have a social networking profile have experienced something that has upset them in the last
year. These experiences include cyber-stalking, being subjected to aggressive or offensive
language, being sent sexually explicit pictures and being asked to provide personal or
private information. However, the greatest proportion of the group (37%) had experienced
“trolling”.120 Alongside this evidence that online bullying is clearly a problem for young
people, the latest Childline statistics show an 87% increase in 2012/13 in the number of
young people contacting the NSPCC for support and advice about being bullied via social
networking sites, chat rooms, online gaming sites, or via their mobile phones. The NSPCC
attributes this trend in part to the increasing ownership by young people of smartphones
and tablets.
89. The results of a July 2013 survey by the bullying prevention charity, BeatBullying,
provide further evidence of both the nature and scale of online bullying and the dangerous
sides of the internet:
•
One in five 12–16 year-olds have interacted with strangers online
•
More than a third of 12–16 year-olds go online most often in their own bedroom
•
One in five 12–16 year-olds think being bullied online is part of life
•
More than a quarter of 12–16 year-olds admitted to witnessing bullying online, but
only half of these did something about it
•
The primary reasons young people gave for not doing anything about the online
bullying was being worried about being bullied themselves or not knowing who to
speak to about it
•
Almost a quarter (23%) of 12–16 year-olds spend more than five hours a day online
during school holidays; more than double the number during term time (10%)
•
The majority (80%) of 12–16 year-olds said they feel safe online, compared to only
60% of the younger age group (8-11 year-olds). But worryingly, one in five (22%) of
12–16 year-olds said they think being bullied online is part of life
119 Ev 111
120 Trolling: the practice of posting deliberately inflammatory material.
Online safety
•
35
For those 12–16 year-olds who did do something about the cyber bullying, most
went to their parents for advice; however, only 38% of parents think their children
are at risk of being bullied online.121
90. Anthony Smythe of BeatBullying told us: “Our research would suggest that one in three
children have experienced cyber-bullying. More worrying is that you will find that one in
13 are subject to persistent cyber-bullying and that is what leads to the cases of suicide and
self- harm that we have seen over the recent summer months.”122 Our own conversations
with young people left us in little doubt as to the corrosive effect of bullying–often
perpetrated by “friends”.
91. Two of the best known social media platforms (there are many) provided both written
and oral evidence in the course of our inquiry: Facebook and Twitter. Written evidence
from Facebook begins by describing its mission “to make the world more open and
connected and to give people the power to share.”123 Facebook is a global community of
more than 1.15 billion people and hundreds of thousands of organisations. Facebook
works “to foster a safe and open environment where everyone can freely discuss issues and
express their views, while respecting the rights of others.”124
92. Twitter told us of their 200 million active users across the world and 15 million in the
UK alone; the platform now serves 500 million tweets a day. “Like most technology
companies we are clear that there is no single silver bullet for online safety, rather it must
be a combined approach from technology companies, educators, governments and parents
to ensure that we equip people with the digital skills they will need to navigate the web and
wider world going forward.”125
The law
93. Evidence from the DCMS makes the general point that behaviour that is illegal off-line
is also illegal online.126 Communications sent via social media are capable of amounting to
criminal offences in relation to a range of legislation, including:
•
Communications which may constitute credible threats of violence to the person or
damage to property.
•
Communications which specifically target an individual or individuals and which
may constitute harassment or stalking within the meaning of the Protection from
Harassment Act 1997.
•
Communications which may be considered grossly offensive, indecent, obscene or
false.127
121 Ev 75
122 Q 5
123 Ev 89
124 Ev 89
125 Ev 92
126 Ev 109
36
Online safety
94. The Director for Public Prosecutions published guidelines for prosecutors when
considering cases involving communications via social media. Relevant legislation
includes: Malicious Communications Act 1988; section 127, Communications Act 2003;
Offences Against the Person Act 1861; Computer Misuse Act 1990; Protection from
Harassment Act 1997; Criminal Justice and Public Order Act 1994; section 15, Sexual
Offences Act 2003 (for grooming).
95. The DCMS cites data from the Crime Survey of England and Wales which shows that,
in 2011/12, 3.5% of adults (aged 16 and over) had experienced upsetting or illegal images.
1.4% had experienced abusive or threatening behaviour. Some of these experiences are
likely not to have met the criminal threshold.128
96. BeatBullying have argued for greater clarity in the law; they told us:
More than 1,700 cases involving abusive messages sent online or via text message
reached English and Welsh courts in 2012. However, cyberbullying is not a specific
criminal offence in the UK. Some types of harassing or threatening behaviour—or
communications—could be a criminal offence. These laws were introduced many
years before Twitter, Facebook and Ask.FM, and they have failed to keep pace with
the demands of modern technology. Unfortunately, serious cases of cyberbullying,
which have often resulted in suicide, have dominated our headlines in recent
months. That is why BeatBullying have been calling on the Government to review
current legislation and make bullying and cyberbullying a criminal offence so that
children and young people have the protection they need and deserve, at the earliest
opportunity, to avoid this escalation.129
97. BeatBullying’s evidence went on to cite the recent Anti-Social Behaviour, Crime and
Policing Bill as a possible vehicle for introducing bullying and cyberbullying as a form of
anti-social behaviour. Jim Gamble told us: “The Prevention of Harassment Act is bullying.
The Prevention of Harassment Act is trolling ... We need to ensure that the laws as they
exist, when they can be applied, are applied.”130 Any changes to legislation, including
consolidation of current laws, which clarify the status of bullying, whether off-line or
online, would be welcome. At the same time, much could be achieved by the timely
introduction of improved guidance on the interpretation of existing laws.
Enforcement
98. On Twitter, users “agree” to obey local laws. Twitter’s rules and terms of service “clearly
state that the Twitter service may not be used for any unlawful purposes or in furtherance
of illegal activities. International users agree to comply with all local laws regarding online
conduct and acceptable content.”131 A question that arises is how more could be done by
Twitter and other social media providers to assist in compliance with the law.
127 Ev 109-110
128 Ev 110
129 Ev 76
130 Q 116
131 Ev 92
Online safety
37
99. Facebook’s detailed Statement of Rights and Responsibilities (“SRR”) describes the
content and behaviour that is and is not permitted on its service. With respect to safety, the
SRR specifically prohibits the following types of behaviours:
•
Bullying, intimidating, or harassing any user.
•
Posting content that: is hate speech, threatening, or pornographic; incites violence;
or contains nudity or graphic or gratuitous violence.
•
Using Facebook
discriminatory.132
to do anything unlawful,
misleading,
malicious,
or
100. Both Facebook and Twitter have sensible terms and conditions attaching to the use
of their services. However, these should be made much clearer, explicit and visible.
People who might be tempted to misuse social media need to be left in no doubt that
abuses online are just as unacceptable as similar misbehaviour face-to-face.
101. Facebook encourages people to report content that they believe violates their terms.
“Report” buttons are “on every piece of content on our site.” “When we receive a report,
we have a dedicated team of professionals that investigate the piece of content in question.
If the content in question is found to violate our terms, we remove it. If it does not violate
our terms, then we do not remove it. We also take action, such as disabling entire accounts
(eg of trolls) or unpublishing Pages, if deemed necessary.”133 Reports are handled by the
User Operations team comprising hundreds of employees located in India, Ireland and the
USA. The User Operations team is separated into four specific teams covering safety, hate
and harassment, access and abusive content.
102. Facebook is aware that many under-13s are falsifying their ages to open accounts, in
violation of the Statement of Rights and Responsibilities. Often parents assist them in
doing so, something that Facebook’s Simon Milner has reportedly134 likened to allowing
younger children to view Harry Potter video works (some of which have a ‘12’ certificate).
Sinéad McSweeney of Twitter told us: “We do not collect age information on sign-up. I
think Twitter has established a reputation in the area of privacy. We minimise the amount
of information that we require from users to sign up so we do not collect age or gender or
other details about our users. Where it comes to our attention that somebody under the age
of 13 is using the platform, their accounts are removed.”135 She went on to imply that a
child under 13 would either have to be aware of this age rule, or read about it in Twitter’s
privacy policy.136
103. Claire Lilley of the NSPCC suggested: “Some of these sites need more human
moderators to look for the fake accounts. That is one part of it, but they also have very
sophisticated algorithms where they look at what language people are using, what sites they
are visiting and what they are talking about online. They can tell with quite a degree of
132 Ev 89
133 Ev 89
134 Facebook admits it is powerless to stop young users setting up profiles, Guardian, 23 January 2013
135 Q 166
136 Q 169
38
Online safety
sophistication a lot about the individuals. Twitter, for example, has a minimum age of 13,
but when you sign up to Twitter, you do not have to put in your date of birth. There is
nothing to stop you. Twitter would say that what it does is to look at its algorithms to spot
the children who are under 13 and therefore potentially on a site that is aimed at an older
audience and containing information, posts and so on that are not suitable for their age. Its
argument is that it uses very sophisticated algorithms, so I think there is a lot more that the
sites could do.”137
104. Twitter’s age-verification process could at best be described as algorithmic and
reactive; non-existent might be a more accurate description. Given that Facebook and
Twitter are aware of the extent to which their services are accessed by younger children,
we expect them to pay greater attention to factoring this into the services provided, the
content allowed and the access to both. The same applies to other social media
companies in a similar position.
105. BeatBullying told us that BeatBullying.org is the only e-mentoring and social
networking site to be endorsed by CEOP. “We strongly believe that our approach to online
safety must be adopted by all internet providers if children and young people are to be safe
online.”138 This website checks all content prior to it being uploaded. As a general policy,
Twitter told us that they do not mediate content. However, there are some limitations on
the type of content that can be published with Twitter. These limitations include
prohibitions on the posting of other people’s private or confidential information,
impersonation of others in a manner that does or is intended to mislead, confuse, or
deceive others, the posting of direct, specific threats of violence against others, and
trademark and copyright infringement.139 Twitter told us that users can mark their own
tweets as sensitive which by default means a warning message is posted to anyone
wishing to view these. This is a good reminder that self-restraint and self-regulation are
crucial aspects of any enforcement regime in the online world.
106. In spite of reassuring words from Facebook and Twitter, it is clear that these
platforms, in common with other social media providers, could do far more to signal
the unacceptability of abuse and to stamp it out when it arises.
107. Offensive communications via social media that do not cross the threshold into
criminal behaviour should, the Government expects, be dealt with expeditiously by the
social media companies.140 We agree. Social media providers should follow the examples
of Facebook and Twitter in having appropriate terms and conditions. We believe there
is significant scope for such providers—including Facebook and Twitter—to enforce
such conditions with greater robustness.
137 Q 22
138 Ev 73
139 Ev 92
140 Ev 111
Online safety
39
Reporting
108. A service for reporting all hate crimes online was launched by the police in April 2011.
The website, called True Vision, is supported by all forces in England, Wales and Northern
Ireland and can be accessed at www.report-it.org.uk. All reports of incitement to racial
hatred content hosted in the UK previously reported to the Internet Watch Foundation
(IWF) should now be reported directly to True Vision. True Vision takes reports about:
racist or religious hate crime; homophobic and transphobic hate crime; disability hate
crime; bullying and harassment; domestic abuse. The National Centre for Cyberstalking
Research commented: “The True Vision initiative for hate crime reporting is an excellent
example of a simple and transparent reporting mechanism, but it needs to be more widely
publicised.”141
109. Social media providers generally offer reporting mechanisms, with varying degrees of
user-friendliness and degree of follow-up. In serious cases, the access providers will also get
involved. TalkTalk told us that they investigate any abusive or threatening comments
posted on sites by their customers when provided with the log information that supports
the complaint. In severe cases, relevant data will be disclosed to the third party solicitor, on
receipt of a fully sealed court order from a UK court.142
110. As noted above, Facebook told us about their ‘report’ buttons. Twitter also told us
that they have now introduced a similar reporting facility. Twitter said that reports that are
flagged for threats, harassment or self-harm are reviewed manually. Twitter advises users
to report illegal content, such as threats, to local law enforcement and refers to working
closely with the police in the UK.143 Stella Creasy MP—who has herself been subject to
bullying and threats via social media—argued for the introduction of an “online panic
button system” to alert sites like Twitter to an emerging problem.144 She told us how she
had been subjected to graphic threats and harassment on Twitter over the course of two
weeks.145 Even this was “just a fraction” of what had been endured by Caroline CriadoPerez who had been receiving 50 rape threats an hour. These threats evidently started for
no reason other than Ms Criado-Perez’s successful campaign to keep female representation
on English bank notes. In January, two people were jailed for their roles in the abuse to
which Caroline Criado-Perez was subjected.146 Another individual has recently been
charged with sending malicious communications to Stella Creasy MP. 147 All were charged
under section 127 of the Communications Act 2003.
111. The NSPCC have suggested that providers of social media services should provide a
range of options for users to report material, with extra support for younger users. They
add that default privacy settings should be the highest possible and there should be
141 Ev w144
142 Ev 85
143 Ev 93
144 Q 70
145 Q 61
146 “Two jailed for Twitter abuse of feminist campaigner”, Guardian, 24 January 2014
147 “Man charged over MP Stella Creasy tweets”, BBC News, 23 January 2014
40
Online safety
adequate use of human moderators.148 Claire Lilley of the NSPCC told us that, even if
children report bullying to social networking sites, “they often feel like nothing is being
done as a result of that and that no action is being taken, so children, when it is happening
to them, [are] feeling extreme vulnerability and humiliation, and a sense of helplessness.” 149
These comments were borne out by one of the teenage girls we talked to in January who
told us that, with Facebook, it was hard to get bullying material taken down or blocked;
when it was eventually removed, the damage had already been done. Twitter continues to
be criticised for not doing enough to combat abusive and threatening behaviour online, 150
even in the wake of the limited and tardy corrective action it took following last year’s case
involving Caroline Criado-Perez and Stella Creasy MP.151
112. Anthony Smythe of BeatBullying said:
I would like more transparency of the websites to hear from the big websites about
how many cases of cyber-bullying are reported to them each year. What did they do
about them? How quickly did they respond to those cases? What support and help
did they offer? It is about having a bit more accountability. They will say that they
probably do that, and if you spend five hours on the internet, you might find that
information annexed somewhere. I would like that information signposted on their
main websites so that parents and young people can have access and understand the
website that they are using.152
113. Stella Creasy MP told us: “One of the other things I have asked the companies to do is
publish their data about the numbers of reports of abuse they get and the numbers of the
concerns so we can get a question of scale.”153
114. Social media providers should offer a range of prominently displayed options for,
and routes to, reporting harmful content and communications. They need to act on
these reports much more quickly and effectively, keeping the complainant and—where
appropriate—the subject of the complaints informed of outcomes and actions.
115. Ofcom should monitor and report on complaints it receives, perhaps via an
improved ParentPort, regarding the speed and effectiveness of response to complaints
by different social media providers.
Advice and support
116. Anthony Smythe of BeatBullying told us: “What is the greatest concern for children
and young people—and now for adults—is the feeling that they are helpless and are
148 Ev 73
149 Q 10
150 “Ex-footballer Collymore accuses Twitter over abusive messages”, BBC News, 22 January 2014
151 Q 61
152 Q 14
153 Q 66
Online safety
41
hopeless in terms of getting advice and support. They do not know where to go to.” 154
BeatBullying develops these points in their written evidence:
Everyone involved with children’s and young people’s use of the internet—parents,
schools, service providers, organisations and children themselves—has a shared
responsibility for online safety. That is why in April 2013 BeatBullying launched a
campaign for better anti-bullying protections called Ayden’s Law. The campaign
calls for a national strategy to tackle cyberbullying and would set out how the
voluntary and community sector, parents and schools would be equipped to (a)
protect the children in their care from harm online and (b) educate and equip
children about internet safety and responsible digital citizenship so that they
understand the issues for themselves.
Any approach to online safety must ultimately be about shaping attitudes and
changing behaviors as much as it is about teaching techniques for staying safe or for
anything else.155
117. Claire Lilley said: “I would say that what bullying on social media comes down to is
about behaviour. We can wave the long arm of the law at children, but what we need to do
is to educate them about the impact of the behaviour in which they are engaging on people
who are at the receiving end of it. We need to do much more to educate them to build their
resilience, both when they are on the receiving end, but also to build their empathy and
their sense of respect for other children.”156
118. The Home Office told us that they undertake to “work with DCMS to ensure we are
linked into initiatives such as Safer Internet Centre and Get Safe Online, which provide
internet safety information and advice alongside a wealth of internet safety resources for
schools and information for parents and children.”157
119. Social media companies could, and in some cases do, provide resources and funding
for educational initiatives. For example, Simon Milner of Facebook, referred to support
given to the South West Grid for Learning which is “particularly helpful”158 for schools
and teachers. He also indicated that a request for funds would be listened to “with a very
open mind.”159 We also heard evidence from the Government, Facebook and Twitter of the
value of the helpline operated by the UK Safer Internet Centre, which operates on
minimum and time-limited funding from the European Union. We believe it is in the
interests of social media platforms, if they wish to avoid a more regulatory approach,
to put their money where their mouths are and provide more funding for the valuable
work being done on internet safety by voluntary organisations and charities.
154 Q 5
155 Ev 73
156 Q 12
157 Ev 107
158 Q 138
159 Q 150
42
Online safety
120. A good deal of advice on the safe use of social media is available already. This
should be signposted more clearly for teachers, who are likely to be in the front line
when it comes to dealing with bullying both in the playground and in the online world.
Anonymity
121. The cloak of anonymity is a useful one for a dissident or free-thinker to wear; but it
can also mask the bully and the criminal. Evidence from Dr Claire Hardaker, a Lecturer in
Corpus Linguistics, identifies anonymity as one of the factors that can lead to harmful
behaviour online (others include detachment and entertainment). She notes: “the internet
offers a perceived anonymity that has no real parallel offline, and this appearance of
invisibility encourages the user to feel that they can do unpleasant things with a highly
reduced risk of suffering any consequences.”160 She goes on to question the feasibility of
removing anonymity:
In a nutshell, this is borderline impossible, if only because it is unenforceable, and
unworkable. Even if all countries agree to legally mandating online identity
disclosure (unlikely) the costs of setting up, administrating, and then enforcing it
would be staggering. Further, we need only consider the risks inherent in having a
child’s name, age, location, etc. available online to realise that online identity
disclosure would actually create more dangers than anonymity currently averts.161
122. These views appear at odds with those of John Carr of the Children’s Charities’
Coalition on Internet Safety. Referring to the “abuse of anonymity” he emphasised the
importance of being able to trace individuals; this would require social media providers to
take greater steps to verify the identity of their account holders. 162 He said: “So the
requirement on internet service providers would be to verify that the individual who has
just signed up with them is not in fact a dog, but an actual human being with an actual and
verified address where they can be reached. That alone would act as a major brake on a lot
of the bad behaviour.”163
123. The Open Rights Group told us: “It is too easy to assume that tackling anonymity
online is a simple solution to abusiveness.” The Group added:
In fact, people are usually not truly ‘anonymous’ when they are online. People leave
all sorts of information that can identify them. It is sometimes possible to use this
information to identify somebody with varying levels of confidence—even if the
person posts messages as an ‘anonymous’ or ‘pseudonymous’ user. For example an
ISP may try to ‘match’ an IP address with one of their subscribers. There are various
legal powers that in some circumstance, require Internet companies to disclose this
data, and which permit the use of it in various contexts for the purposes of trying to
identify a user.164
160 Ev w2
161 Ev w4
162 Qq 22-26
163 Q 24
164 Ev w126
Online safety
43
124. Nicholas Lansman of the ISPA said: “People can attempt to hide themselves online,
but there are technical ways in which they can be discovered.”165 Claire Perry MP referred
to a particularly tragic case when she told us: “I was encouraged with Ask.fm—having
spent a lot of time with Hanna Smith’s father, who was one of the young women who did
indeed commit suicide—that company did set up a facility where users could choose to be
anonymous, but you would know if the user was anonymous when you were exchanging
information with them.”166
125. Anonymity is not just a cloak for cowards who bully; it is used by others to disguise
their criminal activities. In January of this year, the National Crime Agency announced
that 17 Britons had already been arrested as a result of Operation Endeavour, spanning 14
countries. This particular case involved the live streaming of child abuse in the Philippines
for viewing across the world. The prevalence of child abuse images on the internet and the
associated activities of paedophiles provide just one of the starkest of reminders that
keeping children safe off-line includes keeping them safe online too.
165 Q 110
44
Online safety
Conclusions and recommendations
1.
We believe that the Government should, in due course, consolidate the law around
child abuse images into a single Act of Parliament with a view to providing even
greater clarity for the purposes of law enforcement and deterrence. (Paragraph 9)
2.
Given the worldwide nature of online crime, we recommend that the Government
press for wider international adoption of both the Budapest and Lanzarote
Conventions. The Government should ratify the Lanzarote Convention as soon as
practicable. (Paragraph 10)
3.
We recommend that the Government examines whether adequate resources are
being deployed to track down online paedophiles in sufficient numbers to act as a
meaningful deterrent to others. If not, additional funding should be provided to
recruit and train a sufficiently large number of police officers adequate to the task.
(Paragraph 17)
4.
CEOP has been increasingly effective not least because it is not solely a criminal
justice organisation: its education and social care work has also been very important
in increasing public understanding of the problem of child abuse and in offering
means of countering abusers. We therefore recommend that CEOP continues to
publish an annual review which includes an assessment of its ongoing contribution
to all three elements of its mission—education, social care and criminal justice.
(Paragraph 18)
5.
We welcome the increasing use of alerting tools to identify individuals who seek out
child abuse and other illegal material online provided these tools are deployed in
ways that do not unduly compromise the privacy of the law-abiding majority.
(Paragraph 23)
6.
We very much welcome the commitment by the Internet Watch Foundation to
embark on proactive searching for online child abuse images. The sooner these can
be found and removed, the better. However, we are concerned that seven additional
staff might prove woefully insufficient to achieve substantial progress towards what
must be an important intermediate goal: the eradication of child abuse images from
the open internet. (Paragraph 29)
7.
Search engines and other internet service providers have a vital role in ensuring that
access to online child abuse images is prevented and deterred. We expect the
Government to monitor closely their degree of commitment and success and to
consider the introduction of legislation should they fall short of reasonable
expectations. (Paragraph 33)
8.
We welcome the Government’s decision to include pornographic depictions of rape
in the definition of extreme pornography. It has been illegal to publish such images
for many years; outlawing their possession is long overdue. (Paragraph 34)
9.
There is clearly a need to obtain wider international consensus and cooperation in
relation to combating criminally obscene adult material and terrorist material and
Online safety
45
we urge the Government to use all the influences it can bring to bear to bring this
about within a transparent, legal framework. (Paragraph 37)
10.
We believe that the existing obscenity laws already proscribe the publication of adult
material in ways that make it readily available to children. However, we are
concerned that no prosecutions have been brought despite the proliferation of
pornography sites which make no attempt to restrict access by children. We welcome
the Government’s declared intention to legislate to clarify the law in this area.
However, in the meantime, we urge the prosecuting authorities to use the existing
law to crack down on the worst offenders in order to put pressure on all suppliers of
hardcore pornography to make greater efforts to ensure that such material is
accessible only by adults. (Paragraph 51)
11.
The Government should seek agreement with other European Union Member States
to ban on demand programme services that make pornography readily available to
children. We further urge the Government to engage with other international
partners, particularly the USA, with the aim of securing a similar outcome more
widely. (Paragraph 53)
12.
We believe that, as part of its existing media literacy duties, Ofcom has an important
role in monitoring internet content and advising the public on online safety.
However, we are anxious to avoid suggesting a significant extension of formal
content regulation of the internet. Among the unintended consequences this could
have would be a stifling of the free flow of ideas that lies at the heart of internet
communication. (Paragraph 55)
13.
Providers of adult content on the internet should take all reasonable steps to prevent
children under 18 from accessing inappropriate and harmful content. Such systems
may include, but will not necessarily be restricted to, processes to verify the age of
users. (Paragraph 62)
14.
We have no reason to suppose that Nominet has either the resources or inclination
to police the internet. Age verification, while ideal, is not the only way of preventing
children from accessing unsuitable content. However, we believe that no .uk site
should offer unimpeded access to adult pornography to children. This should be
made a condition of registration. (Paragraph 63)
15.
Site blocking is highly unlikely to be a suitable approach for adult pornography or
violent material much of which is legal (at least if it is unavailable to minors) and
which is prevalent on the internet. However, blocking should be considered as a last
resort for particularly harmful adult websites that make no serious attempt to hinder
access by children. (Paragraph 66)
16.
We welcome the introduction of whole home filtering solutions that prompt account
holders with a choice to apply them. We encourage all internet service providers to
offer their customers this valuable service.
Ofcom should monitor the
implementation of this filtering and report back on its level of success and adoption.
(Paragraph 74)
46
Online safety
17.
We value the UK games industry and the many educational and recreational benefits
it provides to children. As filtering technologies continue to develop, as they should,
we trust parents will be empowered to provide the supervision they want of what
games their children play and when. (Paragraph 77)
18.
We agree that the availability and performance of filtering solutions must be closely
monitored, both for efficacy and the avoidance of over-blocking. It should also be
easy for websites inadvertently blocked to report the fact and for corrective action to
be taken. (Paragraph 79)
19.
Websites that provide adult content should signal the fact clearly to enable filters
better to take effect. A failure on the part of the operators of such sites to do so
should be a factor in determining what measures should be taken against them.
(Paragraph 80)
20.
Filters are clearly a useful tool to protect children online. Ofcom should continue to
monitor their effectiveness and the degree to which they can be circumvented.
(Paragraph 81)
21.
We welcome the introduction of ParentPort but believe Ofcom should seek to
promote and improve it further. For example, more use could be made of it to
collect data on complaints concerning children’s access to adult material. (Paragraph
83)
22.
We further recommend that Ofcom regularly reports on children’s access to agerestricted material, particularly adult pornography and the effectiveness of filters and
age verification measures. Ofcom is well-placed to fulfil this role given the work it
does on its Children and Parents: Media Use and Attitudes Report. (Paragraph 84)
23.
We note comments on the state of, and access to, sex and relationships education.
We are aware this is a politically contested subject but believe the Government
should take into account the views of the young people who gave evidence to us of
the value and importance of good quality mandatory sex and relationship education
as policy develops. In the mean time, teachers have many opportunities to use their
professional judgement in advising children both on online safety and on respect for
each other. We believe there is scope for providing teachers with clearer signposting
of the advice and educational resources that are already available. (Paragraph 86)
24.
Any changes to legislation, including consolidation of current laws, which clarify the
status of bullying, whether off-line or online, would be welcome. At the same time,
much could be achieved by the timely introduction of improved guidance on the
interpretation of existing laws. (Paragraph 97)
25.
Both Facebook and Twitter have sensible terms and conditions attaching to the use
of their services. However, these should be made much clearer, explicit and visible.
People who might be tempted to misuse social media need to be left in no doubt that
abuses online are just as unacceptable as similar misbehaviour face-to-face.
(Paragraph 100)
Online safety
47
26.
Given that Facebook and Twitter are aware of the extent to which their services are
accessed by younger children, we expect them to pay greater attention to factoring
this into the services provided, the content allowed and the access to both. The same
applies to other social media companies in a similar position. (Paragraph 104)
27.
Twitter told us that users can mark their own tweets as sensitive which by default
means a warning message is posted to anyone wishing to view these. This is a good
reminder that self-restraint and self-regulation are crucial aspects of any
enforcement regime in the online world. (Paragraph 105)
28.
In spite of reassuring words from Facebook and Twitter, it is clear that these
platforms, in common with other social media providers, could do far more to signal
the unacceptability of abuse and to stamp it out when it arises. (Paragraph 106)
29.
Social media providers should follow the examples of Facebook and Twitter in
having appropriate terms and conditions. We believe there is significant scope for
such providers—including Facebook and Twitter—to enforce such conditions with
greater robustness. (Paragraph 107)
30.
Social media providers should offer a range of prominently displayed options for,
and routes to, reporting harmful content and communications. They need to act on
these reports much more quickly and effectively, keeping the complainant and—
where appropriate—the subject of the complaints informed of outcomes and actions.
(Paragraph 114)
31.
Ofcom should monitor and report on complaints it receives, perhaps via an
improved ParentPort, regarding the speed and effectiveness of response to
complaints by different social media providers. (Paragraph 115)
32.
We believe it is in the interests of social media platforms, if they wish to avoid a more
regulatory approach, to put their money where their mouths are and provide more
funding for the valuable work being done on internet safety by voluntary
organisations and charities. (Paragraph 119)
33.
A good deal of advice on the safe use of social media is available already. This should
be signposted more clearly for teachers, who are likely to be in the front line when it
comes to dealing with bullying both in the playground and in the online world.
(Paragraph 120)
48
Online safety
Formal Minutes
Thursday 13 March 2014
Members present:
Mr John Whittingdale, in the Chair
Ben Bradshaw
Tracey Crouch
Paul Farrelly
Steve Rotheram
** *
Draft Report (Online Safety), proposed by the Chair, brought up and read.
Ordered, That the draft Report be read a second time, paragraph by paragraph.
Paragraphs 1 to 125 read and agreed to.
Summary read and agreed to.
Resolved, That the Report be the Sixth Report of the Committee to the House.
Ordered, That the Chair make the Report to the House.
Ordered, That embargoed copies of the Report be made available, in accordance with the provisions
of Standing Order No. 134.
** *
[Adjourned till Tuesday 18 March at 10.00 am
Online safety
49
Witnesses
Tuesday 15 October 2013
Page
John Carr, Secretary, Children’s Charities’ Coalition on Internet Safety, Claire
Lilley, Senior Analyst, NSPCC, Anthony Smythe, Managing Director,
BeatBullying
Ev 1
Susie Hargreaves, Chief Executive, Internet Watch Foundation, and Peter
Davi es, Child Exploitation and Online Protection (CEOP) Centre
Ev 11
Tuesday 29 October 2013
Stella Creasy MP
Ev 20
Nicholas Lansman, Secretary General, Internet Services Providers’
Association, Dido Harding, Chief Executive Officer, TalkTalk Group, Hamish
Macleod, Chair, Mobile Broadband Group
Ev 28
Mr Jim Gamble, Independent Chair, City and Hackney Safeguarding
Children Board
Ev 36
Tuesday 19 November 2013
Simon Milner, Policy Director, UK and Ireland, Facebook and Sinéad
McSweeney, Director, Public Policy, EMEA, Twitter International Company
Ev 44
Tony Close, Director of Content Standards, Licencing and Enforcement,
Ofcom and Claudio Pollack, Group Director, Content Consumer and External
Affairs Group, Ofcom
Ev 52
Rt Hon Damian Green MP, Minister of State for Policing, Criminal Justice
and Victims, Mr Edward Vaizey MP, Parliamentary Under-Secretary of State
for Culture, Communications and Creative Industries and Claire Perry MP,
Special Advisor to the Prime Minister on Preventing the Commercialisation
and Sexualisation of Childhood
Ev 57
50
Online safety
List of printed written evidence
1
Children’s Charities’ Coalition on Internet Safety
Ev 65
2
NSPCC
3
BeatBullying
Ev 73
4
Internet Watch Foundation
Ev 76
5
Internet Service Providers’ Associ ation
Ev 79
6
TalkTalk
Ev 82
7
Mobile Broadband Group
Ev 86
8
Facebook
Ev 89
9
Twitter
Ev 92
10
Ofcom
Ev 94
11
Home Office
Ev 103
12
Department for Culture, Media and Sport
Ev 107
Ev 68; Ev 111
List of additional written evidence
(published in Volume II on the Committee’s website www.parliament.uk/cmscom)
1
Dr Peter Dawe OBE
Ev w1
2
Dr Claire Hardaker
3
Remote Gambling Association
4
Ann Farmer
Ev w10
5
Kirsty Hopley
Ev w10
6
Dr Peter Nelson
Ev w11
7
British Board of Film Classification (BBFC)
Ev w13
8
Arqiva
Ev w18
9
James Griffiths
Ev w22
10
Christopher G. H. Thomas
Ev w22
11
John Edgar Cameron Lowe
Ev w23
12
Norfolk Library and Information Service
Ev w23
13
Phil Alexander
Ev w25
14
Ben Hardwick
Ev w25
15
John R Edwards
Ev w26
16
Chris Evershed
Ev w27
17
EE Limited
Ev w28
18
BCS The Chartered Institute for IT
Ev w32
19
Dudley Friend Clayson
Ev w36
20
PAPY RUS Prevention of Young Suicide
Ev w37
21
EU Kids Online
Ev w38
22
Rolf Smith
Ev w43
23
Mediawatch-UK
Ev w45
Ev w2 ; Ev w7
Ev w7
Online safety
51
24
South West Grid for Learning
Ev w47
25
British Sky Broadcasting Limited (‘Sky’)
Ev w51
26
Big Brother Watch
Ev w53
27
BT
Ev w56
28
Russell Hopwood
Ev w59
29
Stuart Taylor
Ev w60
30
Richard Browning
Ev w61
31
Timothy Michael Johnston
Ev w62
32
Lloyd Johnston
Ev w62
33
John Carruth
Ev w63
34
Rob Barry
Ev w64
35
Barry Saltmarsh
Ev w65
36
Safermedia
Ev w66
37
John Reiner
Ev w69
38
Family Online Safety Institute
Ev w69
39
Terry James
Ev w72
40
Film Distributors’ Association
Ev w73
41
Information Commissioner’s Office
Ev w74
42
Virgin Media
Ev w77
43
The Police Federation of England and Wales
Ev w84
44
End Violence Against Women Coalition
Ev w85
45
Childnet International
Ev w88
46
Ukie
Ev w92
47
The Children’s Society
Ev w96
48
Eric and Esme Bricknell
Ev w97
49
Stonewall
Ev w98
50
FPA and Brook
Ev w100
51
Michael J Smaldon
Ev w109
52
Prof Andy Phippen
Ev w110
53
Ahmadiyya Muslim Community UK
Ev w113
54
Sex Education Forum
Ev w117
55
Net Children Go Mobile
Ev w119
56
British Naturism
Ev w120
57
Open Rights Group
Ev w122
58
Russell Pillar
Ev w126
59
CARE (Christian Action Research and Education)
Ev w127
60
Malcolm Holmes
Ev w130
61
Microsoft UK
Ev w131
62
The Authority for Tel evision on Demand
Ev w132
63
Intellect
Ev w139
64
Telefonica UK Ltd
Ev w142
65
National Centre for Cyberstalking Research
Ev w143
66
Ethos Capital Ltd
Ev w145
52
Online safety
List of Reports from the Committee during
the current Parliament
The reference number of the Government’s response to each Report is printed in brackets after the
HC printing number.
Session 2012-2013
First Report
The Gambling Act 2005: A bet worth taking?
HC 421
Second Report
Racism in football
Third Report
Library Closures
Fourth Report
Football Governance Follow-Up
Fifth Report
Scrutiny of the Draft Public Bodies (Merger of the
Gambling Commission and the National Lottery
Commission) Order 2013
Sixth Report
Pre-legislative scrutiny of the draft Gambling
(Licensing and Advertising) Bill
HC 905
First Special Report
Football Governance Follow–Up: Government
Response to the Committee's Fourth Report of
Session 2012–13
HC 156
First Report
Scrutiny of the draft Public Bodies (Abolition of the
Registrar of Public Lending Right) Order 2013
HC 506
Second Report
Scrutiny of the draft Public Bodies (Merger of the
Gambling Commission and National Lottery
Commission) Order 2013
HC 256
HC 89
HC 587
HC 509
HC 1104
Session 2013-2014
Second Special Report Scrutiny of the draft Public Bodies (Merger of the
Gambling Commission and National Lottery
Commission) Order 2013: Government Response to
the Committee’s Second Report of Session 2013-14
HC 646
Third Report
Supporting the creative economy
HC 674
Fourth Report
Nuisance Calls
HC 636
Third Special Report
Supporting the Creative Economy: Government
Response to the Committee’s Third Report of Session
2013-14
HC 945
Fifth Report
Pre-appointment hearing with the Government’s
preferred candidate for the Chairman of Ofcom
HC 933
cobber Pack: U PL: COE1 [SO]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 1
Oral evidence
Taken before the Culture, Media and Sport Committee
on Tuesday 15 October 2013
Members present:
Mr John Whittingdale (Chair)
Mr Ben Bradshaw
Angie Bray
Tracey Crouch
Philip Davies
Paul Farrelly
Mr John Leech
Steve Rotheram
Mr Gerry Sutcliffe
________________
Examination of Witnesses
Witnesses: John Carr, Secretary, Children’s Charities’ Coalition on Internet Safety, Claire Lilley, Senior
Analyst, NSPCC, and Anthony Smythe, Managing Director, BeatBullying, gave evidence
Q1 Chair: Good morning. This is the Committee’s
first session examining issues around online safety and
I would like to welcome John Carr of the Children’s
Charities’ Coalition on Internet Safety, Claire Lilley
of the NSPCC and Anthony Smythe of BeatBullying.
Perhaps you could begin by just giving us your view
of the scale of the problem now, because obviously
this has been an issue that has been of concern for
some time. There are measures that have been taken
and there are more measures that are potentially in
prospect, but it would be helpful for you to tell us first
of all what the volume is, how many people are
involved and how easy it is to access images of child
abuse or inappropriate material. Who would like to
start?
John Carr: I will go first. The question of volume is
absolutely central to this whole debate. Like with all
areas of crime, it is impossible to get any precision as
to the real volumes. There are various straws in the
wind that give us reliable indicators. The NSPCC, for
example, did a series of FOI requests to all the police
forces in England and Wales asking them to report
back to them how many child abuse images they had
seized in arrests that they had carried out in the twoyear period ending April 2012. In the time frame that
the NSPCC was working to, only five forces replied,
but those five forces reported that they had seized 26
million images in that two-year period. I spoke to a
professional statistician and he looked at them—sorry,
go ahead.
Q2 Chair: On the 26 million, just to clarify in my
own mind, is that 26 individuals with 1 million images
each, or is it a million individuals with 26? How does
it work?
John Carr: That is an extremely good point—I was
going to get to it later. The numbers themselves can
sometimes conceal more than they reveal. There was
one man, for example, in Cambridgeshire who was
arrested with 5 million images on his machine. It is
not uncommon to find individuals now with a million
or more images. You are absolutely right; it is possible
that those 26 million images represented the activity
of no more than half a dozen highly obsessive
individuals who were downloading large quantities. It
is very unlikely. Just to finish on the point about
numbers, because I do think they tell a story—albeit
not the whole story, definitely—if you gross that up
to assume broadly similar levels of activity by all of
the police forces in England and Wales, that would
have led you to a number in the region of 300 million
that would have been seized by all police forces in
England and Wales in that period.
Now, putting that on one side for the moment, we then
have the revelations by Peter Davies, who I know you
are going to see later today, on ITN at the end of May
this year when he acknowledged that between 50,000
and 60,000 individuals had been identified by CEOP
as being involved in downloading child abuse images.
The period was not very clear, but none the less the
key point that Davies made in that TV interview—I
think it was very courageous of him to do it—was that
that is simply beyond the capacity of the police to deal
with this in any reasonable way. At no point since
records began—for these purposes, by the way, that
was around the mid-1980s—have British police ever
arrested more than about 2,500 individuals for child
abuse image-related offences. If you do the maths and
assume that there are no new offences committed
relating to images, the last individual on that list of
50,000 to 60,000 individuals who are known about
now would not be arrested until 2032.
The answer is not that we simply throw our hands up
in despair and say, “Nothing can be done.” Something
can be done, but it must involve industry in deploying
ever-smarter, better technical tools to deal with it,
because even if we were living in times of
superabundance as opposed to times of austerity, with
those kinds of volumes and those kinds of numbers of
offenders, it is impossible to imagine how we would
ever be able to, or want to, employ the right number
of police officers, and to have the right number of
prisons, courtrooms, judges and all those things to
deal with the volumes that we now know exist. We
have to find better technical solutions for addressing
some of this.
Chair: Thank you. Just before I bring in the other
two witnesses, we have two quick questions on what
you said.
Q3 Mr Leech: Mr Carr, in relation to the scale of the
problem, I think people would be amazed by the sheer
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 2 Culture, Media and Sport Committee: Evidence
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
numbers that you are talking about. Do you get a
sense that the problem is massively on the increase,
or is it simply that the detection of the problem is now
far more straightforward and we are becoming more
aware of what has always been a problem of this
scale?
John Carr: It could be a bit of both. Certainly,
obviously, we are lot better now at detecting it; we
have better tools for spotting it. If we can’t arrest
them, at least we know they are doing it, so that
definitely accounts for part of it. Internet penetration
is not going to go much further, or it does not have
much further to go in this country, so I doubt that
mere growth will lead to increase, but we are getting
better at detecting it. Around the world, that is a
different picture. There are lots of countries that are
coming online only now for the first time, where we
can expect to see similar patterns developing. That, of
course, does in a way pose a threat here, because if
the volumes are increasing elsewhere on the internet,
they could in the end none the less find their way to
the UK.
Q4 Tracey Crouch: Mine is a very quick one. With
reference to the 26 million images, what percentage
were of the level 5, I think it is, which is serious, and
how many were level 1? Can you give some sort of
breakdown of the kind of levels?
John Carr: I do not know exactly. The data was not
broken down in that way, but I would guess,
particularly with volumes of that kind, that they would
probably correspond roughly with the kinds of images
that the IWF is seeing in general, so a very substantial
proportion. I think 70% of the images that IWF deals
with depict children under the age of 10 being
sexually abused so, almost by definition, a very high
proportion of those will be level 5 or level 4—the
worst kind.
Q5 Chair: Sorry to have cut you off, but let me come
back now to Claire and Anthony.
Claire Lilley: Thank you. I just want to add to what
John said about the freedom of information requests
that the NSPCC did, and just to emphasise that that
was five forces, none of which had a large
metropolitan base, so that is not including London,
Manchester or Birmingham, where the figures are
likely to be even higher. I just add that caveat as well.
John has talked a lot about child abuse images. What
I would like to touch on is the other issues and
bringing the children’s voice into the session in terms
of what the NSPCC hears via ChildLine from
children. In 2012–13, we had almost 11,000
counselling sessions from children where the main
thing that they wanted to talk to us about was a
problem that they had had in relation to the internet.
That was a 65% increase on the previous year. In
terms of what children are telling us that they are
experiencing, much greater volumes of children are so
distressed that they are phoning ChildLine to talk
about a problem on the internet. That can be anything
from online bullying, so-called sexting or online
grooming—all those sorts of issues covered within
that.
Specifically in relation to online bullying, for
example, that saw an 87% increase in contacts from
children over the last year. That corresponds with a
4% decrease of wanting to talk about face-to-face
bullying. We are seeing that children contacting
ChildLine about bullying is massively increasing, and
obviously it is an issue about the conduct of children,
but the internet and access to smart phones, tablets
and all these things are facilitating and exacerbating
the problem.
Chair: That probably brings us neatly on to Mr
Smythe.
Anthony Smythe: Yes. I represent an anti-bullying
charity and my main interest is peer-on-peer abuse,
cyber-bullying and trolling. Our research would
suggest that one in three children have experienced
cyber-bullying. More worrying is that you will find
that one in 13 are subject to persistent cyber-bullying
and that is what leads to the cases of suicide and selfharm that we have seen over the recent summer
months. What is the greatest concern for children and
young people—and now for adults—is the feeling that
they are helpless and are hopeless in terms of getting
advice and support. They do not know where to go to.
Our work with practitioners would suggest that a lot
more needs to be done in strengthening roles and
responsibilities and clarifying that so that we have an
approach to tackle cyber-bullying and bullying, and
realise that cyber-bullying does not always happen
online. It sometimes can reflect what is happening in
communities, schools and vice versa now. Children
experience cyber-bullying online, and it is reflected in
the playground, so you have to see those as
overlapping circles with a great deal of movement
between the two.
Claire Lilley: May I come back? Sorry, I meant to
mention as well that I know the Committee is also
looking at the impact of online adult content on
children’s behaviour, and I wanted to mention that
ChildLine has seen a 70% increase in the number of
boys contacting the service because they have been
upset by online pornography specifically, so seeing
pornography and being distressed or feeling guilty
because they have watched it, or feeling somehow
dirty or ashamed because they have seen it, whether
accidentally or deliberately. Also there is an increase
in children being referred to our services that we run
around the country that are for children with harmful
sexual behaviour and watching pornography, with
easy access to online pornography, is playing a part in
their sexualised behaviour.
Q6 Chair: Can you just say a little more about what
evidence you have of the sort of harm that is done by
children either accidentally coming across, or
deliberately coming across, material that is clearly
adult and should not be seen by them?
Claire Lilley: The evidence base, it is fair to say, is
growing. In the past 10 years, we have seen such an
explosion in access to the internet and online
pornography being freely available within a few
clicks—it is very easy to get hold of. While the
evidence base is still growing, the NSPCC would say
that a precautionary approach would be the best
approach, because essentially we are conducting a big
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 3
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
experiment at the minute with our young people
where they have access to all this online pornography
very easily and we do not know what the influence
will be. We have evidence from ChildLine that
children are being pressurised to watch porn by their
peers and they don’t necessarily want to do it. Girls
are telling us that they are being pressurised to act
out sex scenes from porn movies that they or their
boyfriends might have watched, and there is the
sexualisation of women and girls that is being caused,
we believe, by these large numbers of porn being
easily available over the internet.
John Carr: I will just come in on that. I think the
word “pornography” when we discuss it in this
context is in many ways very unhelpful, because
certainly for people of a certain age, the word
“pornography” is mentioned and images of a Playboy
centrefold or pictures of ladies dancing without their
bras on the beach or something spring to mind. Then
this whole thing starts going about, “Well, aren’t you
just being a bit old-fashioned or prudish?” Whatever
view you take about that, the sort of material that we
are in fact talking about that is available at the click
of a mouse on the internet is light years away from
anything of that kind. This is, in essence, violence. It
is sadistic, and very often quite bizarre.
To go back to Claire’s point about the biggest
experiment we have ever undertaken, porn has been
around since time immemorial. Anybody who has
been to Pompeii and visited the House of the Vettii
can see it there in stucco on the wall. But never
before, ever, including in the 1960s and 1970s, when
a lot of the studies on porn were initially done, has
the kind of material that is there now been available
24 hours a day, seven days a week in graphic colour,
with full sound and free. In our evidence, we refer to
the decision in the case of the Crown v. Perrin, which
was decided in the Criminal Court of Appeal in 2002.
This was a case of a man who was publishing extreme
pornography on a website and had taken no steps to
try to keep it away from kids. He had no age
verification system—nothing in place—and he went
to jail, but the judges said, “It is okay for you to
publish porn. Nobody is saying you cannot or should
not, but if you do, you must take steps to keep kids
away from it”.
That law is being honoured in the breach rather than
in the observance. It is principally because most of the
publishers are based overseas and the British police
have not sought to extradite them or go after them,
and that is a great pity. In our evidence, we have made
a number of practical suggestions about how we might
try to get at least companies that are based in Britain
or operate from here to try to observe that particular
law. That is to say, “If you are going to publish porn,
okay, that is your business, but please take concrete
steps to make sure kids cannot get easy access to it.”
Anthony Smythe: Can I just add some statistics? In
terms of bullying and cyber-bullying, the Department
for Education produced some statistics a number of
years ago that suggested that for those who are
persistently bullied, their attainment drops by on
average two GCSEs. Our research looked at persistent
bullying and the link to suicides, and what we found
when we looked back at the records was that 44% of
child suicides had bullying as a contributing factor.
That does not mean that they committed suicide
because they were bullied. It could be that they
suffered a bereavement and were bullied and cyberbullied because of the bereavement, and the totality of
what they were facing led to the suicide. But it is a
major contributing factor to child suicide—peer-onpeer abuse.
Chair: That neatly brings us on because Tracey
Crouch wants to explore this area.
Q7 Tracey Crouch: It does. I am just going to go
further into cyber-bullying, if I may, but can I just
follow up on a couple of the issues that John has
raised with regard to pornography? Could you just
explain to me: are there already certain restrictions in
place for some of the pornographic content that you
were referencing, such as, for example, the more
bizarre or perhaps violent images? Does that already
exist, or can you just go on to any pornographic
website and have easy access to the more bizarre
images?
John Carr: Two clicks of a mouse. In the UK, the
mobile phone networks have, since 1 January 2005,
all blocked access to adult content accessible either
over their own networks or the internet without you
first going through age verification process. If you are
accessing through a mobile phone in the UK, broadly
speaking, only adults should be able to access porn in
that way. But of course there are lots of other ways
that you can do it. Our wi-fi providers—all the larger
ones—have said that they are going to block access
to porn as well. That is a great achievement, by the
way. It is a world first, so we can be very pleased with
that. The only people who are not yet currently taking
steps to block access to those types of sites are the
ISPs. But, as we know, thanks to Mr Cameron’s
efforts, there are currently discussions going on
between all the large ISPs and the Government about
introducing steps to limit access to that type of
material. But right now today within the UK, there is
nothing there routinely that restricts access to the most
bizarre, the most violent and the most graphic types
of pornography—anybody can get it.
Q8 Tracey Crouch: But is the age restriction process
just a case of saying, “I am over the age of 18,” and
ticking the box?
John Carr: No. With the wi-fi, you will not be able
to lift it—it will be fixed. By the way, that will apply
only in public places where children and young people
are present, so if you go to Spearmint Rhino, a casino
or some place that is essentially wholly populated by
adults and access the wi-fi there, you will be able to
get anything that is on the internet. But ordinarily in
Starbucks, at the Odeon, in a railway station and so
on—places like that; public places where children and
young people are going to be—you will not be able
to get access to porn full stop, so you cannot prove
you are over 18 and get it lifted. But with the ISPs,
the process is being debated right now, and we do not
quite know how that is going to pan out yet.
Q9 Tracey Crouch: But it will not impact on those
accessing content at home?
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 4 Culture, Media and Sport Committee: Evidence
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
John Carr: Yes, it would. Sorry, I lost my thread
slightly. The mobile phone companies have quite a
robust age verification system. You have to go through
the same process that the online gambling companies
use, and you have either to be on the electoral roll or
to have a credit card that one of the credit reference
agencies can get. There are a number of bits of
information they collect from you as a mobile user to
determine that you are in fact an adult. What the ISPs
are going to do in the end at home, we do not know.
We have heard what the Government have said they
want them to do. We have not yet heard from them
what they are in fact going to do, and they have until
31 December to tell us. But they could introduce age
verification as well, if they wanted to, because it is
tried and tested technology.
Q10 Tracey Crouch: Thank you. Can I just go back
to cyber-bullying, please? We heard from Anthony
about the scale of the statistics and what they see as
the problem of cyber-bullying. I know that both the
NSPCC and yourself, Mr Carr, have referenced cyberbullying in your written submissions as well. Could
you just expand on what you are seeing from your
own organisations about cyber-bullying, please?
Claire Lilley: In terms of calls to ChildLine, we have
seen an 87% increase in the last year, and huge
volumes of children under extreme pressure, mainly
as a result of bullying from their peers. I think the
characteristics of online bullying that set it apart from
traditional bullying are that because children very
often have phones and access to the internet at home
and in their bedroom, it is not something that stops
when they go home and close the door. They no
longer have that sense of safety, so it is literally 24/7.
Some children are telling us that if they do not
respond to a message within 20 minutes at 4 am, the
bullying will increase—a sort of pack mentality
sometimes where it is no longer just six children
involved in the bullying, but perhaps the whole year
group or going out to the whole school, so the child
is feeling very vulnerable in terms of their exposure,
feeling a huge degree of humiliation, not knowing
who knows, whatever the rumour is, and feeling it is
totally beyond their control and that they are very
unable to do anything to combat it. If it is happening
on social networking sites, for example, it will quite
often be replicated via email and instant messaging,
and even if they report it to social networking sites,
they often feel like nothing is being done as a result
of that and that no action is being taken, so children,
when it is happening to them, feeling extreme
vulnerability and humiliation, and a sense of
helplessness.
Q11 Mr Bradshaw: Do you take up individual cases
with social networking sites at your organisation?
Claire Lilley: No, we do not, because the service that
ChildLine operates is confidential, so we never
disclose anything, except in the case that a child tells
us that they were about to commit suicide. Only then
would we breach that confidentiality.
Q12 Tracey Crouch: Section 127 of the
Communications Act describes the improper use of
the public electronic communications network. Do
you think that more could be made of the
Communications Act to combat bullying on social
media?
Claire Lilley: I would say that what bullying on social
media comes down to is about behaviour. We can
wave the long arm of the law at children, but what we
need to do is to educate them about the impact of the
behaviour in which they are engaging on people who
are at the receiving end of it. We need to do much
more to educate them to build their resilience, both
when they are on the receiving end, but also to build
their empathy and their sense of respect for other
children. There may be things that we could do in
terms of the law, but what we need to do is to build
in much more education of children about the impact
of their behaviour on others and get parents involved
in talking to their children about how they behave
online, how they conduct themselves and how they
treat other people online.
Q13 Tracey Crouch: Anthony, do you think there
should be further legislative change to combat
bullying?
Anthony Smythe: I completely agree. The first
priority is about prevention and education. Sometimes
that does not work, though, so the question is what do
you do next? What you want is a well-trained
workforce around the child—a workforce that
understands their roles and responsibilities and that
can provide that support. As Claire says, bullying is
24/7, so we need infrastructure around the child that
sees bullying from a child’s point of view. For too
long, we have seen bullying as a school issue or a
cyber-bullying issue. Children don’t see it like that.
They do not have an online and offline world; it is
just their world.
I would like the Government to reform their approach
to bullying in the same way as they have reformed
their approach to child protection, where they have
seen child protection from the child’s point of view
and had services rallying around that child. If you do
that, you will get good early intervention, and that
brings in the third sector, teachers and the police. If
we share the responsibility, we will reduce the
workload on one another because capacity is an issue.
In terms of the law, it is not fit for purpose. It does
not protect children from cyber-bullying; it does not
protect adults from cyber-bullying. The Children’s
Commissioner for Wales this morning, I believe, made
a recommendation to make cyber-bullying a criminal
offence. I would endorse that because I think we need
to give out a message that bullying is unacceptable in
our society, whether it be cyber-bullying or bullying.
Under the current legislation, the Communications
Act and what have you, the thresholds have been set
really high. I have not seen many cases in my time
where it has been used to address bullying, so I think
we need updated legislation that is fit for purpose.
Most of the legislation that deals with cyber-bullying
was drafted many years before the likes of Twitter and
Facebook. There needs to be a fresh look at this. There
needs to be a legal definition of bullying, cyberbullying and trolling because we all talk about these
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 5
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
things, yet we are often working off a different
definition. That needs to be embedded into legislation.
In answer to your question, yes, there need to be
stronger safeguards in place. There needs to be an
anti-bullying strategy and an anti-cyber-bullying
strategy for this country, which would set out the
measures for education, prevention, intervention and
support, and how we can strengthen the current
criminal justice system.
Q14 Tracey Crouch: Thank you. A final question:
do you think that social media providers could do a
little bit more to stamp out bullying? You saw the
recent very sad and tragic situation of the Ask.fm
suicide and I just wondered whether you think that
they should be doing more to try to prevent cyberbullying.
Anthony Smythe: A lot of the big ones have done a
great deal of work over recent years and that should
be recognised.
Each time my teams that go into schools do a straw
poll and ask, “What websites are you using?”—my
teams are quite young people, so they are relatively
up to speed with what is going on out there—they find
that children and young people are using new websites
and each time the dynamic of the website changes.
There is a new one now where a bit of chat is posted
straight away and deleted within a short period of
time. The new websites are playing catch-up a great
deal, and we heard a lot over the summer about
Ask.fm.
Ultimately what I would want from websites is for
them to know their customers. Why do people cyberbully? First, because they can; secondly, because they
haven’t had the education to understand the impact it
has on other people. How do we deal with the first
point, which is going back to John’s point about
having verification about the user’s identity? You can
go on to Twitter and set up any account, and you can
go on to all the other websites and by and large set up
an account that allows you to go in and say things
without fear of being suspended or caught. There
needs to be a lot more done at the point of registration
and a lot more investment into moderation of their
websites.
I would like more transparency of the websites to hear
from the big websites about how many cases of cyberbullying are reported to them each year. What did they
do about them? How quickly did they respond to those
cases? What support and help did they offer? It is
about having a bit more accountability. They will say
that they probably do that, and if you spend five hours
on the internet, you might find that information
annexed somewhere. I would like that information
signposted on their main websites so that parents and
young people can have access and understand the
website that they are using.
John Carr: Just to make a quick point, I hope.
Education and awareness—what goes on between a
child’s left ear and their right ear—are always going
to be the best defence for them in life in general,
including on the internet, no question. However, let
me just point you to the results of research published
by Ofcom in August or September, I think: 37% of
three and four-year-olds are now regularly going
online in the UK, with 9% of three and four-yearolds seemingly having their own tablets. Both those
numbers are going to go up, not down. There is a limit
to how much education and awareness can do in the
context of three and four-year-olds.
We have a choice. We can either wag our fingers and
say, “Parents, don’t be so irresponsible. Don’t let your
kids go online at that age,” or whatever, or we can
accept that that is the reality of modern parenting,
which therefore pushes us to looking towards the
industry again for more and better technical support.
Education and awareness are always the best possible
option, no question, but that will not work for very
young children and for vulnerable children. In that
context, I think we do need to see more and better
technical solutions coming into play to help.
Q15 Mr Bradshaw: Mr Smythe, you called for
cyber-bullying to be made a criminal offence. Forgive
my ignorance, but are there any other jurisdictions
where it is already a criminal offence, or where noncyber-bullying is a criminal offence?
Anthony Smythe: Yes. A number of countries around
the world are having this very same debate. Australia
and New Zealand are bringing in new legislation;
Canada is about to bring in new legislation; Ireland is
currently looking at it. There is an element of
everyone is playing catch-up. What I would like to
see is not just legislation around cyber-bullying, but a
criminal offence on bullying and cyber-bullying. As I
mentioned earlier, a child will be bullied in the
playground and that continues online. We need to
make that link because a child will not just see it as
cyber-bullying; they will see it as consistent,
persistent bullying. A number of countries are looking
at this. Every time I have looked at a review, the
recommendation always comes out to make cyberbullying a criminal offence. I think they are just going
through those procedures, so that is happening right
now.
Q16 Mr Sutcliffe: It was interesting to me that the
first mention of parents came in 34 minutes since you
started—that was the first mention of parents. A bit of
context to my contribution: I am 60 years old, and I
have children and grandchildren. My oldest
grandchild is 13. I am trying to come to grips, John
Carr, with the size and scale of this issue. I think that
Anthony’s survey was in terms of one in five 12 to
16-year-olds having interacted with strangers online,
and 33% of 12 to 16-year-olds going online in their
own bedrooms. The size and scale of the problem is
something that we need to get out into the open in
terms of where we are. I do believe that there is
something for the law to do, but I do think there is an
issue around parenting and also content.
Again, as a 60-year-old fuddy-duddy, with the
sexualisation and violence on TV now, I think the
watershed has gone by the bye. If you look at some
of the soaps—look at some of the storylines on
Coronation Street, Emmerdale and things like that—
and the violence and sexual innuendo, is there an issue
for society here around how we deal with these things
in terms of context and content? There are two
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 6 Culture, Media and Sport Committee: Evidence
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
questions: size and scale; and the inevitability of what
is going on in terms of the availability.
John Carr: On the wider point you make about the
sexualisation of society, I think with how what used
to be considered unusual is now considered normal or
typical, there has definitely been a shift, but however
far you look across that spectrum, we are a very long
away off accepting some of the stuff that is on
PornHub or YouPorn, for example, or some of the
more gross sites that are out there. They are a very,
very long away from anything that you might see on
a music video or at 10 am on Channel 4—or even at
1 in the morning on Channel 4. Of course the context
has shifted, and things are not considered to be quite
so outrageous today as they were so long ago, but the
kind of things that we have been concentrating on are
still way outside any spectrum of what would be
considered to be generally normal.
The size and scale is part of the problem. I have been
speaking to legislators and investigators as well,
which is what you are now. In the past, a lot of people,
when they have thought about what to do about the
internet, have said, “It is too difficult. It is global.
Nothing can be done”. A lot of people have a vested
interest in getting everybody to believe that is true.
They want people to believe that nothing can be done,
that Governments are impotent and that all we can do
in the end is to trust big business or all these internet
companies to do the right thing. But time and time
again, we see signs that they do not, and it is only
when they are pressed by legislative measures or
Governments very often, sadly, that they will act.
There are things that can be done. Even if every
country in the world does not do them at the same
time as we do, that is not a reason for us not doing
them. Leadership and showing an example is also an
important part of this whole equation. There are lots
of things that we did first in Britain. Look at what I
was mentioning earlier on wi-fi and mobile phones.
We were the first in the world to do that. Does that
make us mad? No. Other people will catch up; other
people will copy us. There is a normative aspect of
this, isn’t there? There is a question about what are
the right things or what we ought to expect companies
to do? We ought to articulate them and our
Government should press for those things to happen,
however difficult they might appear on the surface.
Claire Lilley: Just on the question of parents, there is
a very good tracker that Ofcom does of parents’ and
children’s attitudes to media. Of the parents of 12 to
15-year-olds, 63% said that they felt that their child
knew more about the internet than they did. I think,
particularly with older children, that that is a problem,
but it is a problem that we can address by encouraging
parents to think of the internet as an extension of their
parenting and that what they have to do is use their
common sense and the same principles and guidelines
that they would use in the so-called offline world and
just apply those from a very early age, as the child
grows, to the online world. We need to give parents
confidence and empower them to think that they can
deal with these issues, because the evidence does
show—work by Sonia Livingstone at the LSE—that
when parents lay down boundaries and guidelines for
their children, children will adhere to those. That is
where the point comes in again about educating
children about how to behave.
Anthony Smythe: Can I just add to that? In my paper,
I said that the most useful parental control is parental
responsibility. There is a need to get more information
and advice out to parents so that they can have those
discussions. They often feel intimidated when it
comes to tech and children. They need not, because
ultimately it is about, as Claire says, having those
discussions about empathy for other users online and
being responsible online. Often you find with cyberbullying that they don’t often realise the damage on
the victim—the person on the other end of the
bullying. When you address that, they start to change
their behaviours.
Q17 Mr Sutcliffe: On that, Anthony, in your survey,
22% of 12 to 16-year-olds thought that being bullied
online was part of life. It has become the norm, hasn’t
it, in terms of there being an acceptance of bullying
by that 20%?
Anthony Smythe: I found that statistic to be quite sad
because they have just given up and we need to make
bullying and cyber-bullying unacceptable. The reason
why people tend to troll and join in with the cyberbullying is because they feel it is okay to do it. Again,
it goes back to that programme about education, about
empathy for other people online and about the fact
that it will have an impact on that person. You may
not be the only person saying it; you may be person
No. 1,000 and they have probably been hearing that
abuse for six months or a year. That is why we see
the self-harm. That is why we see, unfortunately, in
recent months, the suicides. We do need to educate
people online. The internet providers have a role in
there in the social networking sites. They also have a
role pointing users to help and support that they can
get as soon as they make a referral to them. My
charity has counselled over 3,000 children over the
last year on bullying and cyber-bullying, but I have
not received many referrals from the big internet or
social networking sites—they come to us direct. That
is something we need to improve on.
Claire Lilley: I would just like to build on that by
saying that in terms of educating children about these
issues, it is definitely parents, but also the industry.
There is a lot more they could do in terms of the
number of human moderators they have enforcing
their terms of service, being very clear about what
their terms of service are, having terms of service
written in very basic language that is child-friendly—
all these things—and being much more transparent.
At the minute, we have no idea of how many reports
they are getting of online bullying or online grooming
per year. They could be much more transparent about
what they are seeing and what they are doing to deal
with it, and about what the rest of us could do in
partnership with them to help deal with it as well.
Q18 Angie Bray: If I could just continue a little bit
along the lines of my colleague, Gerry Sutcliffe, about
the role of parents in this, because what I see at the
moment is an army of people marching into family
life, in a sense, whether it is the Government or
organisations like your own. It is changing laws and
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 7
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
having to put filters on this, that and the other—which
is information to other people—all because
effectively, it would seem to me, you are saying that
there is a shift in the relationship between parents and
their children. Call me old-fashioned, but there was a
debate many years ago about whether children should
have televisions in their rooms precisely because of
what they might be watching that was not being
monitored. Fundamentally, isn’t the point that parents
should have the confidence—I understand what you
are saying about online and internet and all that—as
they have done in the past and make the time to take
an interest in what their children are doing in the
privacy of their own rooms, however intrusive they
might be worried that might seem to be?
Claire Lilley: I think that is the ideal scenario: parents
who are engaged, who are confident, who understand
their children and who make time for their children.
Unfortunately, that is not always the reality. Some
parents are not confident and don’t feel equipped, and
there are some children who, as a result, are
particularly vulnerable. There are also children who
do not have an active adult in their life, or perhaps
their parents might be substance misusers, for
example. There are quite a lot of children in that
bracket. I think we have a duty to spread the load.
Parents definitely have a big role, but there is a lot
that the rest of us can do. We have mentioned the
sexualisation of society. It is a broader problem than
just within families and the family unit; it is a problem
for the whole of society, so we all need to play a role
in that.
It is why we have had rules and laws about this type
of thing in relation to cinema, for example, and there
was for a very long time. In a way, some of the things
that we are describing are us as a society catching up
with the consequences and impact of the new
technology, and trying to recreate in the virtual space
rules and laws that we have had—not quite since time
immemorial, but certainly for a very long time. I do
not hear anybody arguing, for example, that we should
get rid of the rules about 18-rated movies and make
them open to everybody and anybody. In essence, all
we are doing—certainly when you talk about porn—
is trying to recreate in the reality for children’s lives,
which is a virtual one, the same rules that we have
applied for a long time in cinemas.
Another aspect of this as well, by the way, is the
knowledge gap. There was a debate many years ago
about whether it should be legal for parents to buy
their children a computer or a mobile phone until they
had passed a test like for a driving licence or
something to show that they had at least a rudimentary
understanding of the types of things they were
handing over to their kids. That never got anywhere.
Certainly the industry did not like the idea that this
could act as a brake on the sales of these brand-new
shiny devices. We absolutely put parents front and
centre, but what we also acknowledge is that,
particularly because of the speed at which these things
have changed, a lot of parents—even very dedicated,
loving, nurturing and supportive parents who would
fit any ideal paradigm that you could construct—feel
they need that little bit of extra help, and that is where
I think the state can play a role.
Q19 Angie Bray: I take that point, but fundamentally
the family unit is there. It is, most people would argue,
the most successful start for a child’s life. I just feel
that the stress is now all on what the state can do to
deal with this problem, but there is much less about
what can be asked of parents to take responsibility for
their own children. I wonder whether your messaging
should not start by talking about parents, rather than
talking about changing laws and applying more
regulation. Shouldn’t it start, where it possibly can—
I accept that it cannot always—in the home with a
proper relationship, where the parents expect to be
able to go in, knock on the door and say, “Hi. What
are you up to this evening?”
Claire Lilley: Yes, absolutely.
John Carr: First of all, a lot of this activity is not
taking place in the child’s room. Bear in mind that
internet access is now essentially ubiquitous. You can
walk out of your house with a tablet, a mobile phone
or a games console and connect to the internet
essentially anywhere. The things I was referring to
earlier about the wi-fi providers are an example of
them doing the right thing and essentially helping
parents to create an environment that they would want
their children to be in. What parent would be happy
with the thought that their child can go into a
restaurant or Starbucks and sit next to a dirty old man
looking at porn on his laptop? Nobody would, because
there is a reasonable expectation that that type of
material ought not to be in public view in places
where children and young people are going to be.
Q20 Angie Bray: What you are saying
fundamentally is that this new technology has
fundamentally altered the relationship between
parents and their children because it is—
John Carr: The same potential for control and
intervention simply does not exist. Some parents
think, “Oh, little Johnny is up in his room. He is safe.
He is upstairs. He is not out on the street playing with
the bad boys and the bad girls. He is not behind the
bicycle shed smoking. He is up in his room online.”
In other words, there is a fundamental gap in the
knowledge that we need, and a lot of this you can see
as a way of trying to help bridge that gap.
Anthony Smythe: I can only repeat what John just
said. Even the best parents need help and advice in
terms of addressing this issue and having those
conversations. Ultimately, I do not think it is for
parents alone to do this. Peer-to-peer support is
sometimes the best way to tackle online abuse. The
approach cannot be top-down, because top-down
approaches from Government probably will not work,
but we do need better leadership so that we can help
to clarify roles and responsibilities so that when a
child is cyber-bullied in the middle of their summer
holidays, either the child or the parents know where
to go to get support, to feel safe and for that cyberbullying to stop. I do not think that scenario is an
easy one for people to answer yet. The Department
for Education did a lot of good work in terms of
clarifying roles, responsibilities and powers for head
teachers to deal with bullying in schools. We have not
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 8 Culture, Media and Sport Committee: Evidence
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
seen the same in terms of cyber-bullying, and while it
would be good to get that leadership, it needs to
involve everyone—it cannot be top-down. It needs to
be a conversation; a joint plan, if you like.
John Carr: Sadly, many of the big internet companies
that are not British—they are not based here; their
jurisdiction is elsewhere—do need to be encouraged,
let us put it that way, to act. No business that I have
ever heard of willingly steps up and says, “Please
regulate me. I do not know what to do, so please
would you, Government or politicians, tell me what
to do?” They do not do that. Businesses will go on
ploughing their own furrow unless and until
something cuts across their pathway. That is what we
do, in essence. We try to raise these issues and that is
what Governments do as well.
Q21 Steve Rotheram: Just following on from Angie
Bray’s questions, it would be fantastic to believe that
all parents are responsible and do have that
relationship with their children. While I genuinely
believe that education and awareness are primarily the
ways to tackle the issue, that is not always the case.
There are parents who are absolutely fearful of IT and
do not understand it, and therefore do not understand
the content that can be accessed via it. But you also
mentioned that there are potential technical solutions
to it. There are flash screens that can come up that
say you are about to access a site of adult content or
whatever. There are restrictions that can be put on to
these, such as for anything that is over 18, you have
to opt in rather than out of it. There are blockages, I
believe, that you can do to adult content coming into
the home. I do not think this is all an either/or
solution.
John Carr: Correct.
Steve Rotheram: I think with the education, fantastic,
but there may well be other things that can be done.
What are the most effective ways and what more
could be done to curtail children’s access to adult
content or age-restricted material, including
pornography and graphic depictions of violence?
John Carr: We have never said that porn should not
exist or should not be on the internet. If it is legal,
there is no reason why it should not be published, but
companies that make pornography available should
take steps to make sure that kids cannot access it.
They should employ age verification. The British
gambling industry has done it with tremendous
success. We never hear of cases of kids getting online
and gambling. Why? Because Parliament, through the
Gambling Act 2005, introduced a law making it
obligatory. The Gambling Commission can give you
a licence to operate as an online gambling company
only if you prove to the commission that you have an
age verification system in place that works and that is
robust. That could be done in any other area, including
this one.
There are technical tools available, but they are not
being used—they should be. I do not want to argue
against myself because I am a great believer in
technical tools helping parents to put in place stuff
that they would want, even if they do not know that
it is what they would want, if you see what I mean.
Technical tools can do a great deal to supplement
good parental support and guidance. They will never
be 100% perfect. Nothing in life is 100% perfect. No
self-respecting parent would delegate their parental
responsibilities to a piece of software, but they
definitely can help, and in some situations that
software might be the only help that exists to keep
that child safer than it would otherwise be. There are
tools available. They can be done. They will not solve
everything, but they could do a lot to help.
Claire Lilley: Just to add to that, one of the arguments
we hear against the use of these parental controls and
against the active choice or the opt-in is that it will
lull parents into a false sense of security. If that is the
case, as John has said before, there has been an awful
lot of non-lulling going on for a long time. The
parents who are already using these parental controls
are the parents who are already engaged and therefore
are not complacent and probably already doing a
range of other things to support the use of those tools,
like having conversations with their children and
looking at their children’s browsing history and so on.
But for the parents who are not already having those
conversations and not already using the tools, the optin system will provide a nudge to make them think
about it and engage with this subject, so we think that
is valuable.
Anthony Smythe: My main concern with filtering is
that there is a danger it may filter out useful
information for children and young people, so it could
be information on sex advice or health conditions and
so on, such as anorexia.
John Carr: But that is bad filtering.
Anthony Smythe: That is bad filtering, but there is a
risk that it will happen. The most important thing we
can do regarding technology in terms of social
networking sites is for social networking sites to know
who is on their site at any one time and to have that
verification in place so that people cannot set up fake
websites, because it is those fake websites that allow
you go on and abuse others without fear of being
caught. If people knew that there were stronger laws
in place and that they could be traced online, they
would behave differently. The filtering is fine to a
certain degree, as long as it doesn’t block useful
information. What filtering will not do is prevent peeron-peer abuse.
Q22 Chair: How exactly do you envisage that
happening? Anybody can set up a Facebook account
with an alias just by creating an email address. What
more do you expect them to do?
Anthony Smythe: That is my concern—that it is too
easy at the moment. John has mentioned how it has
been done in other parts of the online world. We need
to look at that age verification. I know DFE was doing
a lot of work in looking at ways to take that forward.
Ultimately, I think that is a job for industry to take
forward. They are earning billions and billions of
pounds. They can find a way, I am sure, if they were
told to and this is where pressure from Government
would be useful. But as a starter for 10, it may be the
use of credit cards—that type of thing—so that you
know who that user is and that they are at a certain
age. At this moment in time, for example, in my
charity, we have a Twitter account for our resident dog
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 9
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
and that is how easy it is to set up these accounts.
There is no point threatening somebody that you will
suspend their account because they know they can set
up a new one straightaway. We just need to review
the whole situation.
Claire Lilley: Some of these sites need more human
moderators to look for the fake accounts. That is one
part of it, but they also have very sophisticated
algorithms where they look at what language people
are using, what sites they are visiting and what they
are talking about online. They can tell with quite a
degree of sophistication a lot about the individuals.
Twitter, for example, has a minimum age of 13, but
when you sign up to Twitter, you do not have to put
in your date of birth. There is nothing to stop you.
Twitter would say that what it does is to look at its
algorithms to spot the children who are under 13 and
therefore potentially on a site that is aimed at an older
audience and containing information, posts and so on
that are not suitable for their age. Its argument is that
it uses very sophisticated algorithms, so I think there
is a lot more that the sites could do.
John Carr: I will just make a fairly obvious point.
The abuse of anonymity lies behind most of the
enduring problems on the internet that we are
concerned about. I do not care if people sign in as
Donald Duck or Ronaldo or whatever. What matters
is not how you sign on, and not necessarily how you
present yourself in a public space on the internet, but
that you are traceable. If people were traceable and
knew that, should they cross a certain line, the cops,
the courts or whatever could fairly rapidly ascertain
who they were and reach out to them, I think that
would have a tremendously beneficial effect across the
internet in all sorts of different ways.
Of course immediately you then get thrown back at
you, “What about political dissidents in North Korea?
What about whistle-blowers? What about people
dealing with sensitive subjects of one kind or another
who need or can benefit from anonymity?” Two things
to say about that: I have no desire to make it more
difficult for political dissidents in North Korea to
overthrow their Government, if that is what they want
to do. I have no desire to make it harder for whistleblowers to do their public service or for people who
deal with sensitive information. But it cannot for ever
be a question of either you make your children safer,
or you make it impossible for political dissidents to
log on.
Q23 Chair: So how do you do it? What is it that you
would like to happen?
John Carr: One potential route would be to develop
a trusted third-party institution of some kind in whose
integrity one could have a very high degree of
confidence that could issue IDs that would not be
disclosed to anybody unless and until due process had
been followed. If you were dealing with a sensitive
health issue or you were a whistle-blower, or if you
were a political dissident in a particular country or
whatever—although God knows, after the revelations
of what the NSA gets up to with Mr Snowden,
whether you would ever trust the internet again
anyway—you would be able to be behind a shield of
anonymity without putting yourself entirely beyond
the pale.
Q24 Chair: So you have that, but what I am asking
is what are you expecting Facebook, for instance, to
do to satisfy themselves?
John Carr: Sorry, right. So the requirement on
internet service providers would be to verify that the
individual who has just signed up with them is not in
fact a dog, but an actual human being with an actual
and verified address where they can be reached. That
alone would act as a major brake on a lot of the bad
behaviour.
Q25 Chair: At the moment, I go on to Facebook and
call myself Donald Duck, I have created
[email protected] or whatever—
Steve Rotheram: No, I did Donald Duck.
Chair: Oh, Steve has Donald Duck. But whatever it
is, I have a Facebook account. What is that you want
Facebook to do before signing me up?
John Carr: What the gambling companies do now is
to say, “If you want to place a bet, that is fine; that is
what we are here for. We just need to check that you
are who you say you are”. They ask you for various
bits of information—four or five different pieces of
personal information that they confirm—and
assuming everything holds up they give you your
account. You can then call yourself Donald Duck; it
does not matter. That particular Donald Duck will be
traceable back to John Whittingdale, SW1 and you
will behave accordingly, one assumes.
Q26 Chair: Do you think there is any prospect of
Facebook or Twitter agreeing to do this?
John Carr: Not voluntarily, no, but with all the
revelations Mr Snowden has made and continuing
escalation in criminal behaviour online, and law
enforcement around the world feeling more and more
beleaguered and besieged, I do not think the current
situation is sustainable in the very long run. Will big
companies on the west coast of America seek to resist
this? Yes, they will. It will be a tussle of will between
political institutions and Governments as to who wins
that battle or how quickly it will be won, but I have
no doubt the current situation is not tenable in the
long run.
Q27 Steve Rotheram: Chair, I think there could be
an evidence session just on that one aspect of what
we are looking at. I have met Facebook, Twitter, the
police, the Home Office and the DPP on trolling as an
issue. I thought I had a simplistic solution whereby the
verification could be done with somebody’s national
insurance number, but it costs to go through the
verification process and the likes of Twitter and
Facebook are not prepared voluntarily to cost
themselves several billions of pounds because there
are literally billions of people who are on social media
platforms, so it would take something from the
Government to force their hand.
There are also issues of death threats and certainly
rape threats that happened to a Member of Parliament,
and the algorithm did not quite pick that up because
it was a little bit more subtle. There were issues about
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 10 Culture, Media and Sport Committee: Evidence
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
that, so that had to be monitored by an individual over
a two-week period and somebody was arrested for
that. The laws are in place. The laws around this are
very complex. We have picked up on seven different
pieces of legislation that can be covered for what
happens online and offline, so it might be time for
Parliament to look at this as an issue and to take it
seriously. There is something called the Trolling
Academy that promotes individuals specifically to
troll MPs, for instance. After this, who knows? We
might all get what I got last time I mentioned this,
which was lots and lots of Twitter feeds or whatever
they are called—tweets—and Facebook feeds on our
statuses attacking us. But there you go, that is part
and parcel.
If we could just move on to the issue about child
abuse images, given that these images are illegal
throughout the world, to what do you attribute the
apparent prevalence of the material that was
mentioned before? Was it 300 million pieces that are
known to exist in the UK alone?
John Carr: I think it is a profound shock to realise
that so many people are interested in that type of
material. There is an appetite for it and people go and
get it, and they circulate it. They are getting away with
it because we do not have the capacity to stamp it out.
Q28 Steve Rotheram: We do not have the capacity.
John Carr: We do not at the moment, for sure. I do
not know if Peter Davies has arrived in the room yet.
Look at Peter Davies’ interview on TV at the end of
May. He said he would like to arrest all 50,000 to
60,000 people who CEOP had identified as being
involved in exchanging or trading these images, but
that is not the situation we are in, so they go after
the—
Q29 Steve Rotheram: Is it lack of capacity or
resource? We have the ability, because we can detect
them.
John Carr: I suppose if Peter had 100,000 police
officers and an unlimited budget, he could probably
get them all, but we have not, and I do not see any
scenario where that is going to happen.
Claire Lilley: The internet has caused a huge growth
in the number of people who are looking at these
images. It has removed a lot of the key inhibitors that
would have once stopped men looking at it. You can
now get these images completely anonymously, they
are free or very affordable, and they are very easily
accessible. Because of those three things, we have
seen a growth in men—it is usually men—who once
would not have gone down this route, but now they
are looking at it. That is why the NSPCC believes
there is a need for a public education campaign. From
studies of offenders, their motivations and the reasons
why they have done it, quite a lot of them will
maintain that they have not done anything wrong, that
it was just an image and that they did not touch a
child, so they did not actually commit child abuse. We
need to educate people that these are not just images;
that by looking at these images, they are committing
an offence; that a child is being re-victimised every
single time an image is looked at; and that they are
potentially affecting their own sensitivity around the
issue and escalating their own ability to go down that
route and end up abusing in the physical sense
themselves. There is quite a lot of research now about
the crossover between non-contact and contact
offending and CEOP would put it at about 50%. There
are a couple of studies that put it at between 40% and
55%, which is a very high level of crossover. It is not
just looking at an image; it is much more dangerous
than that.
Q30 Steve Rotheram: You said that they are easily
accessible images. Does that in some way excuse the
fact that people might innocently stumble on them?
My understanding is that it is not easy to click on
something and then for these graphic images that we
are talking about, even though there are 300 million
of them, to be readily accessible—there are other
things that you would have to click on to. So if you
just put a name up or something and a whole list of
images came up, they would not be of the graphic
nature, surely, of child images.
John Carr: It depends. You can get pretty quickly to
some of them. Some of the ostensibly legal adult sites
will have sections with names like “jail bait” or
“barely legal” and very often that is a code for saying
if you keep going just a couple of clicks through there,
you will find stuff that is not legal and definitely is
jail bait. If you go on to peer-to-peer networks, you
can type in certain words and pretty quickly get at
them. You are right that you need a little bit of
determination, but not a great deal.
Claire Lilley: Anyone who innocently stumbles across
them accidently should be reporting them immediately
to the IWF. Lots of people do, though not everyone
by any means. I am sure the IWF will talk about that
in the next session.
John Carr: Just be clear, we are not blaming the
internet for this. People do things, not machines. What
the internet has done is open pathways that have
allowed people to walk down them in a way that just
was not there before. That is the shocking and
surprising thing—the numbers of people who have
decided to walk down that path.
Q31 Steve Rotheram: But internet providers cannot
abrogate responsibility completely for those images
being on their systems. Surely there must be some
people who could block those images.
John Carr: Absolutely; I 100% agree. We do have a
fundamental issue with the law, of course, which is
that the EU-wide e-commerce directive expressly says
that you have no legal liability or responsibility for
anything on your network unless and until you have
actual knowledge of its existence. Some companies
take that as a get out of jail card. They say, correctly,
“We are not accountable. We cannot be prosecuted or
pursued for failing positively to go forward to try to
block things.” To their credit, many choose to,
certainly in the UK, but not all do. That is because
our law says there is no liability for this material
online unless and until you have actual knowledge.
Q32 Steve Rotheram: Even though the content itself
is illegal.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 11
15 October 2013 John Carr, Claire Lilley and Anthony Smythe
John Carr: To be honest, they must know that that
illegal content is on their network, but unless they
have actual knowledge of it, they cannot be held
liable. I get that, up to a point—you cannot find
somebody guilty of something if they do not know
specifically that it is there—but I think it is a shield
that scoundrels can hide behind. We ought to simply
acknowledge that this stuff is out there and there
should be an expectation that companies running
internet access providing services or internet ISPs
ought to have an obligation to try to intervene to
minimise or reduce this type of material, even if you
do not hold them legally liable.
Q33 Steve Rotheram: Do we know how many sites
or images are taken down on a regular basis?
John Carr: You have the IWF appearing in a minute.
Chair: Absolutely; that is for our next panel.
Claire Lilley: Just one last point. There is a lot of
information we do not know and one of the things that
we do not have in this area is some sort of high-level
strategic information-sharing forum. There is a lot of
partnership working going on, but there is no one
place in which it is held in the same way as I believe
there is for terrorism, for example. We obviously have
UKCCIS, but it does not deal with child abuse images
online because it largely takes the view that those are
illegal and therefore that is dealt with. I think we do
need some sort of more strategic forum at which
industry—CEOP, IWF, everyone—gets around the
table and shares information and best practice about
what can be done.
Chair: That very neatly leads us to our next panel, so
may I thank the three of you very much?
Examination of Witnesses
Witnesses: Susie Hargreaves, Chief Executive, Internet Watch Foundation, and Peter Davies, Child
Exploitation and Online Protection (CEOP) Centre, gave evidence
Q34 Chair: May I welcome our next two witnesses:
Susie Hargreaves, Chief Executive of the Internet
Watch Foundation, and Peter Davies, the Chief
Executive of CEOP?
Perhaps we can continue from the point we reached
with the last panel: can you first of all give us your
idea of the volume, the scale of the problem and the
number of people who are involved; and perhaps
address Claire’s point about the need for greater coordination in gathering information?
Peter Davies: After you.
Susie Hargreaves: Thank you very much, Mr
Chairman. Would it be useful if I just gave a brief
summary of how the IWF works and the key areas in
which we operate?
Chair: Yes.
Susie Hargreaves: The IWF is the UK hotline for
reporting criminal content, specifically child sexual
abuse content hosted anywhere in the world, nonphotographic images of child sexual abuse hosted in
the UK and obscene adult content hosted in the UK.
We are a self-regulatory body founded by the UK
internet industry 17 years ago. We are an independent
charity with an independent board and independent of
law enforcement and Government. Our operations are
covered by a memorandum of understanding with the
Association of Chief Police Officers and a service
level agreement with CEOP. We work very closely
with CEOP on a day-to-day basis. Very quickly, we
are funded by the internet industry—80% of our
money comes from them and the remaining 20%
comes from the EU as the Safer Internet Centre for
the UK along with Childnet International and the
South West Grid for Learning.
Specifically, the way we work is that we currently take
reports from the public—although this is about to
change—and assess those reports, and if we find that
child sexual abuse content is hosted in the UK, we
issue a notice and takedown. We typically have the
content removed in under 60 minutes. If it is hosted
outside the UK, which the majority are, we notify the
sister hotline in that country or international law
enforcement, and until such time as that content is
removed, we place it on a blocking list, which is used
across the UK and in many other countries. It is a
very dynamic list and it is updated twice a day, and
the content stays on the blocking list until such time
as it is removed at source. The most effective way to
remove content is to remove it at source. It is our view
that blocking will only stop inadvertent access and
will not stop the determined. We also have a
keywords list.
My final point is to say that 17 years ago, when we
were founded, 18% of child sexual abuse content was
hosted in the UK. It is now down to less than 1%
and that is because of the tremendous relationships we
have. It is about partnership working. It is about the
UK and international internet industry, and our
relationship with law enforcement, the Government,
the public and civil society.
To answer specifically your question about the extent
of the problem, the truth is that nobody knows how
many images are out there. John has mentioned a
number of images, but we do not know. Nobody
knows how many unique images are out there. The
majority of the images we see are duplicates and I
imagine that the majority of the images within those
numbers that John quoted were also duplicates. To
give you an indication, we see about one to two new
images a week. The scale of the problem to us is that
last year we had about 40,000 reports of which we
actioned just under 9,500 individual webpages. A
webpage can contain from one image up to thousands
of images or videos. Of those just under 10,000, 35
were hosted in the UK. In terms of the type of content,
81% were of children who appeared to be 10 and
under; 4% were under 2; and 53% were levels 4 and
5, which are the rape and sexual torture of children.
Q35 Chair: Just before Peter Davies comes in, I have
a couple of questions. You said one or two new
images a week. That does not sound very many, given
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 12 Culture, Media and Sport Committee: Evidence
15 October 2013 Susie Hargreaves and Peter Davies
that previously we have been talking about millions
and millions of images. These are just the same
images being circulated round and round? You are not
suggesting that only one or two new images are taken
every week; it is just the ones that you happen to
stumble across, is it?
Susie Hargreaves: Yes. The majority of images that
our analysts see are images they have seen many,
many times because images are recirculated. This is
one of the issues that we try to get across, particularly
in terms of young people’s behaviour on Sex Teen, is
once an image is out there, it is out there for ever
because it will just be recirculated and redistributed.
What it means is that when we do see a new image,
we can immediately work with the police to notify
them, because if it is an unidentified child, we can say
it is a new image and pass it on to the police.
Just also to clarify, our role is around the removal of
content, and we pass on and share intelligence with
the police, but we are very much not part of the police
side of the work. We are not about the child protection
side or the safeguarding of that child, or in fact the
investigation of the perpetrators. But in terms of the
new images, we see very few new images, but we see
an awful lot of recycled images.
Q36 Chair: You said the best way of dealing with
this is at source and that you talk to your sister
organisations. But in the world where there is no
equivalent of you, where people are likely to host this
material, are the particular countries where it is hosted
that you cannot get to?
Susie Hargreaves: There are 42 hotlines in 39
countries. The majority of them are in Europe and
North America and Australia—those sort of countries.
We are working with the Commonwealth and with
ITU, the UN International Telecommunications
Union, to look at ways in which we can provide a
back-office solution for countries without a hotline
where they are having increasing internet penetration
to be able to report. We provide them with some
support to do that, because what we are seeing, as
John Carr talked about, is that as internet penetration
grows in developing countries, people are starting to
access these images.
In terms of where the content is hosted, about 50% of
it is in North America—the US and Canada. We are
seeing less content hosted in Russia than we saw in
previous years. Patterns change slightly—we are
seeing some emerging markets out in Asia, Cambodia,
Vietnam—but the majority is hosted in the United
States.
Q37 Chair: Yet the United States presumably has
pretty tough laws about this, does it not?
Susie Hargreaves: One of the reasons we are so
effective in the UK is because we self-regulate, so we
do not have any mandatory powers to enforce internet
companies to remove the content, but they do it on a
voluntary basis, which means they can remove the
content very quickly. In the States, internet companies
are obliged to notify the National Centre for Missing
and Exploited Children—Peter will be able to answer
this much better than I can—and they then have to go
through a quite lengthy judicial process to have the
content removed. That means that it takes a lot longer.
Since we have been in operation, it used to take 20
days to remove the content in the US. It is now down
to 10 days. We are working very closely with them.
We now have permission to go directly to
organisations that are members of ours and are based
in the States and simultaneously to alert them to the
fact that they are hosting content at the same time as
we tell the police in the States. We are trying to bring
that time down, but it is to do with the judicial system
and how law enforcement works in the States, but I
defer to Peter’s expertise on this.
Q38 Mr Bradshaw: You just implied that the speed
of taking content down is part of the reason why a lot
of it is hosted in the US. Did you mean to suggest
that connection?
Susie Hargreaves: No, I do not mean to do that. The
reason why a lot of it is hosted in the US is, first of
all, that a lot of companies are based there. The
internet industry is very big in the States. Also, there
is all sorts of legislation around freedom of
information and ability to host certain sites, which is
to do with their judicial systems. But, Peter, you
would probably answer that better.
Peter Davies: Chairman, thank you. Should I perhaps
wind back to an introduction, if that is okay?
Chair: Yes, I think that would be helpful.
Peter Davies: But I will happily address any questions
raised. I am grateful to Susie for leading off.
My name is Peter Davies. I am now the Director of
CEOP, which is one of the four commands within the
National Crime Agency, which commenced last
Monday, 7 October. The CEOP centre has existed
since 2006 and was previously affiliated to the Serious
Organised Crime Agency. Our mission, whether as a
centre or as a command, as we are now, is to protect
children from sexual exploitation and sexual abuse.
When we were founded, there was heavy emphasis on
online protection. Our remit does not limit us to online
activity, but what we try to do is to become a national
hub for intelligence, a national contact point and a
source of expertise and specialist support on any
aspect of child sexual exploitation that would benefit
from that approach. We still have a very strong
heritage and a very high level of expertise in child
abuse imagery and other related forms of harm.
Our threat assessment, which is published every year,
was most recently updated in June this year and in
that we identified four key threats to which we address
our minds, two of which are specific to online. One
of those is the proliferation of indecent images of
children. Images, as I am sure you will know, are a
harm in themselves. They are scenes of crime. They
are depictions and recordings of children being very
seriously raped and sexually abused and the
possession of these is a crime—and rightly so. The
possession of these is also, as Claire Lilley referred
to, quite a strong indicator of a sexual interest in
children that also extends to physical abuse, and so
anybody who possesses indecent images of children
is a criminal, but is also somebody who might present
an additional risk to children, as if that were not
enough.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 13
15 October 2013 Susie Hargreaves and Peter Davies
I can go back to talking about the overall level of
prevalence, but I also want to highlight the other
aspect of threat that we are very concerned about,
which is what we call online child sexual exploitation.
This is less well understood. I would not estimate the
prevalence of it, but what we can say is that this
appears to be an emerging trend of offending and its
consequences for the victims extend to fatal
consequences and are in many cases very severe. This
is not so much about taking an interest in indecent
images of children as about using the medium of the
internet to contact, ruthlessly manipulate and exploit
and sexually abuse children to the point of causing
them physical and psychological harm, and in some
cases pushing them towards suicide either because of
the acts they have been compelled to do, or in some
cases because they have carried out acts online that
are then used to blackmail them for money—money
they do not always have. So these are the two key
areas of threat.
We have excellent relationships with a number of
external bodies, including the Internet Watch
Foundation, which I think is a world-leading
organisation at doing what it does.
Our remit does not extend to all the issues that I know
you are examining because we have to focus in.
Everybody has a view about the risks associated, for
example, with children accessing adult pornography
online. We have to restrict our interest in that
professionally to the extent to which it makes them
more vulnerable as victims, or maybe increases the
possibility that they will be offenders in future.
Turning to prevalence, on the second threat I referred
to, online child sexual exploitation, we do not know
and we would not pretend to know. We deal with a
number of cases, and a growing number of cases, of
increasing severity. The type of offending is hugely
harmful and also complex to investigate because it
necessitates no physical contact at all between the
offender and the victim. It is not uncommon for us to
encounter situations where offenders in one country
will target and harvest victims in a completely
different part of the world. The complexities that
presents are obvious. But we do not know how much
is going on and it is in the nature of any form of
child sexual abuse or exploitation that it will be underreported by the victims because the whole process is
designed to minimise the risk that they will report and
designed to induce shame and embarrassment and all
the things that might stop a young or vulnerable
person coming forward and saying what has happened
to them.
Q39 Chair: Where is it principally happening? Is it
Facebook?
Peter Davies: It is important to identify that what we
are trying to tackle is appalling aspects of human
behaviour, to some extent facilitated by the internet.
At the moment we are seeing roughly half of this kind
of activity—online child sexual exploitation—taking
place through social networking sites. Of course
Facebook is a major one, but not the only one. We are
also seeing it taking place through chat environments
and other forums. But it is important to realise, as I
think John Carr said previously, that the medium is
not to blame. The medium might be managed better
so that it is safer for its users, but what is to blame is
human behaviour, albeit through the internet
amplified, multiplied and in some cases almost
industrialised to a quite remarkable degree.
In terms of the proliferation of indecent images of
children, we are a little bit clearer about this. I would
not estimate a total number of indecent images
circulating on the internet; I would not estimate the
number that are duplicates or unique. I would
probably observe that regrettably there are more now
than when you started sitting this morning and that is
one of the major challenges that we have. You don’t
need me to explain what has to happen for an image
to be created.
Q40 Chair: That does not quite equate with the one
or two new ones a week, if you see what I mean.
Peter Davies: Our experience is different, because of
course we are a different organisation and we receive
reports through different routes. It is absolutely true
to say that the majority of images that we encounter
are repeats. I do not have a figure on the proportion
that are new, but a lot of our investigations are not so
much into the images themselves but into the child
abuse they portray or the child abuse that they give us
an opportunity to interdict, and so it is understandable
that we may not come up with the same figure. As I
say, I would not estimate the total number circulating.
There are always more being produced and obviously
that again is a harm in itself.
Where we have got somewhere as a centre over the
last two years is getting a sense of how many
offenders there are in the UK, and I think John Carr
has already referred to this. In our thematic
assessment published in June, we estimated that there
may be around—I use that term carefully—50,000
people in the UK who commit offences at least to the
level of possessing indecent images of children, which
is quite a stark figure to contemplate. But I am firmly
of the view that it is not enough to deal with a handful
of cases—and deal with them well, which I believe
CEOP and our colleagues in law enforcement,
supported by our partners, do—but have a sense of
the overall scope of the problem, because it is that
sense of the overall scope that drives a lot of other
people who can help to want to do something as well.
I would echo roughly what Susie says about the trends
in images. We check trends in images every year and
our experience recently has been that the victims
appear to be getting younger and younger and the
level of abuse portrayed appears to be getting worse
and worse. So there is no comfort in any of these
aspects of our analysis.
In terms of reporting, if I may just cue in based on
Susie’s point, we do receive reports from the Internet
Watch Foundation. We have a very clear service level
agreement and we are looking to revise and update
that into a collaboration agreement very shortly. We
are keen to work together with the IWF every step of
the way. We also have partnerships with other
organisations that report incidents to us, the most
high-volume one being the one that Susie referred to,
which is the National Centre for Missing and
Exploited Children in the United States, from whom
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 14 Culture, Media and Sport Committee: Evidence
15 October 2013 Susie Hargreaves and Peter Davies
we receive on average maybe 800 reports per month.
We also have reports from UK organisations
including, very importantly, the IWF. We are also
available for any member of the public who wishes to
report any concerns to us online. That public reporting
has probably doubled in volume over the last year or
year and a half, and we generally expect 300-plus
reports into CEOP online from the public every month
about incidents of concern.
Susie Hargreaves: Mr Chairman, could I clarify the
issue around the number of new images? When we
talk about new images, it might be a series of images,
but you have to remember that that means two new
children who have been sexually abused. So it is not
just an image, it is two new children who have been
sexually abused in itself, and it will be the worst of
the worst stuff.
Also, I just wanted to say there is a lot of information
out there about potentially how many images there are
and of course it is a changing field from day to day,
but I do think there is still a need for some very robust
research. I would certainly echo what Claire was
saying earlier on. Although we have the UKCCIS
board, of which we are both members, that is very
much around the whole field of child internet safety
and online protection. In the field of online child
sexual abuse content, we do need a kind of high-level
strategic thinking body to get together a couple of
times a year to start to share that high level of
intelligence. It only works in the UK because we work
in partnership and all do our specific bits, but it is
really important that we understand the whole and
the context.
Q41 Mr Bradshaw: If most of this content is being
generated in the States, who are these children? Have
they been—
Susie Hargreaves: Sorry, could I clarify? It is not
necessarily generated in the States; it is hosted in the
States.
Peter Davies: If I can help, the process by which it is
identified is a specific piece of legislation that exists
in the United States called the Protect Our Children
Act 2008. It places an obligation on communications
and internet service providers to identify and report to
NCMEC, the National Center, incidents that appear to
depict child sexual exploitation. There is a piece of
legislation in the States that places this obligation and
it is in the discharge of that obligation that they report
everything—well, the performance varies, but they
report things to the National Center and the National
Center identifies the location of these through internet
protocol addresses and so on. What we receive in the
UK is effectively, for want of a better expression, the
UK slice of all that which is reported to the National
Center.
Q42 Mr Bradshaw: So where is it being generated?
Who are these children? Have they been kidnapped?
Are they within families?
Peter Davies: It is hard to tell. There is one more
thing I could add. There are very few internet
transactions that we can do that do not travel through
more than one country, and because a great deal of
the internet is hosted or serviced in some way in the
United States, that is why this piece of legislation is
so powerful. It may well be that it is a UK individual
sending to a friend an indecent image of a child, but
because of the way that the internet works, it passes
through the US. Depending on the extent to which the
service providers are attentive and surveil these
things, that is where it might get picked up.
Q43 Chair: IWF, you have been talking about
content hosted on websites that people can access.
You have just mentioned direct peer to peer. Is it
possible to give us a rough estimate of what
proportion is done peer to peer and what proportion is
done through a website?
Peter Davies: Yes, it is. We do not share all the
information we have or the means by which we
acquire it, for reasons I am sure you will understand,
but our estimation is that the highest volume of
offenders is in the peer-to-peer category, which means
they are not offending on the open internet. They are
in groups of people whom they know, or think they
know, and they have an increased level of trust and
confidence that within those groups they can share
child abuse imagery, child abuse material and so on.
So the highest volume of offenders is in the peer-topeer area.
There are offenders who use what is called the hidden
internet, also known as Tor or The Onion Router or
something similar. This is harder to penetrate, but I
believe not impossible. These offenders are far fewer
in number. The issue with them is they quite possibly
pose the most serious risk because they are technically
sophisticated and some of the harm that we see taking
place in that part of the internet is some of the worst
that we see.
Q44 Chair: And Tor, do I need a computer science
degree to use it? I know you cannot find—
Peter Davies: Absolutely. I do not want to advertise
it, but it is possible to download it and start using it
on a normal home computer. It was designed by the
US Navy in the 1980s, I believe, and released to the
world. Some of our most sophisticated offenders use
Tor. It is also important to realise that some other
serious organised criminals use Tor as well. People
may be aware of the arrest of somebody who was the
proprietor of something called The Silk Road, so Tor
is also used for drugs distribution, distribution of other
criminal materials and obviously communication
between criminal groups.
Q45 Steve Rotheram: IWF and CEOP have
demonstrably been successful, if we have gone from
a fairly high proportion of the images being generated
in the UK to 1%. That in itself is a success rate.
However, behind each of these individual images
there is a victim. How successful are the police in
pursuing the perpetrators of the crime against the child
who has ended up on the internet?
Peter Davies: An excellent question, and I am not
going to be able to provide a simple answer. As a
centre and as a command, CEOP is getting more
successful year by year in converting the intelligence
that is derivable from the internet into outcomes in
terms of safeguarding children. Last year we put out
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 15
15 October 2013 Susie Hargreaves and Peter Davies
more intelligence packages than ever to law
enforcement and also to children’s services in the UK,
precisely with that intent. It is also important to
highlight that within police forces there are very often
technically proficient and committed online child
protection units, so it is not as if CEOP is the only UK
law enforcement asset in this area. The larger forces,
particularly, often have real experts doing great work.
I think we are in a transitional phase at the moment
where we need to understand better what the overall
capability and capacity is of the whole of UK law
enforcement in order to tackle this kind of offending,
and there is a plan in place to do that. We do not have
the sufficiently high-quality robust data for me to give
you proper arrest figures and things like that, and I
can only really tell you at the moment that which gets
reported back to CEOP. Again, there is a plan to
improve and develop that because I think it is
important for me, or the person sitting in my chair, to
have a really good answer to that question and to be
able to provide it for you. My observation is that there
is a lot more law enforcement activity going on than
it is possible to capture at the present time. This is
something that HM Inspectorate of Constabulary has
picked up on and noticed, and is working on.
John Carr has a way of putting things very simply and
of course if there are 50,000 offenders out there and
they are detectable, we should not really be resting
until they have been detected. What we have to do in
the face of a gap between provision and what we
would like to do is prioritise, and that is what law
enforcement and others charged with public safety
always do. What we try to do is prioritise the people
who do pose the highest risk to children by bringing
in risk factors that help us identify them. We focus on
the worst offenders—the people who are at the heart
of networks that cause greater harm—so numbers are
not the whole story; it is the seriousness and the
numbers put together. The best evidence of whether
we are arresting the right people of those that we can
arrest is the number of children who are safeguarded
as a consequence. I can tell you that in the annual
report that we published in May this year, it was very
clear that we are getting more successful than ever at
doing the most important thing I think we can do,
which is safeguarding children who are at serious risk.
Q46 Steve Rotheram: Is there any way that we can
quantify that? While I absolutely accept the success
of taking down images and the fact that it is
demonstrably and obviously a criminal act to
distribute the images, surely identifying the poor
child—the individual who has been exploited—should
be the primary concern.
Peter Davies: Absolutely, it is.
Q47 Steve Rotheram: Do you work with
international law enforcement agencies? If we can
identify an image that might not have been taken in
the UK but might have been taken outside, do we still
pursue that outside of our boundaries?
Peter Davies: Very much so, apologies. We have a
victim identification unit in CEOP that has huge
expertise in looking at images and looking for clues
within them about how to track back to the child. This
was a specialist unit that was featured relatively
recently on Crimewatch, for example.
We always seek to cultivate the best possible
relationships with law enforcement partners in other
countries. CEOP was a founding member, and is still
a board and management member, of an organisation
called the Virtual Global Taskforce, which now
comprises eight different countries from around the
world. Well below that strategic level, there are really
good relationships, and indeed technically facilitated
virtual private networks that enable, for example, my
victim identification team to deal directly with
colleagues around the world.
Some of our best successes have been quite
remarkable regarding the speed with which we have
received intelligence on UK activity from another
country. We have seen opportunities to identify where
that is taking place from the image and we have dealt
effectively with the local police force concerned.
There are occasions within my time at CEOP where
that whole process from end to end has taken less than
a day, and as a result children who were definitely
at serious risk have been safeguarded and offenders
arrested. That is the ideal. That is what we would like
to happen every day. It is in the nature of the fluidity
of these things and the complexity of international
relationships that that does not always work, but it
works sufficiently often to make us believe that it is
worth pressing on with that. There is no occasion
when we would not share information with law
enforcement elsewhere that might give them the
opportunity to do the right thing in their country as
well.
Susie Hargreaves: We work also in the field of victim
identification. We work very closely with CEOP; we
are trained by CEOP. When our analysts see a new
child, they look for particular clues within the picture
and do some analysis of those new children. As a
result, we also work very closely with international
law enforcement—with Interpol; with Europol; with
international locally-based law enforcement—and
certainly through our work over the last three years,
12 children have been rescued. There are statistics that
we can show in relation to rescuing children. It is
about a partnership approach, but very much working
very closely with law enforcement on an
international level.
Q48 Steve Rotheram: Are there any changes that
either organisation would like to see to either the law
in this country or international law that would help
you to do your jobs better?
Peter Davies: Yes. In terms of international law, we
need to start from the point of recognition that the
UK has just about the best suite of child protection
legislation that there is, and that we are more of an
example to others than we are in need of catching up.
But when you are in the business of child protection,
that gives you no cause for complacency, but it is
important to note at a Committee such as this. There
are two conventions, details of which I would rather
supply separately, but they are the Budapest
Convention and the Lanzarote Convention. One is
mainly about online matters and one is about child
protection. My sense is that if countries sign up to
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 16 Culture, Media and Sport Committee: Evidence
15 October 2013 Susie Hargreaves and Peter Davies
and work up to both these conventions combined, that
provides probably the highest straightforward level of
assurance you can have that they will be equipping
themselves to protect children online.
Turning to the UK, it is quite remarkable to me how
far the criminal legislation, for example, around
indecent images of children, which was, I believe,
conceived and passed before the internet was around,
has still pretty much stood up to the new world of
child abuse online, and I do not think what we need
is a basic root-and-branch piece of work. It is my firm
operational view, and I have articulated it previously,
that if there is one piece of legislation that would most
help us tackle online child abuse, it would be the
provision of a clear, consistent and properly enforced
regime for retaining and accessing communications
data, because we are regularly in a situation where we
are unable to convert, for example, an IP address into
a name and address for lack of that precise thing. I
appreciate that it has been under consideration. I am
merely repeating an operational view that I have
articulated previously.
Q49 Chair: You can get a court order to get an ISP
to divulge the address.
Peter Davies: It is possible. The issue is that it is
not always an ISP; it can be a communication service
provider. They are sometimes configured differently.
My experience is that some are more able to provide
the information we need than others, and when this
legislation was under debate a little while ago, I took
the view that our general attrition rate—the proportion
of questions we would like an answer to that we do
not get—is roughly 25%. Now, if you are dealing with
a repeat victim or repeat offender, there are ways of
getting around some of those, but it is implicit in that
missing 25% that there will be lost opportunities for
us to do the right thing. At the expense of sounding
repetitive, I still think that is the most important single
piece of legislation that would be of assistance to us.
Susie Hargreaves: Three things, if I may. One is that
we do have some of the most robust legislation in the
world in the field of child sexual abuse online, and
certainly this is one of the reasons why we are setting
up a back-office solution for countries without a
hotline, because we will assess it under UK law and
it means that it will be dealt with very thoroughly. So
on that front, we do not feel that we need new
legislation.
What we do feel though is that it is very important
that people do not conflate the two issues between
people accessing criminal content, which is what we
deal with, and young people’s access to adult content,
and this is not particularly helpful. Certainly within
Westminster, the issues need to be kept totally
separate. Where, as a citizen and a mother, I might
care that my teenager has access to adult content, it is
entirely different to anybody having access to criminal
content. We need to keep these very separate. We also
need to have an awareness that we are a world model
in relation to the internet industry support for the IWF,
and in terms of the internet industry’s commitment to
support the removal of child sexual abuse content.
They are currently working with us on a membership
review to look at ways that they can step up that
support, and it is being done in a very positive frame
of mind, and certainly it is one in which the industry
does take its role very seriously. We only exist
because of the industry.
The third thing I just wanted to say was that in relation
to tackling this, it is our experience that young men
aged between 18 and 25 are the most likely to stumble
across it in the first place, and they will do that
through the open internet, not through peer-to-peer
networks. Because it is such a huge issue, we need to
do much more to raise people’s awareness to report
and do the right thing, and to know that it is wrong.
It is not just a question of looking at an image; it is a
child who has been sexually abused. It is criminal.
The consequences of looking at it are potentially
extremely serious for everyone concerned, so it is to
raise awareness to stop people going there in the first
place.
Q50 Mr Sutcliffe: As a Home Office Minister in
2006, I visited CEOP under the leadership of Jim
Gamble and, as you quite rightly said, Peter, it was
set up as a unit because there seemed to be no
consistency of approach across the 42 forces in
relation to this issue. Seven years later, with all the
experience that CEOP has gained, is the incorporation
of CEOP into the National Crime Agency a good
thing or a bad thing?
Peter Davies: I think it is absolutely a good thing.
You will no doubt have followed our progress in the
meantime and be aware that when I came in three
years ago, it was very much a major issue. Since then,
we have set about the business of contributing to the
construction of the National Crime Agency, and also
to making sure that we were absolutely able to dock
into it effectively. There are a number of different
aspects that I can offer up as evidence that it is a good
thing, but the simple answer is, “Yes, absolutely no
question.” However, may I just offer one or two
examples none the less?
Firstly, SOCA, and therefore CEOP, when you visited
us, were never subject to a piece of legislation called
the Children Act 2004, which places an obligation on
public bodies—and others where appropriate—to
have regard to the safeguarding and welfare of
children in all their activities. Every police force was
subject to that expectation, but not SOCA and
therefore not CEOP. The new National Crime Agency
has asked to be and is subject to that obligation, and
I can tell you, as the Director of the CEOP command,
that is influencing not just how we do the work that
CEOP has traditionally done, but also influencing the
way that the Agency thinks about the interests of
children in its other activities, and that is quite a
significant difference.
During the time I have been Chief Executive and now
Director, the resources and the number of people
available to do CEOP’s work have increased, and that
increase will continue into the NCA. There are aspects
of what we have been doing in the past that are a
major test of resilience, and the NCA is taking on
some of those aspects in order precisely to help to
take some of the load that that creates. The alternative
is, as we have had to do in the past, that we have to
move people from other parts of CEOP to support
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 17
15 October 2013 Susie Hargreaves and Peter Davies
some of these functions, which is not the best use of
their time.
Another major gain is that it is not just about CEOP
any more. Our mission is shared by the whole
National Crime Agency and the structure of the
agency is such that where the intelligence is
sufficiently compelling and where the balance of risk
dictates, the NCA can mobilise 2,000 of its own law
enforcement officers to carry out operations to do with
child sexual exploitation. One of the first tasks, other
than set up the agency, that its Director General, Keith
Bristow, was asked to do was to conduct a further
investigation into child abuse at the Bryn Estyn
children’s home in north Wales, which was an
operation carried out under the operational name
Pallial, and there have been a number of releases of
information and updates about that.
I could go on, but all those things point to the NCA
being the right place for CEOP to be, and also to the
major gain that it is not just about CEOP against the
world—we are part of an agency that shares our
objectives and ethos, and that is keen to help.
Q51 Mr Sutcliffe: I am heartened by that, because
clearly it was a high-profile agency when it was set
up, but what you are saying now is, with maturity, it
is now back within the family, if you like. That then
leads me on to the relationships with police forces
in terms of whether CEOP just seen as, “You refer
everything to CEOP,” or do forces actively deal with
this as an issue in their parameters?
Peter Davies: I think forces have actively dealt with
this for a long time. It has not always been fully
acknowledged. I myself, before I came to CEOP, was
involved in a major crime inquiry to do with a news
group permitting the sharing of indecent images
online, which we did in partnership with CEOP, but
that was a local police force doing the right thing, and
there is a fair bit of that going on.
Another aspect of the agency, which we have to
approach with caution, is that ultimately it has a
responsibility not just to manage how it uses its own
4,500 officers, but for making sure that the collective
efforts of UK law enforcement deal with the risk in
the most efficient and effective way possible. There is
no big problem there, because we experience very
good relationships with police forces. We co-operate
very well, we try to find the right balance of resources
so they can do what they do best and we can support
them, or sometimes co-ordinate them—or very
occasionally lead them, as appropriate. So the picture
was never that bleak.
My experience, more anecdotal than out of
certainty—refer to my previous answer—is that police
forces are investing in the skills to do this area of
work. Where I am less clear and less well able to
answer with the accuracy I would like is on exactly
how much capacity is invested in this across those 42
forces, the extent to which it looks similar and the
extent that it can be interoperable, which means that
officers from one force use the same systems and have
the same approach as officers from another and can
easily work together when possible. We have a plan
for how we are going to learn that. The architecture
of the NCA allows us to pursue that sense of national
consistency in a way we were never able to do before,
and that is another significant gain.
Q52 Mr Sutcliffe: May I just follow that through
because when I have briefings from our local police
forces about what goes on within their force, it is
about robberies or a variety of different things that are
there, but this never gets reported, in the sense that
they do not tell me about what work they are doing in
this area. What I am slightly worried about is how we
get—because a lot of this is about education and
making sure that people understand the nature of this
crime—the information out there? What are the
mechanisms for raising the awareness of those forces
to raise awareness within their communities?
Peter Davies: Yes. I think there is a whole variety of
things. You may have seen when you visited CEOP
that we have an education team and we have a very
strong internet presence. We have a website to which
the public are more than welcome to get advice and
to access materials. We also have a network of
ambassadors and volunteers, and we create new
education packages every year with the intention that
our ambassadors and volunteers can convey them to
children. This is a real success story for CEOP,
because out of a team of maybe three or four people,
we manage to reach over 2.5 million children a year
directly in a structured classroom or similar setting.
Our website regularly attracts more than a quarter of
a million unique visitors per month and recently, when
we did an information campaign about online child
sexual exploitation, the number of people coming to
our website was higher than it has ever been before.
Of course, there is always a gap, and there are far
more children out there who receive the message. This
is one of the success stories, and we engage with any
police force that wants to, in the same way as we
engage with teachers or others to use the materials
and take them out.
In terms of what police people talk about as their
priorities, which is another aspect of what you were
saying, I genuinely believe that child protection—
online and offline—has never been closer to the
mainstream of law enforcement in the UK than it is
now. The police service has always been populated by
experts—hugely passionate, committed people who
make immense sacrifices to do the work that they do.
But it is not just that. It is inescapable for police forces
to have noticed the increased number of cases of what
is called street grooming or localised grooming,
another form of child sexual exploitation. There is a
national policing plan in place to make sure every
force has the proper capability and capacity to deal
with that, and I am involved in overseeing the delivery
of that, but that plan comes from Chief Constables
and is being implemented by local Chief Constables,
supported by their PCCs. This is an interesting area,
because it is quite technically difficult and it requires
a lot of innovation and flexibility and quite a lot of
expertise to get. Generally speaking, law enforcement
does not lack that, but keeping up with internetenabled crime is always going to be a challenge.
I look forward to the day—I do not think it is far off—
when police people talk just as much about the child
sexual exploitation operations they are doing and what
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Ev 18 Culture, Media and Sport Committee: Evidence
15 October 2013 Susie Hargreaves and Peter Davies
a great result they had safeguarding children through
an online inquiry as they would be talking about
armed robbery and so on. Probably, in some police
forces, we are already there.
Q53 Tracey Crouch: A lot of my questions have
already been asked, but could I just ask: do you think
that the social media providers could do more to help
combat illegal behaviour?
Susie Hargreaves: I would say that all the internet
industry regardless of social media—search engines,
ISPs, mobile operators—are all extremely aware of
their responsibilities in the field. We work very closely
with all of them. Following the Secretary of State’s
summit earlier in the summer, we were asked to look
at proactively seeking content in the future, which
means we will be the first hotline in the world with
the powers to do that, and that will mean we will need
many more analysts. In order to support that, we have
been working with the internet industry to look at
ways that they will adjust their fees accordingly. We
are in the middle of some very, very positive
discussions and everybody has stepped up. In terms
of their responsibility generally, I think all of them are
really aware of what their responsibilities are and they
are all doing what they can.
Certainly, I speak regularly at child online protection
conferences and safety conferences, and you will have
Facebook and Twitter there, and you will have those
organisations talking about what they are doing to
address the problem. I think they are aware—there are
always things that all of us could do—of what they
need to do and they are stepping up. But it is very
much a group approach and a partnership approach,
and in the same way as the police cannot solve it on
their own and we cannot solve it on our own, we all
have the responsibility to solve it. Facebook can do
so much, but people also have to adjust their
behaviour and so the education and awareness side
is absolutely crucial—young people knowing how to
protect themselves online. All these messages are
really important, and that is part of our partnership
in terms of the Safer Internet Centre with Childnet
International, which goes out to schools and do all the
awareness raising, and South West Grid for Learning,
which gives support to professionals in terms of
making sure that people can keep themselves safe. It
is part of the overall picture, so I do not think it is
enough just to say that one organisation needs to do
more. Collectively, we all need to do more.
Q54 Tracey Crouch: Do you have support networks
in place for your staff because this is pretty horrific
stuff?
Susie Hargreaves: Yes, we do. We are currently
recruiting for seven analysts, and we go through a
very intensive recruitment process. We have a threestage interview that culminates in them having a
session where they see some of the images. We have
quite a good dropout rate at that level. Anybody who
sees any of our images has very intensive training.
They cannot work on their own for six months—well,
no one can ever work on their own, but they cannot
take assessments without a second person looking at
their work for six months. They have mandatory
monthly counselling, mandatory group counselling on
a quarterly basis and an annual psychological
assessment. So we do an awful lot on the welfare
front, because obviously it is a real worry to us
because they are looking at this content all day.
Of course, the public report to us stuff that is outside
our remit and they still have to look at it. Sometimes
they might look at something—they have to open a
video to have a look at it—and it might be something
very distressing, but is kind of outside of our remit,
so that is the stuff that takes them by surprise. Yes,
they are constantly seeing really very, very distressing
stuff, but they are the most incredibly sane group of
people and they do it because they are very committed
to the issue. Certainly, when we rescue a child or we
are involved in the rescue of a child, it is the greatest
feeling in the world, and they do it because of their
commitment. But we do protect them, yes.
Q55 Tracey Crouch: I assume it is the same at
CEOP as well?
Peter Davies: Yes. I feel like saying, “What she said,”
but that is not quite adequate. We have a very similar
approach. People are psychologically assessed before
they start at CEOP, regardless of what job they are
doing. That is refreshed on a basis that is
proportionate to the level of exposure they have to
material. It is really important that people have an exit
strategy if it just becomes too much for them, and that
is one of the huge virtues of CEOP being part of a
larger organisation. People do not have to leave and
find a new job; they can simply transfer to a different
kind of duty, and that way they can move on and
develop professionally, and continue to contribute in
some other way to public safety.
We also have compared our provision with provision
in similar law enforcement units around the world
because—I am sure Susie would say the same—only
absolutely the best is good enough in terms of taking
care of our people. They do an exceptional job that
most people would understandably shy away from,
and we should be thankful that they are there to do it.
Susie Hargreaves: Absolutely.
Q56 Tracey Crouch: My very final question is
probably relevant to you, Mr Davies. I have recently
been doing quite a lot of work on modern-day slavery
and saw quite a shocking statistic for someone who
represents a south-eastern constituency: 49% of all
instances of modern-day slavery are in the south-east.
Are you able to give a figure as to perhaps what
percentage or what proportion of modern-day slavery
would be related to some sort of child sexual
exploitation?
Peter Davies: I can give you a rough figure. Of
course, as with all these things, it is based on that part
of it that we can see, and while I may be getting it
wrong by a few per cent, we would be looking at
maybe 15% to 20% of human trafficking, which is
probably our previous definition of what we now call
modern-day slavery, to have a sexual exploitation
element. Of course, some exploitation or slavery has
more than one dimension to it and there is an overlap.
We have dealt in the past with child trafficking cases.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:11] Job: 035254 Unit: PG01
Source: /MILES/PKU/INPUT/035254/035254_o001_steve_CMSC 15.10.13.xml
Culture, Media and Sport Committee: Evidence Ev 19
15 October 2013 Susie Hargreaves and Peter Davies
There are other parts of the National Crime Agency,
including the UK Human Trafficking Centre, which
has a specific lead on that and is part of the Organised
Crime Command, which is one of the other commands
in the agency. More generally, our view is that we
want to support those efforts precisely because the
element of it that is child sexual exploitation is not
the majority of it. There are a number of other factors
to it: forced labour; street crime; organ harvesting;
relationship abuse; domestic servitude. All those need
to be taken into account, so the issue of modern-day
slavery is something where we would like to support
national efforts on it, but where we do not own it, and
that is where I think it should be.
Q57 Tracey Crouch: So you would welcome the
Government’s intention, albeit it is very early days, of
bringing in a piece of legislation that will help to give
you those resources?
Peter Davies: I am not familiar with the proposed
piece of legislation.
Tracey Crouch: Not yet, but it is an intention.
Peter Davies: From an operational point of view, I
welcome the intention to take a very serious look at
this issue and to do more, because within that thing
that we call modern-day slavery are some of the most
vulnerable people in this country, in desperate need of
rescue—some will be children, some will not—by a
law enforcement officer, not just a child protection
person. Those people need help, so anything that puts
the spotlight on that, and that challenges us to do
better and do our best at that, is welcome to me.
Q58 Chair: Can I just ask you a couple of quick
questions? Reference has been made to your
comments about how you had identified 50,000 to
60,000 individuals but you just did not have the
resources to arrest them all. Can you give us an
indication of how far short your resources fall from
being able to pursue all those leads?
Peter Davies: It is not just our resources; it is
everything. This is an extremely large body of
offenders and we have never been able to attach a
number in which we had sufficient faith to that body
of offenders before this year. I am told that there are
a total of 85,000 prison places in the UK, and that
gives you some indication of what would happen if
we got up one morning and arrested all these people.
Q59 Chair: Is it at 5% or 10% you are able to—
Peter Davies: To be honest with you, I am really not
able to tell you at the moment. One of the things we
are doing as a command within the NCA, and
obviously with the support of our colleagues in NCA
and so on, is to identify how many of those people we
can take on, but to have a plan for what we would do
with the rest as well. Because I meant what John
quoted me on, if I could, I would arrest all 50,000 of
them, and I think that is what people would like. But
it is not just about law enforcement capacity; it is
technical capacity, the courts, the prison system and
the child protection system. None of this is resourced
for a volume of offences of that nature, which means
that we need a plan that goes beyond enforcement.
The truth is that for most high-volume or serious
crime, you have to have a strategy that goes beyond
enforcing it. You cannot enforce your way out of it,
which is why our partnerships with the IWF and
others, our work with industry to make the internet a
tougher place to commit crime and our work with
children to make sure they are less vulnerable to this
kind of criminality all matters, because it is an issue
for society as a whole and the solutions lie in lots of
different places. Our contribution is to be absolutely
focused in pursuing the worst of the worst—pursuing
as many as possible within the obvious resource
constraints that the whole of UK law enforcement has.
Q60 Chair: Can I just follow up your point about
pursuing the worst of the worst? Susie Hargreaves
rightly warned us to distinguish between material that
is illegal and extreme and material that is not illegal
but should be viewed only by adults. John Carr was
suggesting that the police should be taking advantage
of the Perrin judgment to pursue websites that are
hosting adult content and taking insufficient measures
to protect their content from being viewed by minors.
That is a whole new area of operation. Would you
have sympathy with the idea that the police should
also be looking at that, or is your view that we should
just concentrate on the really bad stuff?
Peter Davies: I do not think I entirely caught that part
of his evidence. We have to prioritise. In law
enforcement—in public protection—we have to
prioritise every day. That is a simple fact of life.
Anybody who suggests that you do not have to and
that anybody has enough resources to do everything
you would like to do is not unfortunately in the same
world as the one I occupy. In terms of priorities, I
would view the existence of child abuse imagery, with
all the harms that it brings, and the existence of online
child sexual exploitation, as a higher priority in terms
of protecting the public than seeking to stop people
who are below a certain age accessing things that we
think are inappropriate for them. I make no apology
for considering that as a priority. I am not entirely sure
that John was suggesting that that is an area of work
the police should get seriously into, but I was not here
to hear his evidence.
I would certainly sympathise with the idea that given
that it may not be something that law enforcement
is necessarily able to prioritise, or that CEOP would
prioritise, that somebody somewhere, if they are
sufficiently concerned about that as a source of harm
and have the ability to do something about it, should
certainly do something about it. It may well be that
age verification or active verification by the
organisations that benefit from these people using
their services is a way forward.
Chair: I think that is all we have. Can I thank the two
of you very much for coming this morning?
Peter Davies: Thank you very much.
cobber Pack: U PL: COE1 [SE]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 20 Culture, Media and Sport Committee: Evidence
Tuesday 29 October 2013
Members present:
Mr John Whittingdale (Chair)
Mr Ben Bradshaw
Tracey Crouch
Philip Davies
Paul Farrelly
Mr John Leech
Steve Rotheram
Jim Sheridan
Mr Gerry Sutcliffe
________________
Examination of Witness
Witness: Stella Creasy MP gave evidence.
Q61 Chair: Let us get straight in. This is the second
session of the Committee’s inquiry into online safety.
We have three sessions this morning, and I would like
to welcome for our first session Stella Creasy, the
Member of Parliament for Walthamstow, Shadow
Minister for Business. Stella, once you have your
breath back, perhaps you would just like to begin by
telling us your experience.
Stella Creasy: Sure. Can I start by thanking the
Committee for inviting me to address you this
morning? I come very much in a private capacity, as
somebody who has experienced both the positive and
the negative elements of online engagement.
I presume the Committee is familiar with what
happened over the summer. A brief précis is this: I
got involved in a campaign to keep female
representation on bank notes—not a particularly big
issue, but one that nevertheless we felt was
important—led by a young lady called Caroline
Criado-Perez. When that campaign was successful, we
had a day where there was a lot of national press
attention about it. Twenty-four hours later, Caroline
contacted me to say that she was receiving threatening
and abusive messages on Twitter. I contacted her to
check that she was all right. She said that she felt she
was. Twenty-four hours after that, she said, “Actually,
I’m not okay”. I then began reviewing the messages
that she was getting and it was absolutely horrendous.
She was getting 50 rape threats in an hour. It was an
avalanche of abusive, very, very explicit, threatening
messages from a wide range of accounts.
At that point, which was the Friday of that week—we
had the campaign success on the Wednesday; by the
Friday this was happening—I felt a duty of care to
Caroline, having been involved in that campaign with
her. I contacted the police that I had been working
with in my shadow ministerial capacity when I was in
the Home Affairs team to say, “Look, this is
happening. We need to do something to protect this
young lady”. I was extremely worried because these
people were trying to post up her home address and
they were trying to find out personal information
about Caroline.
We then also tried to contact Twitter to say, “Look,
can you help?” I have to say, Twitter’s response, of
all the accounts that we could find for Twitter, was to
block Caroline and then to claim that they were the
victims of harassment. Twitter has since apologised
for that and I think they recognise that that was a
completely inappropriate response; that, when
somebody was reaching out for help because they
were under that kind of attack, it was right for them
to get support.
Over the course of that weekend, I continued to liaise
with the police to try to identify the level of risk that
Caroline was at, because the threats were relentless
and escalating by that point, and also to try to hold
the companies to account for their failure to act.
Unfortunately, on the Sunday evening I then became
a target for these people because I had spoken out
publicly, both in the media and online, to say that I
thought it was inappropriate. I then, along with a
number of other women in the public eye, was
subjected to the same sorts of threats over the course
of the next two weeks. These threats were very
graphic, very explicit.
I have always been a passionate defender of free
speech. I also think you have to recognise that free
speech comes when everybody feels that they can
participate and the nature of these threats was that
they were absolutely designed to intimidate, to harass,
to close down Caroline and myself. They were very
graphic, very specific. They were the sort of language
that would be illegal offline and the argument that I
made, and still make, is that there should be no
distinction about this kind of behaviour. Indeed, one
of the challenges we have here is to understand how
what we have learnt about these sorts of activities
offline can be translated into making sure that our
online spaces are safe.
It continued for about a week and a half. There have
now been a number of arrests. That case is ongoing,
so I do not think it is appropriate for me to talk about
that per se today, but I can tell you, from a personal
perspective, however tough you think that you are, to
get that kind of relentless onslaught—and I have had
just a fraction of what Caroline has had, and Caroline
is an amazingly brave young lady—is horrific. It is
absolutely harassment. This is absolutely designed to
intimidate, to scare, to frighten you, and it is
absolutely an issue that we need to see that both the
technology companies and the police understand is
part of the modern world.
I think it is inevitable when you do this job that, while
you might personally be affected by something, you
feel, “What can we do to fix this?” You start looking
at where the legislation can help and also where the
culture needs to change to understand that, just
because it is online, it is not somehow trivial. This is
part and parcel of modern day life and indeed it is not
Twitter or Facebook or Ask.fm that makes people say
these things. It is something in these people that
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 21
29 October 2013 Stella Creasy MP
makes them say these things and so we need to be
able to identify these people and engage with them
and hold them to account for their behaviour because
if we do not, the consequences could be quite severe.
In many ways, it was an odd series of events because
I had covered both the Data Communications Bill in
my previous job and also the protection from
harassment law and the stalking legislation. I felt both
sides of the concern, which was why it was important
what information the police could be asking of these
companies and what possibly the risk could be to
Caroline. The concern I had on that Friday night was,
“Well, is this one person sending 50 rape threats,
trying to find her home address, trying to target this
woman, or is this 50 people?” We have to find a way.
Just as offline, if you were in a pub, say, and it was
one person who was persistently aggressively talking
to somebody in that way, or 50 people, it would be a
different type of risk. We do not have the capacity and
the protocols in place as yet to identify those different
ranges of risk and, therefore, there is a very real
danger that we could miss something very serious as
a result.
All the way through this, both as a politician and as
somebody who was being affected by it, I had those
two things in mind. We do not know what the risk is
that we are facing and if people focus on the online
element of it, they miss that fundamental question:
what is the danger that we are trying to deal with
here?
Q62 Chair: Were you able to get an answer to the
question of how many people were involved in this?
Stella Creasy: No. As I say, there have been some
arrests. It is an issue I have raised with the companies.
The internet has come from the 4chan culture. We can
have a long academic discourse about the gendered
nature of that 4chan culture, but there is a very strong
thread of concern that there should not be intervention
in the internet, that what is wonderful about the
internet is that it is creative. I absolutely agree. What
that should not stop us doing is saying to the
companies, “But, look, the way in which people are
using these technologies is evolving and you
yourselves say that you want to make sure they are
safe spaces”. One of the arguments I have made to
the companies is that they themselves could identify
whether the same IP address was being used to send
threats without having necessarily to pass that
information on to the police.
If you recognise that there are different ranges of
threats, if you stop thinking this is about online and
offline, and start thinking that this is a range of
offences that might be created—one of my bugbears
about this is people kept trying to say this is about
malicious communications and, therefore, the content
of the messages being sent. I pushed back and said,
“No, you need to understand this is about protection
from harassment and the fact that these messages are
being sent at all and the impact they are designed to
have on the victim. Why does it matter whether it is
50 different people or one person?” If this person has
said, “Please stop sending me messages”, and they
continue to get these messages, that is harassment.
This concept of escalation—somebody who does not
leave somebody alone because they are fixating on
them, which we know very well from all the research
into stalking and into domestic violence and intimate
violence—we have not yet translated that
understanding to online behaviour. We know that 50%
of stalking cases involve online activity. To
understand what was happening to Caroline and then
to myself for those companies was not just a question
of, “Is somebody sending a message that is
malicious?” but, “Is their behaviour something that
could be escalating?” They could, in the way that they
have emergency protocols for dealing with reports of
somebody who has gone missing, have emergency
protocols where they do not necessarily have to share
that data with the police, but they could themselves
identify whether it is different people or the same
accounts or the same IP address.
One of the things that we kept seeing was people
taunting. They would get shut down. We know that
Twitter was then monitoring my account and
Caroline’s account, and Hadley Freeman and
Catherine Mayer, other women who then proceeded to
get these messages in that 10-day period. They were
immediately suspending accounts, only for an account
to open up two seconds later. I had a gentleman who
was “killcreasynow”—I apologise for using these
words at this time in the morning—“slutcreasynow”,
“eatcreasynow”. He set up a whole series of accounts
where he was targeting me with messages and
taunting the police that they could not catch him
because he could just set up new accounts. If you were
the companies, you could identify whether that is
different people all spinning off each other because
they are caught up in this moment, or one person
using the same IP address or even a very similar IP
address to be able to say, “This is a different type
of threat”.
Q63 Chair: Do you have any view as to whether
there was a co-ordination taking place, that somebody
was encouraging others to follow suit and send you
similar messages?
Stella Creasy: No. That is interesting because, in
terms of the actual messages that I would classify as
relevant for the police, and I was always very clear
that there were some I was reporting to the police, and
there were some people I was just sending pictures of
kittens to because they needed to calm down because
they were being hysterical about the idea that
somehow this was going to be the end of internet
freedom if somebody did something about it. Within
that there was a morass of people talking about the
issue, because the wonderful thing about Twitter and
all sorts of online channels is that we can
communicate. Certainly there were people sending
messages and badgering other people to get involved
in the debate.
There were one or two accounts that were goading
people to attack Caroline, certainly; whether those
were the same people—Then, two-thirds of the way
through, we came across a particularly nasty
phenomenon, which was the faux friend. I had people
sending me messages, going, “I think I know who is
doing this to you and if you follow me I can send you
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 22 Culture, Media and Sport Committee: Evidence
29 October 2013 Stella Creasy MP
details of them. I am really trying to help you. Look,
why are you ignoring me when I am trying to help
you, save you from these people? Oh, look, Stella
Creasy does not want to be helped. It is all about
publicity. Oh, look, Caroline is not taking help”.
There is a wonderful academic called Claire Hardaker
at Lancaster, who has just received some money—she
was sat behind me—and who is just about to do some
work on this about that element of behaviour. There
was certainly an element of unconscious
collaboration, I would argue, in that people identified
it. Some people certainly felt it was sport to get
involved in that debate and say, “You are obviously
not tough enough. Can’t stand the heat, get out of the
kitchen”, this kind of rubbish. There were other
people who were seeing it as an opportunity to express
what they thought was important about the internet,
this concept that it is without boundaries. I would
argue as someone who is a passionate defender of free
speech that is a misunderstanding about the concept
of free speech and how it operates.
There was no suggestion, as far as I am aware, that
the people who have been arrested knew each other,
but that is something, if this proceeds to trial, that
perhaps might come out. Certainly you could see in
some of the online communication the storm effect;
that people get drawn into it and people become aware
that you are responding to it. I think that is something
people have noticed with trolling—that there is an
element where people are trying to get people’s
attention, what my mum used to call a professional
wind-up merchant. I think that is slightly different
from this behaviour, which was, “We want, in an
aggressive, difficult fashion, to cause you pain, to
cause you harassment”.
Certainly some people who were involved were
enjoying the idea and, therefore, were drawn into it,
because here was an opportunity to get pleasure from
somebody else’s pain. Whether they had seen the
conversations and, therefore, felt—sorry, I am not
being very clear. There were the people who began it,
then there were the people who got involved in it, and
then the people who kept it going. I would say that
there are probably three different groups in that and,
within that, probably those latter two categories are
people that you are talking about; who wanted to be
co-ordinated by what was going on, but perhaps were
not necessarily in touch with each other. I am not
entirely sure. Maybe they were all direct messaging
each other. I do not know. There was certainly that
sense that they were being drawn into it by each other.
Q64 Tracey Crouch: Stella, you are high profile,
prolific tweeter and obviously this level of abuse has
gone too far. What sort of abuse have you been
subjected to prior to this latest event?
Stella Creasy: What, by my mother or just in general?
Tracey Crouch: On Twitter.
Stella Creasy: We are all extremely conscious, I am
sure, as all people around here who use Twitter and
who are in public forums know, that you get a degree
of banter, a degree of offensiveness, a degree of
people trying to get a rise out of you. I respond
robustly. I think I have had robust debates with a lot
of people around this table as well. This is very, very
different. When I looked at what was happening to
Caroline, it was very, very different. This was not
people saying, “I disagree with you”, or even using
colourful language to say they disagree with you or
even just being trite and telling you that they think
you are an idiot or using more colourful language to
say that they think you are an idiot. These were very
specific threats to do very, very graphic things that
were repetitive; that were matched to, as I say, trying
to identify someone’s address, trying to identify where
you could find this person to carry it out. I got an
email from one of these people, saying, “My accounts
have all been suspended. If I apologise to you, can I
have them back?” I looked at what this man had been
sending and thought, “No, I’m just going to help the
police make sure they have this information about
you” because it was so different.
Prior to this, I have had one gentleman who, I would
say, had edged on this sort of behaviour and indeed
went after my mother, who was also on Twitter. That,
for me, was an absolute red line, and prior to this
summer he was the only person I had ever blocked. I
think that is one of the things I really want to get
across. As much as it is very easy in this debate to
think this is about people not wanting to hear negative
messages or not wanting to hear people disagree with
them, for me there is a very clear divide. Prior to this
summer, I had never blocked anybody, except for this
gentleman who had been so utterly offensive to my
mum, because I felt it was important to be open.
I felt, as much as it is sometimes unpleasant what
some people might say to you, it was important to
show that you are willing to debate and discuss and
engage with people. A rape threat is not debate,
discussion or engagement. A repeated rape threat or
a repeated bomb threat, is not debate, discussion or
engagement. It is designed to shut down and it is
designed to intimidate, and we need to treat these
things differently, just as we would in a pub. If
someone came and said this stuff to you, you would
draw a line between that sort of behaviour and banter
and maybe someone being a bit off. You would make
those clear distinctions.
Q65 Tracey Crouch: I have not had anything as
similar to this experience as you, but I have had death
threats and lots of abuse in that sense. Like yourself,
I have never blocked anybody, but Twitter’s response
to you from this was to just block people. Your
response to that was that a rape threat is not a matter
of bad manners. Do you think that Twitter were quick
enough to respond to your concerns? Do you think
they were helpful? Do you think they could have
done more?
Stella Creasy: I think Twitter themselves would say
they did not get it right. What is interesting, having
spent some time now talking to them about these
issues, is they are on a journey themselves about how
different people in different countries are using their
technology. I think they absolutely get that that
weekend, where people were reaching out and saying,
“Look, this is an attack”—the initial thing was, “Well,
report abuse.” There was a lot made of an abuse
button. If you are getting 50 rape threats an hour, it is
not feasible to ask somebody to report every single
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 23
29 October 2013 Stella Creasy MP
message. That pressure alone on somebody, let alone
the technology, even if it is reporting, “Here is another
one, here is another one”, it is not feasible to do it.
Twitter are on a journey of recognising that that is
how people are using their platform, as are Facebook,
as are Ask.fm; that these things are evolving and how
people can use them to send those messages.
What I was very clear about is that if somebody is
willing to say that stuff online, we should not treat
it any differently from its being said offline, and if
somebody is thinking in that way, there is potentially
a problem because it is about this concept of
escalation. The argument I had with both the police
and the companies was not that it was for us to cope
with this or to just block people. There was something
about those people that needed to change. Therefore,
we needed to be able to identify those people to work
out whether they were on that path of escalation; that
the kind of person who was trying to post up a young
woman’s postal address so that she could be found
obviously had some issues, and to simply say it was
the responsibility of that young woman to keep herself
safe, rather than to deal with the threat that this person
posed, was not acceptable. You had to see it in the
context of what we know about violence against
women and what we know about that escalation
process.
Q66 Tracey Crouch: You said there were 50 rape
threats in an hour. That surely is an unusual event, or
do you think this is commonplace behaviour on
Twitter?
Stella Creasy: One of the other things I have asked
the companies to do is publish their data about the
numbers of reports of abuse they get and the numbers
of the concerns so we can get a question of scale. I
know that the Metropolitan Police over the summer
have said, since this all happened, they had had an
increase of 1,500 reports from people about this. I
welcome that because I think we have to recognise
the scale of what is going on and, therefore, the
potential risk. Talk to Women’s Aid or talk to Refuge
about the numbers of people they might deal with who
have been victims of violent assault or victims of a
former partner who is using this technology to commit
the crime, who is using it to continue to harass people.
I think to see that as somehow separate, or not to
have ways of identifying if somebody is using online
forums to harass and to continue a programme of
harassment misses a trick, because we would not let
that go offline. Women are subjected to a particular
type of misogyny online and I think that comes from
that culture of the internet. That has to change as
women come online and I do think that will change.
What I would say is that, as much as these
technologies have these problems, there are also great
opportunities to redress some of that and to have a
public debate about these issues.
There are also other victims of this behaviour. We do
not know the scale of how people are using these
technologies. My point is that it is not the technology;
it is the person underneath it we have to get to grips
with. If the technology allows us to identify that
person, that is a positive move to me because it allows
us to intervene at a much earlier stage. When you look
back at some of the histories of people who have been
victims of intimate violence, you can see points where
they have had this behaviour and everyone has said,
“Well, just don’t go on Facebook” or “Don’t go on
Twitter”. Leaving aside their own personal freedoms,
that is a missed opportunity to understand that the risk
to that person is escalating.
Q67 Mr Leech: I do not think any reasonable person
could try to justify the sort of abuse that you and
others have received in this case. It has clearly stepped
over a line, but there is also a very grey area on social
networking about what one person might think is
overstepping the line and what others do not. I
suppose my question is, how do we come to a
conclusion about what is acceptable and what is not?
How much of it is important for us to make sure what
we are doing online is very much on the right side of
the line, rather than the wrong side of the line, to set
a good example?
Stella Creasy: Could I just flip that back to you, John?
You said there is a grey area online. Do you think
there is a grey area offline?
Mr Leech: Yes, I do, absolutely.
Stella Creasy: That is exactly the point. It is not that
the online world is somehow different or separate or
distinct in terms of these questions about what is
offensive speech or what is a threat. We have laws
and we have guidance that the CPS and the police are
supposed to use offline. That we do not necessarily
think about how we can apply them online is the
challenge here, not that those debates do not exist
offline as well.
Q68 Mr Leech: Is it not the case, though, that more
people are prepared to overstep the line online, rather
than offline, because they feel a lot braver behind their
computer, rather than face-to-face with someone?
Stella Creasy: I think that is absolutely true. That does
not negate the point that this concept of a grey area is
not distinct to the online world. I do not think we
should be drawing distinctions between online and
offline behaviour. I think we should be trying to
understand how people are using different forums,
whether in person—I have had poison pen letters; I
am sure other members of the Committee have as
well—or whether online people are sending these
messages. There is a risk that we trivialise something.
One in five people in the world are on Twitter. That
is only going to grow.
This is not a separate, distinct world in which we have
to make up new rules. We have to make this work in
reality, in the messy reality, of the lives we live. You
are absolutely right that anonymity makes it easier to
be offensive to somebody, easier to threaten
somebody. Someone threatening to rape you or to put
a bomb under your house or posting up your address,
or coming around to your house, is as threatening and
as intimidating online or offline.
Q69 Mr Leech: Do you not think, though, when
people stretch boundaries online, it then encourages
some people who do not recognise boundaries to
overstep the line?
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 24 Culture, Media and Sport Committee: Evidence
29 October 2013 Stella Creasy MP
Stella Creasy: I agree, but I do not think that is to do
with being online. I think that is part of modern
culture. That is why we have guidance and we need
to make sure that we are able to understand the risk
that people might face and also to be robust about
defending the right to be offended. We have also
recently changed the laws around the right to be
offended, haven’t we? That will apply online as well.
We should not let the technology get in the way of
saying these are live debates in our society. What
people will say to each other in any different context
could be offensive or could be designed to intimidate
and we need a protocol and way of understanding
what that is.
Q70 Mr Leech: Changing the subject slightly, you
said earlier that it was not clear whether it was one
person making 50 threats or 50 people making one
threat. Do you think there is a problem with the
systems that the likes of Twitter or other social media
sites have in place, that they cannot necessarily
distinguish between whether it is one person with 50
threats or 50 people with one threat?
Stella Creasy: No, I think it is perfectly technically
possible to do that. Justin Bieber gets millions of
tweets a day. You would expect that account to get a
lot of traffic. I have talked to Twitter about the concept
of an online panic button. Because of this, I have now
had a panic button installed in my house. I have a
button that if I feel in distress, I can press and the
police know, because of the history of what has gone
on, that possibly my property might be at risk and,
therefore, to come around. Online, if you say, “I am
being attacked”, and your account is getting a high
level of traffic compared with what you might expect
it to get—so say the numbers of followers you have
or the number of people you follow, or the type of
language being used—those algorithms are not
particularly difficult to put in place. You have to have
the culture as a company that says, “We need to be
scanning. We need to be looking for that”.
One of the things I have tried to follow up with
Twitter is obviously if they are going to have an abuse
button and there is every concern, rightly, that people
might then—in fact, quite a few of the people on
Twitter said, “Right, we are going to tag everything
you tweet as an abusive message to see how you like
it”. But if people are going to use these technologies,
what do you do with the information? They hold an
awful lot of information about people already, in terms
of the account that you set up, where it was set up,
what kind of messages it is sending, for them to be
able to do that kind of analysis; to be able to say,
“There is a distinction between the Justin Bieber
account, which we expect to get millions of tweets a
day, and what suddenly is happening here. Is that a
cause for further investigation?” If my property was
getting a lot of reports, the police might flag it up and
go, “Hang on a minute, we might need to investigate
this a bit further”. An online panic button system
would allow someone to say, “Look, this is a
problem”, and for Twitter to say, “Oh, you seem to be
getting a lot of information here. Yes, maybe there is
something going on. Can we help you?”
Q71 Mr Leech: But my question was not about
whether it was technologically possible. It was
whether Twitter and others were using the technology
to its full potential and whether you think there are
other things that they could be doing to make that—
Stella Creasy: The equivalent of an online panic
button is absolutely something I have discussed with
Twitter. I would like to see other technology
companies having it. The answer to your question is
no. As far as I am aware, none of them do this at the
moment, although I think they are all thinking about
it because this is an evolving debate, particularly in
the UK, about how people are using these
technologies.
Still now somebody sends me an abusive message that
crosses that line. I still get one or two of these. Those
accounts get suspended very quickly. That process of
a kind of online temporary restraining order,
essentially, seems to me to be something that these
companies can explore, again if they have protocols
in place. One of the things I am concerned about is
for that not to be abused. In order for that not to be
abused, there needs to be a protocol for it. None of
these companies have this in place at the moment. I
think they are all starting to look at it because of this
public debate.
Q72 Mr Leech: Why do you think they have been
so reluctant to put these sorts of things in place? Is it
just because of an attitude about free speech or is it
more about the cost?
Stella Creasy: I think we have all seen these issues as
issues around malicious communications, rather than
understanding that there could be other forms of
behaviour at stake here. How you would then act and
how you then, as a company, would abide by local
laws, as they all say they would, is very, very
different. If you are only being asked to deal with one
type of offence, which is malicious communications,
then you need a protocol to be able to look at the
communication and decide whether it is malicious. If
you are being asked to deal with a harassment case
where somebody might be sending the most
innocuous messages, but it is the contact and it is the
fact that somebody is continuing to contact somebody,
continuing to set up accounts, continuing to pursue
somebody that is the problem, that is a very different
type of issue.
I do not think these companies have even begun to
scratch the surface of how their technologies are being
used to commit those sorts of offences and so what
they might be able to do to help protect their users to
make it a safe space. The question for all of us as
public policymakers is, can we get them to have that
conversation and do it in a way that is fair and
proportionate so that we do not close down the
internet; we open it up for more people because it will
be a safer space for more people?
Q73 Steve Rotheram: There is not a country in the
world where freedom of speech is absolute. Before we
get those people perhaps who are listening or reading
the transcript claiming that somehow we are trying to
close down freedom of speech, let me say it is
absolutely not the case, and, in fact, as you just
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 25
29 October 2013 Stella Creasy MP
mentioned, we are trying to defend freedom of speech
and encourage more people to take part in the likes of
Twitter debates and stuff. You spoke earlier about
your personal experiences of being a victim of
harassment on Twitter. In your view, are the laws to
tackle harassment fit for purpose in the online world?
Stella Creasy: Thank you, Steve, and I just want to
pay tribute to the work I know you have been doing
about trolling as well, because it was you and I who
first started, well before this ever happened to me, and
I never ever thought I would need to know quite as
intimately and directly what the law would be. I think
the honest answer is we do not know. Obviously the
protection from harassment law was updated about a
year ago. We have seen the first initial statistics about
it. That is not just about online behaviour. I think there
is a way to go for matching that understanding of what
forms of behaviour online are similar to behaviour
offline and, therefore, potentially problematic for our
police and for the technology companies. I think we
have to press to do that because one of my concerns
is, from a public value-for-money perspective, the
sooner we identify these challenges, the sooner we
can get the right training in and the sooner we can
prevent some of the crimes that might end up being
very, very costly. There is a real value in getting
everybody up to speed in the way in which these
technologies can be used.
I think the Protection from Harassment Act has a way
to go in terms of its application. There are concerns
for me about the 4A, 4B—I apologise to the
Committee if that is too technical—about how it can
be implemented. We have to see. The initial figures
look a bit mixed about whether that has been used to
truly tackle these problems. I think we have legislation
at the moment, but we do not necessarily have the
culture at a local level to be able to put it into practice.
Certainly I found, when I was trying to get the police
to engage in this debate, that I kept getting put through
to the person who held the Twitter account for a
particular part of the police rather than somebody who
could understand that on a Friday evening there was
potentially a young girl whose address could be on
the web and she could be the target of all sorts of
malicious behaviour.
That is not about having separate departments or
separate specialist people when this is so much a part
of everyday life. You want your police to know the
rules of the road. You want your police to know how
the internet is being used and to be able to understand
it, because that is what they are going to have to
police.
Q74 Steve Rotheram: Do you believe it is a problem
with shortcomings, in effect, of enforcement, rather
than in the legislation itself?
Stella Creasy: I think we will not know whether the
legislation matches what we need to be able to deal
with until we have more enforcement of it. It is just
too early to tell. I think there is a wider point about
understanding both the concept of escalation, the
concept of harassment, and then how that might
happen online as well as offline and what you can do
about it, and what you can ask companies to do about
it already. What I would say is that, since the summer,
I have noticed the police are willing to engage with
those threats. They need help. I think that is why we
have talked about having cyber specialists coming into
the police, because this is about a set of skills
matching a cultural change.
Q75 Steve Rotheram: You mentioned Claire
Hardaker earlier, who is doing some fantastic work on
looking at perhaps the reasons behind why the culture
is so different online and offline. Just to broaden that
out a bit, are there any changes to legislation you
would like to see to combat harmful or illegal content
on the internet, and not just in the social media field?
Stella Creasy: From a personal perspective, I think
some of this is about the CPS and how they respond
to the powers that they have. I would like to see the
CPS being very proactive about drawing distinctions
and defending that right to be offended, that right for
people to be close to the bone—what is lovely about
British culture is that we have that stream of quite
out-there comedy, out-there satire and out-there debate
and discussion—and, in order to defend that, saying,
“Here is the stuff that crosses the line. Here is the stuff
that we have to be clear is unacceptable, is illegal. It
is illegal offline. It should be dealt with the same way
online and we will act on that to make sure that we
draw that boundary”, so that we can have all the
lovely banter, the jokes, the debate, the discussion and
the humour that is so much part of British culture.
It is genuinely difficult for me to tell on a personal
level, until we have seen this case through, whether
the law matches that standard. The messages that
Caroline has received, for me, certainly breach that
level. I would want to see action and I want to see the
people who sent those messages being held to account
because I have seen the damage that it has done. I
have seen how affected she has been personally by it
and that is what those people were trying to do, and
that is an offence. It is an offence offline and it should
be an offence online.
Q76 Steve Rotheram: I agree with that critique of
current legislation. I am beginning to believe perhaps
there is slightly too much in regard to things falling
between the cracks. I have identified at least seven
different Acts of Parliament that can be used in the
online world for those people who are trolls or who
do these despicable acts. Obviously all that legislation
was before Facebook and Twitter and there is no
mention of trolling. It is not identified. The word
“troll” is not even identified in the legislation. Perhaps
what it needs is just a clarification, rather than a
reworking of any of the laws we currently have.
Stella Creasy: Yes. I have been repetitive with the
Committee about this. I do not want to draw a
distinction between online and offline because that
makes it somehow mean that, if these people were
doing it offline, it would be easier to do something
about it than online. That is just about ease of access
to them, rather than necessarily the technology. We
have laws. If they are not necessarily being applied,
we have to work out what it is about the way in which
the online world operates that makes it harder for
people to see the connections between those things. I
am also resolute that it is not for those people using
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 26 Culture, Media and Sport Committee: Evidence
29 October 2013 Stella Creasy MP
the internet, like Caroline, to cope with this pressure.
It is for us to find a way of helping to bridge that gap
and change the debate on it.
Q77 Mr Bradshaw: Do you know how many arrests
there have been?
Stella Creasy: I believe there have been five arrests.
Mr Bradshaw: Do you expect there to be
prosecutions?
Stella Creasy: On a personal level I think the case is
there, but it is for the police and the CPS to make that
decision. I absolutely respect that. The evidence and
the behaviour seems to me to fit everything I have
ever seen but, again, I am the victim here in that sense
and so it is not for me to make that judgment call.
Mr Bradshaw: But you agree that an important part
of the law is to act as a deterrent and that prosecutions
could act as a deterrent in this case?
Stella Creasy: Prosecutions would be about holding
people to account for their behaviour. I do not think
there is any disagreement about this, that that kind of
behaviour over the summer was absolutely horrific,
was completely illegal. The question for all of us is
whether we are able to meet the test to make an
effective prosecution.
Q78 Mr Bradshaw: What this case seems to have
exposed to me, which is very depressing, is an
underlying current of violent misogyny that is still
there in society. The way you deal with expressions
of this is, in other media, through the law. It is not an
exact parallel, but I suspect that people are now much
more careful about what they say about other people
following the action against the Speaker’s wife, for
example. We are learning how to use Twitter. What I
am saying to you is I hope what this case achieves is
a change and that is what I am looking to see: whether
you are optimistic, whether this could act as a
positive change.
Stella Creasy: I have great optimism about the future
being more equal online and offline, because one of
the things these technologies and communications do
is allow us to have those debates. It is like the No
More Page 3 campaign. It has been entirely run online
and that has opened up a space to say, “The impact of
seeing these images on women is X and this voice
needs to be heard”. We would not have had the
capacity to communicate in that mass way before. As
much as, yes, prosecutions can act as a deterrent, I
also think what has been positive is that there has been
a broader debate that says, “This is completely
unacceptable”.
I found it very striking that, in retweeting the
messages that I considered to be illegal, quite a lot of
people said, “I had no idea this was happening”; in
the same way that the EverydaySexism Twitter
account and all that campaign—again, which is all
online—has allowed people to see a side of society
and to see some of that misogyny that people had
never really engaged with or become aware of. I
would love to think that a prosecution would be—
and I do not think addressing misogyny, addressing
inequality, is going to take that cultural debate. What
is fantastic about these technologies and why I am so
passionate about making sure we have free speech is
that they allow a vehicle for some of that change to
take place as well.
Q79 Jim Sheridan: Stella, many, if not all of us, are
discovering the nature of our job attracts through
social media and emails some of the unsavoury
characters in the world.
Stella Creasy: Yes, I am sorry about that. My mum
just gets bored. She sends emails. I have told her.
Jim Sheridan: Does she send them to me?
Stella Creasy: I would rather she was sending them
to you than me, I tell you; some of the things she
tells me.
Jim Sheridan: No, but seriously, as I say, we do
attract so many nutcases in the country, but what you
have described goes way beyond anything that should
be allowed. Just in terms of legislation and to follow
on from Steve’s point, the 2008 Criminal Justice and
Immigration Act, it sounds to me—this was before
social media kicked off—that it is something we could
revisit and look at again and the value of it, if indeed
that is still working. The main question I am asking is
about the international scene; for instance, in countries
where this kind of behaviour may be legal. What can
we do to stop it coming from countries that might find
this sort of behaviour acceptable?
Stella Creasy: I have no reason to believe that any of
the people sending messages to myself or Caroline or
indeed any of the other figures over the summer were
from anywhere other than the UK and, indeed, all the
prosecutions have been in the UK. One of the
suggestions that one of the reasons why we can’t
bridge that gap between what our law does offline and
online is the global nature of the internet. I do not
necessarily think that is the case. For these technology
companies, the awareness of how people are using
their tools to create crime is an interesting question
because that then could apply to other jurisdictions in
terms of if you were to see escalating behaviour. It is
interesting. The woman who is in charge of online
safety at Twitter comes from a domestic violence
background and said to me, “Oh, yes, I totally get the
point you are making”, and you are thinking, “Okay,
so why don’t you have the systems in place to then
be able to address this?”
I think we have to be wary of the idea that, because
the internet is global, there can be no local response
or no local engagement on any of these issues.
Nothing has disproved that to me so far. We also have
to be very careful and one of the things that is so
important about the internet remaining a space where
people can have freedom and anonymity is that it has
also been a force for good in terms of human rights
issues. It has been a force for good in terms of giving
a space for people speak freely, where they are not
given freedom in their own countries.
We have asked Twitter to make it clearer that
harassment was not acceptable in their guidelines.
That cuts across all their jurisdictions. At that point,
Toby Young accused me of being like a Chinese
dictator for suggesting that harassment should not
happen online. I wanted to point out to him that this
was already illegal in this country. What is harassment
may well be in the eye of the beholder in some
countries and, therefore, it is about the systems that
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 27
29 October 2013 Stella Creasy MP
these companies have to assess the complaints that
they get, which is why we need to see the scale of the
complaints. In the UK they say they comply with local
laws. We have the Protection from Harassment Act.
The question is, how are they then using that
information to be able to make sure that they are
responding to any concerns that are raised with them?
They have a way to go yet. So do a lot of companies.
That Twitter are engaging in this debate for me is very
welcome. I would like to see that then develop.
Q80 Jim Sheridan: If you are receiving horrific
emails or tweets or whatever it may be, for me it does
not matter where it comes from. Do you think there
is anything more that the providers can do to stop this
in a global market?
Stella Creasy: As I say, what is interesting is whether
it is happening in other countries and I think there
are cultural variations in how people are using these
technologies around the world. These companies are
also growing. Twitter did not exist five or six years
ago, so it is growing as an organisation to have the
capacity to deal with it. They now have a new team
of people looking at online safety. I would expect
them to take the lessons they are learning from the
UK and think about whether they can be applied to
any form of behaviour on their platform around the
world. That will be a question for Twitter. It would be
a good question if you guys are speaking to them
about it. What we have to focus on is how we keep
our systems safe in the UK. How do we make sure
that every system in the UK has the freedoms that we
want them to have, online and offline? How do we
make sure, when we get instances like this, that our
own local laws are being used to best effect to open
up that space?
Q81 Philip Davies: I would like to think of myself
as a great champion of free speech, but clearly what
you and Caroline and others have faced is beyond
what anybody would hopefully think was acceptable.
I just wondered whether you felt, given the scale of
Twitter and the pressures on, for example, the police,
this is something that is just going to be in many
respects unsolvable through the normal course of the
law in the sense that, if Twitter is already at the extent
it is and is going to grow further and police resources
are so stretched, is it going to be the case that the
police, even with the best will in the world, will
simply not have the resources to be able to pursue all
these particular cases?
Stella Creasy: I appreciate what you are saying,
Philip, about recognising that this is about the freedom
of speech of people like myself and Caroline to not
be subjected to this kind of behaviour.
Philip Davies: Absolutely, yes.
Stella Creasy: I think the question for all of us is,
what does that behaviour embody? The person who
uses this platform to repetitively send those kind of
messages—what else might they escalate to do? How
else is this connected to other forms of crime? That is
the way to look at it, which is not to say that somehow
there is something unique about online behaviours but
to ask where they interact with offline behaviours.
When you do that and, as I say, when you look at
some of the cases around stalking and harassment and
you see people getting messages online and people
saying, “Well, just don’t go on to Facebook” or “Don’t
use these forums”, rather than recognising that the
person who is sending those messages is fixating on
somebody, could then become an even more severe
risk to that person, and the sooner we intervene, the
sooner we might be able to address that person’s
behaviour, prevent in the future much more damaging
and indeed more expensive and resource-intensive
crime from taking place.
Keeping people safe—there is a gender issue in this
and it is disproportionately women who face these
challenges—will be both better for the public purse
and receive better outcomes. I understand the concern
about resources. What I am saying is that the online
world is an opening to seeing and identifying crime
and detecting it quicker. We should see that as an
opportunity both to save money and get better
outcomes, rather than another set of issues to deal
with.
The other example from this for me was that the EDL
were using Twitter and Facebook to organise coming
to cause problems in Walthamstow. There was a timelag between the police understanding that, seeing that
information and engaging with it—in the same way
during the riots we saw a time-lag between the police
engaging and interacting with that information—to be
able to then use the online world to detect and prevent
crime. There is a big value-for-money argument about
how, if you understand that is how people are using
these technologies and you have a way of dealing with
them, you can get things at source in a way that if you
wait until much later on, they will become much more
expensive. We are nowhere near there yet, and in fact
what people are trying to say to me is, “Well, don’t
we need specialist units?” which to me is a bit like
saying, “Don’t the police need a specialist unit for
dealing with the telephone?” This is so much a part
of everyday life, we need our police to be able to use
their own personal experience of it to understand how
it might then lead on to further crime.
Philip Davies: Obviously I am delighted that there
have been some arrests and I hope there will be
prosecutions and convictions and, like Ben Bradshaw,
I hope that that will act as a deterrent so that people
realise that they can’t get away with this. Hopefully
that will be the case. I just wondered whether, when
you have received these threats and people like
Caroline have received these threats, which must be
horrendous, you felt—and do you know whether
Caroline felt—that they were genuine threats in terms
of people were intending to carry out these things that
they were saying or whether they were just
unacceptable abuse?
Stella Creasy: I think if someone is trying to post up
your home address, Philip, that is pretty frightening
and that is very, very distressing to somebody. To then
receive continued threats, continued aggression,
continued details about what someone is going to
do—these people went into tremendous detail about
what they might do to my dead body once they had
killed me and once they had done horrible things to
me and how they might do it again. I defy anybody
not to find that deeply threatening, deeply unpleasant.
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 28 Culture, Media and Sport Committee: Evidence
29 October 2013 Stella Creasy MP
This is the thing about the Protection from
Harassment Act. It is not about the content. It is about
the impact on the person. Nobody can be unaffected
by receiving that kind of avalanche of threatening
behaviour over a persistent period of time, to not
think, “Actually, is that person outside, who I don’t
know, the person who has been sending those
messages?” The thing about the anonymity of the
internet is that it could be anybody sending those
types of messages to you. It becomes even more
distressing in a way, because you do not know who it
is who feels the need to send you that sort of message.
Q82 Philip Davies: No, I do not disagree with any
of it. I am simply playing devil’s advocate. I am trying
to think about some of the defences that some of the
internet social media sites might give. I was just
anticipating what you thought if they were to argue
that, if there are these people out there and they are
serious about these threats, it is better for the law
enforcement agencies that they are out there in public
so they are able to be traced, rather than to be still
there but untraceable and unknown; that in some
respects it would be easier for the law authorities to
deal with these people, if they are serious about doing
this thing, if it is up there for them to trace.
Stella Creasy: We should use people as bait to draw
these people out?
Philip Davies: I am merely asking for your response
to what may be a line of defence that these social
media companies give.
Stella Creasy: Here is the thing, Philip. Surely what
matters here is the victim. The victim was Caroline
and her life was dramatically changed by the
behaviour of these people. Our job is to make sure
that nobody’s life should be curtailed and nobody’s
free speech should be curtailed by the behaviour of
these people; that it is not for Caroline to cope because
“at least we know who these people are”, as a
containment strategy. It is for us to say both in law
and in culture, “This has to change, because it is not
part of the society we want to live in”.
Philip Davies: No, I agree. I was not giving you my
opinion. I was trying to pre-empt what people might
say.
Chair: I think we had better call a halt at this point.
We have another two panels. Stella, can I thank you
very much for coming?
Stella Creasy: Yes, and can I just thank the
Committee for looking into this? I think it is difficult
to get right, but it is welcome that you guys are
looking at how you could get it right and how you
could support both the police and the technology
companies to work with us.
Chair: Thank you.
Examination of Witness
Witnesses: Nicholas Lansman, Secretary General, Internet Services Providers’ Association, Dido Harding,
Chief Executive Officer, TalkTalk Group, and Hamish Macleod, Chair, Mobile Broadband Group, gave
evidence.
Q83 Chair: For our second session, can I welcome
Dido Harding, the Chief Executive of TalkTalk Group,
Hamish Macleod, the Chair of the Mobile Broadband
Group, and Nicholas Lansman, Secretary General of
the Internet Services Providers’ Association, and
invite Gerry Sutcliffe to start off?
Mr Sutcliffe: The topic of our inquiry is online safety
and, notwithstanding the evidence we have just
received, we recognise we are talking about a growing
situation here, not a closing down, and that is not our
intention. Clearly there are responsibilities and there
are issues that flow from that in terms of what needs
to happen. It would be interesting to hear, right from
the start, what you do already to prevent access to
harmful or illegal material online; just a general
flavour of what it is about at the moment. Whoever
wants to start off.
Nicholas Lansman: Shall I kick off with some
general points?
Mr Sutcliffe: Yes.
Nicholas Lansman: First, you have to treat illegal
material very separately from legal, which might be
age-appropriate. With illegal material, right across the
industry, whether mobile or internet, have done a lot
of work over many years here, primarily setting up
the Internet Watch Foundation in 1996, and illegal
material is dealt with in a variety of ways. The priority
is to remove illegal material from the source. In the
UK we now have much less than 1% of illegal
material that is hosted in the UK. Secondly, to try to
remove it even when it is abroad and, thirdly, to
create, with the IWF, a blocking list so that ISPs can
block illegal material before people can stumble
across it. That is illegal material and what I mean by
that is child abuse material, but also terrorist material.
I think the industry has been extremely effective in
dealing with that side of things.
In terms of age-appropriate material, there have been
huge developments more recently and I will allow
Dido to contribute to that discussion. Principally, there
have been developments in terms of creating an active
choice for customers when they purchase broadband
to make that decision of whether they want to engage
filtering. Filtering is now going to be provided for free
and a whole-home solution. It is very important that
it applies to all devices in the home, whether it is PCs,
laptops, tablets or indeed smartphones. Those filtering
technologies will be easy to use.
Q84 Mr Sutcliffe: Easy to use, but is there not an
education and development process to go alongside
that? It is all right having the technology to deal with
it, but how do you encourage people to be active?
Nicholas Lansman: That is a very good point.
Technology alone is not the only solution and
education and awareness is fundamental to the
process. ISPs over many years have provided material
on websites, on leaflets. They have also been visiting
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 29
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
schools to educate teachers and parents. There is also
the UK Council for Child Internet Safety. We also
have Safer Internet Day. A great deal is being done to
educate. Still more needs to be done and this is about
creating the tools to ensure that parents become
confident with the use of the internet and that they can
also communicate that and have that dialogue with
their own children.
Dido Harding: I have said a number of times that I
think internet safety is the road safety of our
generation. When I was a child, my parents were
advertised that with Clunk Click. We had the Green
Cross Code. Seatbelts were just being rolled out into
cars. The airbag had not yet been invented. You had a
whole load of technology innovation, public
awareness and education in schools. I can remember
the Green Cross man coming to my school. Internet
safety is as broad an issue, if not broader, than road
safety for our generation. There is a lot that is
happening on the technology front. There is a lot that
is starting. We in TalkTalk have lobbied heavily to see
internet safety in the primary curriculum and we are
delighted to see that it will be coming in. There is
a lot that needs to happen in education and public
awareness. There is not one single bullet. We all have
a role to play in making sure that the digital world is
safer than it is today.
Q85 Mr Sutcliffe: I like the analogy of the road
safety because I think you are right, but part of the
problem that we have, in terms of the evidence and in
our own lives, is that parents do not go in to watch
what the kids are watching online and things like that.
The analogy about if a child had a TV set in the
bedroom, years gone by, there would be an issue about
what they were watching. What is concerning me is
that the technology is there, but there is not the
cultural change in terms of parents being able to
understand what they need to be doing.
Dido Harding: Nicholas mentions filtering. At
TalkTalk we have had a whole-home filtering product
called HomeSafe available now for just under 2.5
years. We have about 1.2 million of our 4 million
customers use it and 400,000 customers have
activated the Kids Safe element of that, choosing to
block certain content in their home. Just as you say,
we have learnt that it is important to keep it simple
because most of us find the technology baffling. You
do have to have something that covers the whole
home and you also have to work hard, and all four of
the major ISPs have committed to work on a public
awareness campaign, because all of us are learning
how to use this. It is not just enough as an ISP to
provide the gadget. You have to help people
understand how they might use it. The most valuable
thing you can do as a parent with young children, and
I say this as one, is talk to your children about what
they do online. If you don’t engage with what they
are doing online, no technology is going to help them.
Mr Sutcliffe: I think that is spot-on.
Hamish Macleod: Can I just make a mobile
contribution? As far as the Internet Watch Foundation
and the illegal side of things goes, our approach is
very similar to the fixed ISPs. We take the list. It gets
updated twice a day and any attempt to access sites
on the illegal list will be frustrated. As far as the ageinappropriate content is concerned, people first started
to use the mobile to access the internet about eight or
nine years ago. At that time we published a code of
practice and part of the code was that we offered a
filter for the internet browsing service on the mobile.
For all prepaid phones, though, the filter was put in
by default. For most contract phones it was also put
on by default and in the next 12 months for all
contract phones it will be put on by default.
Mr Sutcliffe: The filters will be basically free? They
will come as standard?
Hamish Macleod: Absolutely, there is no charge for
them.
Q86 Mr Sutcliffe: Is there a need for consistency of
approach across the range of providers and is that
happening?
Hamish Macleod: The approach we have taken on the
mobile is that, when we first published the code back
in 2004, we appointed a body called the Independent
Mobile Classification Body to achieve this level of
consistency. It was a subsidiary of the premium rate
regulator, who was called ICSTIS at the time. The
reason for that was that most of the content about
which there might have been some concern was
premium rate adult content, so it made sense to work
hand in hand with the premium rate regulator. We
carried out a review of that earlier this year and, as a
result of that review, we have appointed the British
Board of Film Classification as our independent body
who give us editorial advice as to how we should
calibrate the filters.
Q87 Mr Sutcliffe: What are the linkages like with
law enforcement agencies and trying to help them
with the speed of the technological changes that are
taking place? Do you think that is working well or
could more be done?
Nicholas Lansman: I think it is working very well.
There is a long-standing process in place where,
according to the regulations, ISPs will be approached
by law enforcement to hand over data and that is done.
It seems to work well and we know that, in a variety
of cases, law enforcement have taken and followed
those investigations and will make prosecutions where
people are behaving illegally online.
Hamish Macleod: Can I make one point on that,
though? A few months ago there was a summit held
with the major ISPs and mobile companies about the
work of the Internet Watch Foundation and whether
they should be more proactive in seeking out illegal
images on the internet because, up to that point, the
memorandum of understanding specifically with the
CPS was that you could only respond to public reports
and then you could go and look for the stuff. Now we
are moving to a point where they will go out and look
for the stuff. That should be more efficient because,
of the level of public reports coming in, only about
one in 10 of them are illegal. That should be more
efficient and should lead to fewer public reports, we
hope, because they are getting to the content before
the public does.
However, the IWF is part of a safeguarding
ecosystem. They are not an island. They are being
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 30 Culture, Media and Sport Committee: Evidence
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
more proactive and generating more intelligence to
pass either to our own law enforcement agencies or
internationally, to overseas. That will generate more
work in other parts of the ecosystem. If we are going
to put more resources into the IWF, more resources
need to go into other parts of the system too,
otherwise we are just creating traffic jams in other bits
of the net.
Q88 Mr Sutcliffe: It is good to hear that there is a
level of consistency there now. You have talked about
the easy work, if you like, the illegal, which is clearly
identified as illegal and, therefore, it has to be stopped.
One thing that concerns me is the cultural issues that
are around. In evidence at the last session, 30% of
children under the age of 12 thought it was okay to
be bullied or had experience of being bullied on the
net. Notwithstanding the massive amount of work that
it has done to deal with the illegal situation, how do
you deal with the cultural changes and trying to get
different responses across because of some of the
problems that people are already facing through the
surveys that already exist?
Dido Harding: I think step by step would be my
rather simplistic answer. The number of things that
can happen to a child online are just as many as can
happen in the playground, only it is more immediate
and they have the whole world at their fingertips. The
wonder of the internet, as Stella Creasy set out, is that
it has a double-edged sword element to it. That is why
I go back to road safety. We have to think about all of
the different types of behaviour that happen in the
playground.
On bullying, I do not think that the fundamental issue
is any different from being in the playground. The
adult supervising the playground has a role to play.
The children have a role to play, as do their parents,
and there is a bigger public awareness challenge on
all those dimensions because it is so much less visible
than your child being physically bullied in the
playground. I don’t think you can overstate the
importance of people such as yourselves thinking
deeply about this. The social and moral framework for
the digital world does not exist yet.
I do think there is quite an important role for
legislators to play in thinking through where the
absolute black and white rules are that we need to put
in place. There is a clear role for all of the owners or
builders of the playgrounds, whether those are the
social networking sites or the internet service
providers themselves, to think through, “How do you
provide parents, teachers and children with the tools
to make that environment safer?” There is a clear role
for all the educators to make sure that we are keeping
up with that technology, not just for our children but
in educating parents and teachers themselves.
I wish I could say that there was a magic bullet. There
is not. We are the generation that has to quite calmly
and logically think through each of these issues.
Q89 Chair: Without wishing to stray too far into the
data communications debate, which would keep us
going probably for the rest of the day, can I ask you,
first of all, how much data in terms of which
individuals are accessing sites, downloading images
and sending abusive messages do you keep and how
easy is it for the police, for instance, to have access
to it?
Dido Harding: I am very happy to answer for
TalkTalk and you can fill in the gaps. Obviously we
work under RIPA. From a TalkTalk perspective, we
currently provide call data, IP addresses, subscriber
details both for our mobile customers and our fixed
customers, session details, so how long they have
been online, IP log history and data access setup, so
how they are accessing it, and we will also provide
email addresses, payment details, customer notes and
email headers. All that is governed by RIPA. We store
those data in the same way that we would store our
own customer data. We retain data for 12 months. We
respond as quickly as all the Government authorities
accessing through RIPA require of us, and are
compensated for them for doing that. I think that
process works reasonably well. I would be
disappointed if you heard otherwise from those
authorities.
Chair: Would that cover Twitter postings?
Dido Harding: No. We do not store browsing history;
so exactly where you have gone and what you have
done, which pages you have looked at, we do not store
any of that. That is not governed by RIPA today. I do
think it is important that, in this, the ISPs are not
choosing their own rules. This is something that is
clearly for you, not for my colleagues and me, to make
the judgments. We act absolutely within the letter of
RIPA and store the data that is required of us there.
Q90 Chair: RIPA itself is now a little ancient in
terms of the speed with which things are advancing.
Do you think it needs to be revisited?
Dido Harding: I think that the debate that is currently
ongoing is very important, precisely as you highlight,
Chairman. The tools and techniques are changing so
much. Yes, I think it is very important to have that
debate and to make sure that we keep up to date.
Chair: Hamish and Nicholas, do you wish to add
anything?
Nicholas Lansman: If I can just add to that, as I was
explaining, that is a very similar setup to the other
internet service providers. I think it is important that,
with an extension to the data requirements, the Home
Office make their case and pass legislation and that is
fully debated. It is very important to note that the
ISPs, as they comply with RIPA, will comply with
amendments to it if that is the case. Hamish made a
point earlier about the fact that we have a lot of data
and I think in a previous session Peter Davies, the
CEO of CEOP, outlined the fact that there are possibly
thousands of potential investigations that perhaps
should be taking place if resources would allow. Just
collecting more and more data that may well be useful
will point out an even greater problem in the lack of
resources to follow up.
Hamish Macleod: I do not have anything in particular
to add to that.
Chair: Basically your industry follows the same
practices that TalkTalk has described?
Hamish Macleod: Yes. There are a few technical
differences but it is, policy-wise, exactly the same.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 31
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
Q91 Chair: It does not matter if I am using a mobile
device or a fixed broadband device? The same
provisions will apply?
Hamish Macleod: More or less. I am not entirely sure
what the fixed situation is. In mobile there is not
necessarily a one-to-one mapping between the IP
address that is being used and the user because we do
not have enough IP addresses to go around. We use
something called NAT, network address translation—
he says quietly—and that generates an absolutely
enormous amount of data.
Q92 Chair: If I used a mobile device in Starbucks,
are you able to identify that it is me?
Hamish Macleod: Connecting how?
Chair: On a public wi-fi system.
Hamish Macleod: I don’t know I am afraid.
Nicholas Lansman: I am pretty sure that the
providers of the public wi-fis, whether it is in
Starbucks or anywhere else, are also covered under
the legislation. Whether it is BT or other cloud
providers, they will be able to provide that data to
enforcement. It is worth underlining—
Chair: But it is not much use to know the person who
has accessed inappropriate material from Starbucks on
Victoria Street.
Nicholas Lansman: I think another discussion that we
will possibly have is the issue about anonymity and
whether that can mean you can’t trace people. That is
not the case. As we found in the previous session,
people commenting on Twitter and making comments
are subject to prosecutions, just like anyone else. It is
worth underlining that our systems have been in place
for quite some time and work well. We have single
points of contact within the Home Office. They can
be in contact with internet service providers simply.
The system effectively works well.
Q93 Tracey Crouch: Could I just follow up, just to
be clear in my mind, Ms Harding, on what you were
saying about collecting all the data and keeping it?
That is available upon request under the legislation,
but if somebody was accessing illegal content, such
as child abuse images, does that flag up in your
systems and you can then pass that over, or does it
have to come from the enforcement agencies before
you give that material?
Dido Harding: As it currently works, we respond
reactively to requests from CEOP. It is one of the
things that we have been discussing with the Internet
Watch Foundation as part of moving the IWF
activities to being more proactive; so proactively
identifying sites and then potentially proactively
identifying offenders. This is all fraught with moral
hazard and, in doing something like that, I would only
want to consider it with the full legal authority and
not as an ISP decide for myself that I fancy doing it
with my customers.
I feel very strongly the weight of responsibility as an
internet service provider in that our customers place
their faith in the fact that it is their data, not our data,
to choose what to do with and, therefore, we need a
clear legal framework on what we store and what we
do not store. That is why I say we store the
information that we are required to by RIPA and we
store the information that we need in order to serve
our customers. We do not keep browsing history of
where our customers browse every day of the week—
that is their data, not ours—unless or until there was
a change in legislation that required us to.
Q94 Tracey Crouch: But you are able to keep
internet browsing and other communications data
under the Data Retention (EC Directive) Regulations
2009, I believe. You could, if you wanted to, keep
that data.
Dido Harding: Technically we could. There are
different views from the lawyers on whether our
customers would be expecting us to store browsing
history. For example, when we launched HomeSafe
and when we first trialled HomeSafe, which is, based
on where you are browsing in your home, you have
set a series of filters that in your home, for example,
suicide and self-harm sites can’t be seen. That requires
us to store for you that you want suicide and selfharm sites blocked. A number of organisations and
the ICO themselves were very concerned that that did
not mean we were storing the history of where you
were browsing. We went through a number of stages
to give the ICO confidence that we were not storing
unnecessary information that our customers would not
necessarily want us to. That is why I say I am very
cautious about what data we store, if it is our
customers’ data as opposed to stuff that we need in
order to run the business for them.
Q95 Tracey Crouch: While I completely understand
the moral issues that you highlight about people
accessing sites that they want to—I can appreciate that
around, for example, online pornography—the issue
about illegal images of child abuse, for example, is
something completely separate.
Dido Harding: Yes, I agree. I do agree.
Tracey Crouch: If you are seeing information or data
from your customers, who are accessing sites that are
illegal, does that not have a moral duty on you as
a provider?
Dido Harding: I completely agree. If you are starting
to think through the stages at which it would be
appropriate to expect ISPs to pass that information
through, the easiest and most obvious is illegal child
abuse images, which are illegal to look at, illegal to
create, illegal to pass on, and it is a criminal rather
than a civil offence. I do agree with you and that is
why, through the IWF, we are working through a
process to do that. It is much less obvious as you work
your way back, because I do not think our customers
want internet service providers to store their history
of which car sites you have visited so that we can sell
it to car manufacturers for you to be bombarded with
emails about buying cars. It is very important that the
ISPs do not feel like it is a free-for-all to use that data.
Our customers do not expect that of us. We have been
thinking through, if it is entirely criminal, how we get
to a place where we can do that proactively with
CEOP.
Q96 Tracey Crouch: In our first oral evidence
session, we heard some shocking statistics about child
abuse images on the internet, talking about millions
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 32 Culture, Media and Sport Committee: Evidence
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
of images. 81% of those images involved children
under the age of 10, 4% of children under the age of
two, and over half of them were level four or level
five images. Do you think that the ISPs are being
effective at blocking these sites, these images?
Dido Harding: We are only as effective as
collectively through the IWF we are effective at
identifying the images. I think we all recognise that
the reactive approach, although it has taken the UK to
a much better place than a number of other countries,
has not solved the problem. Hence all the ISPs are
committing to increasing the funding of the IWF so
that the IWF can proactively go out and identify sites
that we could all block. No, I do not think anyone
working in the industry can say, “Yes, this is a job
well done”. It is a job that we know we have to do a
lot more.
Tracey Crouch: Do you have anything to add?
Nicholas Lansman: I think it is fair to say that there
is no complacency in this area. While the internet
industry is quite proud of having set up the Internet
Watch Foundation, it is an ongoing battle and I think
that several changes are taking place over time to
make the IWF more effective. It is true to say that, of
child abuse images, less than 1% are hosted in the
UK, but they are still going around the world. The
IWF also works with bodies abroad to make sure they
are trying to get them blocked at the source, because
I think that is the priority. Blocking is part of the
solution, but for the internet industry it is more
important that we try to remove this content at the
source, whether it is in the UK, more likely to be
elsewhere, in the US and so forth.
In addition to the extra resource that the IWF will be
getting through additional membership payments from
members, there is also a development in terms of
delivering splash pages. If someone accidentally
comes across child abuse images, rather than getting
an error message that just stops them accessing the
site they will now start to get splash pages that will
identify it is illegal material and they should not look
at it. These are some of the developments that are
taking place, but you are right. There is no
complacency and the internet industry continues to
work with the IWF, but also with CEOP and under
areas of law enforcement to make sure that the battle
to get rid of child abuse images and protect children
online continues.
Hamish Macleod: The Government has announced
that it is joining various international alliances to
tackle the problem. At the moment, the impression I
get is that it is all still a little bit aspirational and they
are just finding their feet. Those alliances need to have
clear deliverables and a clear action plan to improve
the situation because, for a great number of years now,
the vast majority of these sites have been hosted
outside the UK and international action and cooperation is what is going to lead to them being
removed and the people who are putting them up
there prosecuted.
Tracey Crouch: Do you think that requires a will or
money, or both?
Hamish Macleod: I am sure it will require both, yes.
Q97 Tracey Crouch: I understand that TalkTalk
automatically downloads the blocked list of known
websites containing illegal images on a daily basis. Is
that right?
Dido Harding: Yes.
Q98 Tracey Crouch: Two related questions. First, do
you think that a daily basis is fast enough and what
time of day do you do it? In my mind, if you are doing
it in the morning, by the time people are accessing in
the evening, probably new sites have popped up or
been created. It is just out of interest.
Dido Harding: I am ashamed to say I do not know
what time of day. I will check for you.
Tracey Crouch: I would be interested to know,
because it is such a fast-moving beast. Maybe I have
some sort of prejudice that these things are only being
accessed in the evening, but I just wonder—
Dido Harding: You are quite right. Let me go back. I
will get that for you. I can’t speak for the other ISPs.
For TalkTalk, it is currently a manual process. There
is a temptation to believe that telecoms companies are
all entirely automated. I wish we were. My goal is to
be able to automate it. Then I should be able to
consume it much faster.
Nicholas Lansman: I will check, but I understand that
twice a day there is an update to the list.
Chair: That is updated twice a day?
Nicholas Lansman: Yes.
Dido Harding: Then we consume it once a day.
Chair: You only do it once a day.
Q99 Steve Rotheram: I think Tracey has elicited a
response for the first question I had, so if we move
on. Is there a case for more effective filtering of legal
adult content, which could be obscene if it was viewed
by a child?
Hamish Macleod: The mobile filtering has been in
place since 2005 and it covers what we describe as
age-inappropriate. The framework that decides if it is
age-inappropriate or not is given to us by the BBFC
and it covers violence, pornography-type stuff,
behavioural stuff such as promotion of suicide,
promotion of anorexia, those sorts of sites; all the
topics that are of major public policy concern are
covered by their framework, whether that is obscene
or just 18-type content that you would see in the
cinema.
Q100 Steve Rotheram: Information is filtered. There
are filters. How effective are the filters?
Hamish Macleod: I believe they are very effective.
The accusations that are thrown at us more often are
about over-blocking of content, where we get a
handful of reports each month about accidentally
over-blocking. Again, part of the reason for the move
to the BBFC is to make that process a little bit more
slick so that if a website owner or a member of the
public believes that a site is being over-blocked, or
under-blocked, they are able to report it and the first
port of call is for their mobile operator to look at that.
Most of the time it is accidental and it is fairly easy
to speak to the filtering company and say, “Please can
you just change the categorisation of that website so
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 33
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
that it either is not blocked or is blocked, as
appropriate?”
Q101 Steve Rotheram: They are filtered at network
level?
Hamish Macleod: Yes, that is right.
Steve Rotheram: How about at mobile device level?
Hamish Macleod: Not very often, no. The products
that are available in the market to filter it at device
level tend to be a bit blunt. It is either about “turn on
the internet” or “turn off the internet”, or they
whitelist a very few sites or blacklist a very few sites.
From our point of view, it is an area that we would
like to see discussed a little more in the public domain
because a lot of the focus has been on what can be
done at network level, but what can be done at
browser level, what can be done at operating system
level, what can be done at the content provider level
or what can be done at device level? They all have a
contribution to make to this so that people can tailor
the services and the protections they want as best they
can, but we do the filtering at network level in
accordance with the framework that is independently
provided for us.
Dido Harding: I would say that there is still an awful
lot to do to improve the quality of filtering tools that
you give parents. We have had HomeSafe in the
market for 2.5 years. I would be the first to say that it
is not perfect and that there is a lot more to do to
make it something that is easier for parents to
understand and give some of the tools. Firstly, I think
it is important to remember that parents are not just
interested in blocking adult content. They are, if
anything, more worried about suicide and self-harm.
They want to block gaming or gambling. At certain
times of the day they want to block access to social
networking so that their children do their homework.
It is important to look at this more broadly than
simply the, admittedly, very serious issue of the
wealth of adult content online.
First, thinking about different types of content. Then
also we think whole-home network-based solutions
play a very important role, but they are not the only
solution. If you have teenagers and very young
children in the same household with multiple different
devices, you will want a more sophisticated filtering
solution that enables your 16-year-old to do their
homework, while it absolutely ensures that your threeyear-old cannot stumble on something that they are
not expecting. The parents themselves may want
access to some of those different sites.
We get customers complaining that they have turned
on HomeSafe and all the filters, one of which is for
alcohol-related sites, and that blocks Waitrose wines.
We know that the technology is not quite there yet.
When my competitor ISPs launch their network
blocking solutions, I am expecting that they will be
better than mine in some ways and that will be a good
thing. This is still an area where innovation can get
you a march against your competitors and there is real
consumer demand for tools that are simple to use but
meet their needs. I would want the Committee to feel
that both mobile and fixed-line telcos recognise this is
something that we have to keep working at, because
our consumers tell us loud and clear that we do not
quite have it right yet.
Nicholas Lansman: Just a couple of issues. I have
previously mentioned over-blocking and underblocking, which is taken very seriously, but I think
it is important to note that, while the technology is
improving and will continue to improve to make sure
that parents have a tool that they can use in their
homes to make sure it is age-specific and to ensure
their children see only content that is appropriate for
them, technology is not a replacement for good
parenting.
I know the industry is working with Government and
charities to ensure that parents have adequate
information and that they can feel confident about
how the internet works and also confident in using the
technologies, but also to have that dialogue with their
children, which is very important. These are some of
the points that came out of various Government
reviews such as the Byron Review and indeed the
Bailey Review more recently, where it talked about a
partnership approach to child safety. Industry has an
element of responsibility to provide those tools, but
equally there is the responsibility for parents and the
responsibility for Government also to help with that
education process.
Dido talked about how we have had many decades of
education to deal with how you drive, use seatbelts,
drive safely and do not drink and drive. Equally, there
have been campaigns for years about learning to
swim, children playing safety with matches, how to
climb ladders safely and even recently how to
barbecue your sausages. I think it is important to note
that this education, which some of your questions
earlier referred to, is how you get this change in how
we understand the technology and use it safely. The
answer is it will take time and industry is certainly
stepping up to the mark and providing those tools and
some of the education. We feel it is also incumbent
on parents, teachers and the Government to also assist
with that education role.
Dido Harding: What we have learnt is that parents
value being prompted to think about this. Since we
adopted active choice, if you go on to your account
page with TalkTalk and you have not turned
HomeSafe on, it will pop up and prompt you to say
you haven’t and, if you have children in your house
at any time, we strongly recommend you do. We find
that the vast majority of our customers welcome that
interruption and the vast majority of them with
children at home turn it on as a result. I do think this
is a space where proactive action from all the
participants is required. It is not just going to happen
by osmosis.
Q102 Steve Rotheram: But any parent of a certain
age will know that if you have a technological
problem then the first person you go to is the 12-yearold or 13-year-old. How easily can children
circumvent these filters?
Dido Harding: Nothing in life is perfect. We lock the
doors of our houses, even though we know a burglar
could get in. I would not want to pretend that our
safeguards are 100% perfect, but if you change the
HomeSafe settings the account holder gets an email
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 34 Culture, Media and Sport Committee: Evidence
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
immediately saying that you have done this. The only
way, in our system, that your 12-year-old could get
control of the filtering is if your 12-year-old is also
paying the bill for your home phone and broadband.
In order to do that they will have to set up a direct
debit, which they will not be able to do at at age 12.
Clearly, some very clever 12-year-olds may well have
masqueraded as 25-year-olds and be paying the home
phone bill, but all our analysis suggests that is very
tiny percentage. That closed loop where the
accountholder is automatically informed if any of the
settings are changed and you can only set them if you
are in the account space where you pay your bill is
our closed loop.
Hamish Macleod: In mobile, you would go through
an age verification process before you can take the
filters off. On the education point, this something we
have been again plugging away at for years and years
and years, both individual operators and collectively;
distributing materials within the schools to support
teachers. It is not just about inappropriate content, of
course. It is about inappropriate behaviour. It is about
making sure you have physical security of your
device—do you have a PIN number on your mobile
to prevent unauthorised use, to prevent from people
going on to it and spoofing you—and protecting your
personal data and all that sort of thing.
For years we have been putting this stuff into the
schools. I am not fully convinced that the Government
is giving it enough support within the school system,
because that is the easiest point to access children in
a formal setting. It should not just be about the ICT
doing it. It is about all teachers. Through the teacher
training programme, teaching them how to use
technology in the teaching environment, but also how
to use it responsibly and safely and encouraging
children to behave responsibly because it is their
behaviour that is so critical.
Q103 Mr Leech: Can I just ask why TalkTalk
decided to become the first ISP to have this
unavoidable choice?
Dido Harding: We decided to build HomeSafe about
five years ago before I became the chief executive. It
was my chairman and my chief technology officer’s
brainwave that we could build a whole-home networkbased solution. We have a much younger, simpler
telecoms network—techno-babble for a second, an
all-IP network—that makes it technically simpler and
easier for us to do it and our customers told us that
they were really worried about it.
Unashamedly, we launched HomeSafe because we
thought it was brilliant for us as a business, that it
made good business sense to offer our customers
something they wanted, and to bundle it in for free
with our product. I can’t take any of the credit for
that. It was my chairman, before I even joined. We
stand by it. It cost us of the order of £25 million to
build and a lot of people thought it was not possible
to do. There are not many things, to be honest, where
in TalkTalk I am saying I have done something that
my competitors are about to do. We think it stood us
in very good stead, that our customers who use it are
much more loyal to TalkTalk as a result. It has been a
good investment.
Q104 Mr Leech: What stopped other providers from
doing it before now?
Dido Harding: I am probably not the right person to
answer that question.
Nicholas Lansman: I do not want to speak on behalf
of the other three main providers.
Mr Leech: You can have a guess.
Nicholas Lansman: The point is now that this code
of practice has been agreed by four of the main
consumer-focused ISPs. They all agree that active
choice or this unavoidable choice—so when you have
a new customer, they ring up or contact you on the
internet, they have to make that decision about
whether they want to turn on parental controls—is
now agreed across the board. It is good that TalkTalk
were the pioneers, in that sense, but I know the other
players in the industry are stepping up to the mark.
We have introduced a whole range of child protection
measures. We have talked about choice. We have all
talked about the filtering for free, right across the
board, through the whole home. These are just some
of the developments that have been taking place over
the last year and will continue into 2014. There has
been a huge investment and some major developments
of which the whole industry is quite proud.
Dido Harding: I would say, on a broader topic, what
you have in the UK telecoms market, both fixed line
and mobile, is two very highly competitive markets.
The BBC ran something yesterday on, “Why is it that
in the US you have to pay so much more for your
broadband?” and the basic answer is because there is
no competition. What you have in the UK, in both
fixed and mobile, is two very competitive markets.
Fundamentally, we did this in TalkTalk because it was
something that we were willing to take a risk on that
our much bigger, maybe slightly more bureaucratic—
you never know—competitors were focused on other
things. As you as a Committee are looking at this
industry as a whole, recognising the power of
competition to make sure that you have technological
innovation is important in this sector.
Q105 Mr Leech: You said it cost you £25 million
and, from what you said, I guess that it would have
cost more if you did not have such a simple system.
Dido Harding: Yes.
Mr Leech: So, is it cost that has put off the other
providers before now?
Nicholas Lansman: I do not think it is. I do not know
what the costs have been for the other systems put in
place by the other main consumer ISPs, but it is a
very competitive market. All ISPs will have different
priorities in terms of where they are in the
marketplace. Obviously providing good, fast
broadband is an absolute priority to all the companies
and providing it at a price that people can afford.
There are still 7 million people that are not yet
connected to the internet in the UK; not all of those
have children. There is a range of issues that
companies focus on to deliver fast broadband at a
price that is affordable for the UK market.
Q106 Mr Leech: One of the issues of debate at the
moment is whether the active choice should be about
whether you opt in to adult content or opt in to filters.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 35
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
From TalkTalk’s perspective, do you have a difference
in take-up of filters from people who buy broadband
face to face or buy broadband over the telephone, as
opposed to over the internet where it is all far more
anonymous and people perhaps are less embarrassed
to take a particular view on whether they should
have filters?
Dido Harding: 99% of our customers are buying over
the telephone or online, rather than physically face to
face. We do not see any difference in the take-up of
HomeSafe based on whether it is an online sale or a
telephone sale. As a general rule, in all things we sell,
we see much higher customer satisfaction when they
buy it online, when they are in control of their own
settings and it is easier for customers to configure
online than it is to have an agent. As we have just
been grappling with in the last hour, the technology is
quite complicated. Describing it to somebody who is
not hugely digitally savvy over the telephone is not
easy. I always worry that people buying the product
over the phone might not understand what filters they
have turned on as clearly as someone buying it online.
Q107 Mr Leech: Do you find, once people have
opted in for filters or not opted in for filters, they often
change their mind and change it at a later stage?
Dido Harding: Yes. They turn them on and they turn
them off again. I remember keenly when my husband
was at DCMS himself discovering that the Racing
Post was default blocked in our house because I had
blocked all of their HomeSafe categories and, as
racing was one of his areas, he was a bit cross.
Immediately you will want to change that.
Mr Leech: It is one of your areas as well, though,
isn’t it?
Dido Harding: Yes. You see, that is what our
customers tell us. They do want to be able to change
the filters because their children change, their lives
change and, to keep the tools simple enough—it is
still quite blunt. As I say, you blocked alcohol because
you do not want your 12-year-olds going on to alcohol
abuse sites but then discover that wine retailers are
blocked, so you want to change that.
Q108 Mr Leech: You have recognised that the filters
are getting better?
Dido Harding: Yes.
Mr Leech: Have you seen a decrease in the number
of complaints or queries about a site being blocked or
not being blocked, in the time that you have been
offering the service?
Dido Harding: As the base has grown and our
customers are getting better at understanding it, we
are seeing an increasing desire for us to make the
product better. Our customers would love us to have
the ability for them to white-list a particular site
within a category. For example, the national lottery
would count as a gaming site. There are an awful lot
of families who are buying a lottery ticket each week.
They want to check the winning numbers, but they
want gaming sites and gambling sites to be blocked.
At the moment, our HomeSafe tool does not enable
you to do that. We have seen an increase and I think
it is a good thing that our customers are engaging in
the product, because the thing you do not want is that
they set the filters and think that the internet is safe
now for their family and they do not need to worry
about it.
Q109 Mr Leech: How far are you away from being
able to white-list individual sites?
Dido Harding: Not very far.
Mr Leech: How far is not very far? Six months, six
years, 60 years?
Dido Harding: With some of my competitors in the
room, I would rather not give you a specific timing,
but we are a pretty entrepreneurial business that tries
to do things quite fast.
Mr Leech: I am guessing it is quite technologically
difficult, then?
Dido Harding: No. I am not an engineer, so I bat
away all the hard work, which is very unkind of me,
but a lot of this is more around the amount of changes
you can make at any one time to a stack of code. It is
not that it is technically difficult to block a specific
site and give a customer permission. It is just hard to
get all the work done that everyone would like to do
in it.
Q110 Mr Leech: Turning to social media, how easy
is it to trace people who set up fake identities on social
media sites?
Nicholas Lansman: I know you are taking evidence
from some of the social media companies later, so I
do not want to go too much into this area. I mentioned
before that I think there is a difference between
anonymity online and that does not mean you cannot
trace people. In the previous session with Stella
Creasy we had good examples of where people try to
hide their identity online, yet they have been in the
process of prosecution. People can attempt to hide
themselves online, but there are technical ways in
which they can be discovered. We talked about IP
addresses and so forth.
Mr Leech: Is it straightforward, though,
technologically?
Nicholas Lansman: I do not think it is
straightforward. I am not technologically competent
enough to give you a direct answer to that, but we can
go away and perhaps provide some further evidence
to the Committee on this point. A lot of this is
possible, but some of the issues that were considered
by Stella Creasy about people creating multiple fake
identities is an issue the social media companies are
looking at, at the moment.
Q111 Mr Leech: Do any of you have a view on
whether social media companies could do more to
combat illegal behaviour?
Dido Harding: I would take the view that all of us
participating in this space can do more. I would not
pretend to know enough about what social companies
could do specifically, but I think you should be
pushing us all.
Mr Leech: What, in your view, is the reason why
they are not doing more at the moment? Is it because
of this view about freedom of the internet or is it
about money?
Dido Harding: I thought that, in the earlier session,
Stella Creasy was right. She was talking specifically
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 36 Culture, Media and Sport Committee: Evidence
29 October 2013 Nicholas Lansman, Dido Harding and Hamish Macleod
about Twitter. These social media companies are just
very young and they have grown exponentially fast
and they are catching up with some of these issues. I
do not think you should just label it, “Is it money or
is it technical ability?” A chunk of it is immaturity of
an industry, that we are all learning how to do this.
That is why you should be holding us all to account
to think it through.
Q112 Mr Leech: When you say, “Immaturity of an
industry” though, is it something to do with the
uniqueness of social media companies that seem to be
able to grow exponentially with a very small number
of staff and just literally do not have the capacity to
do the work that needs to be done without a vast—
Dido Harding: If you take my road safety analogy, it
is the sheer speed at which these organisations have
gone from being a garage business with a group of
people in their early 20s to being a Fortune 100
company. The speed of that growth does put pressure
on any organisation to catch up, in a world where the
moral and social framework does not exist yet.
Nicholas Lansman: Added to that also is that many
of these social network companies are global and the
Committee is also looking at this concept of how
companies can know how to comply—well, they need
to—with the law and to interpret it. It is not so much
that the companies are not willing. It is a complex
situation where, even where laws apply online as well
as offline, it is how to get to grips with them and how
they can work with law enforcement as well. As Stella
Creasy said previously, it is an evolving area and I
think technology companies, whether social network
or otherwise, are part of this learning process.
Hamish Macleod: The volume is high. A number of
years ago, I appeared in front of this Committee and
I remember quoting a statistic that something like 10
hours of YouTube video goes up every minute. I heard
that statistic the other day and I believe it is now 10
times that going up every minute. It is an enormous
volume that is being dealt with.
Dido Harding: In all the work that I do with the social
media companies, I see no lack of willingness to
tackle these issues. It is a scale and scope challenge
for all of us.
Chair: I think that is all we have. Thank you very
much.
Examination of Witness
Witness: Mr Jim Gamble, Independent Chair, City and Hackney Safeguarding Children Board, gave evidence.
Q113 Chair: For our third sessions, I welcome back
before the Committee Jim Gamble, who is now
Independent Chairman of the City and Hackney
Safeguarding Children Board, but whom we also
remember as the first chief executive of CEOP. You
will have heard some of the evidence that we have
taken, both this morning and in our previous session.
Can you give us your view of the current degree of
the problem both of the extreme child sex abuse-type
images and also on the problem of children having
access to inappropriate material that is more
appropriate for adults?
Jim Gamble: I will deal with the inappropriate
material first because it is the easy one. I think
inappropriate material is a parental decision for those
individuals who have duty of care of the young people
to make and I think active choice is absolutely right.
If parents and others are prompted to make a decision,
I do not think you can do more than that. You are not
going to go out into their homes and look after their
children for them. We are not going ensure that they
do not drink more alcohol than they should or
whatever else. I think active choice means that you
are prompting a decision that a parent must take.
Apart from that, the reason that it is inappropriate and
not illegal is because it is not against the law and
thereby it is quite appropriate for the parent or carer
or person with the duty of care to make that decision.
There is far too great a focus and emphasis on that
side of the debate at the minute and it is clouding
other issues. If you look at Miley Cyrus or if you look
at some of the other pop stars and their behaviour,
that has a far greater and much more easily accessibly
influence on young people today than seeking out
adult or hard-core pornography for that matter. There
is no debate or discussion of any value on that because
you have young pop stars who young people want to
emulate behaving in a highly sexualised way, whereby
there is no filter. That is that issue.
On the second one, I think we have stood still for a
number of years because of inertia in Government.
Let me say before I start that I am going to speak on
the basis of the Laming principles, following his
review after Baby P, and that is that we must have lots
of professional curiosity and appropriate and
respectful challenge. Anything I say is not to offend
anyone. It is simply to bear in mind those principles.
You have too many images still on the internet. We
have too few people deterred from going on and
looking at those images. We have diverted the debate
from how we stop people from accessing, abusing and
capturing images of our children to how we manage
an internet industry that has not stood still; who have
blocked images for as long as I have been involved in
this, from 2002, 2001; who have put in place
processes, perfect or imperfect, to deal with those.
What we need to do now is address the issue that we
are not being creative or imaginative enough in how
we turn the tables. Let me give you an example of
that.
Peter Davies, in one of his statements to the media,
talked about 50,000 to 60,000 people downloading
indecent abuse images at any given time. What we
know about the propensity of those individuals to go
on and commit hands-on offences is sufficient that, if
you are a parent, you should be extremely concerned.
Ultimately, though, we are not having the debate
about how we resource child protection teams and
how we identify, locate and rescue the children
trapped in these images. I have read the previous
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 37
29 October 2013 Mr Jim Gamble
evidence. I do not accept, and I have rung around, that
very few of the images are new. Too many of the
images are new. The images that you talk about that
IWF will access are not hidden on peer-to-peer. That
is where paedophiles nest to share their images. How
do we identify, locate and rescue those children and
how do we identify, locate and hold to account those
predators that represent a threat?
You talked about education earlier. Unless you have a
deterrent—in other words, we arrest 2,500 people and
prosecute them for this—you have no deterrent to
point to that is credible in the classroom. Young
people understand that. There is a case in America
at the minute where two girls have been arrested for
bullying. It will be an interesting case to look at later.
But let us accept that we have a problem with finance.
What I would propose is the Government would be
better investing £1.4 million in recruiting 12 detective
sergeants, 12 trainers and co-ordinators and using the
enthusiasm that is on the ground, because we have
seen organisations like Let’s Go Hunting. Is that a
good thing for them to do? No. Does it represent the
manifestation of the frustration that we talk about one
issue while our children continue to be abused? Yes,
it does.
Let us for one second be a little bit imaginative and
take the special constables’ role. Why can we not have
special constables online? The figures I have given
you are ready-reckoner figures, so they take account
of not just the salary but also the on-costs of the
equipment. If you had that for each region, every
police force could recruit up to 10 special constables,
just in theory, in their area that would be trained,
vetted and supported to come in and give of their time.
As a retired police officer, I might say, “Well I’ll
actually go into Belfast to the PSNI, and I will give
X amount of nights a week”, the same way as specials
do anywhere. If I had worked in IT or currently do
that, I might want to give up time.
If you did that, you would have 520 people online at
any given time who could manifestly cover 10, 20, 30
or 40 chat-room environments, where you would have
people out there protecting our children. Not rogue
vigilantes, but properly vetted, properly trained and
accredited and properly supervised officers online. To
me, that is what we should be thinking about so we
attack the root cause, which is people. Not get
involved in what I see in the one instance as trying to
cover up the fact that—and I have read your evidence,
that everything is fine in CEOP. Well, I would suggest
you scratch the surface on that if you think that CEOP
has not been diminished by going into the National
Crime Agency.
To move the argument to, “How can you deal with
this problem?” We have all become seduced by the
technology. So we talk about filters and we talk about
blocking. We have been doing that for a long time. I
sat and listened to friends from the internet industry.
They have been doing a good job for a long time. I
have not always been their greatest fan because they
should be pushed and they should be challenged, but
I have not heard any talk about the Click CEOP button
that children can go to in Facebook if their parents
have downloaded it, and report, and we fought long
and hard to get that put in there.
In 2004, when I was deputy director general of the
National Crime Squad, we created a site that was
bogus. When you came to it looking for child abuse
images, it popped up and said, “This is a law
enforcement site. Your IP address is—. You have
committed a criminal offence.” Those issues were
okay in their day, but it seems to me we have stood
still. Matt Bishop, vice president from Microsoft who
sat on my board in CEOP, said, “The thing you must
remember”, and I would challenge you to remember
as you go through this, “is that in the world of
technology, standing still is falling rapidly behind”.
I believe we have fallen behind. I believe the early
success of CEOP—not because of me but because of
people like Helen Penn and the education team, who
is now working for some other unit in SOCO, not
dealing with children, a fabulous child protection
expert, developed all the early CEOP training and
education materials, had worked for Childnet charity,
but is now working in some other part of an organised
crime entity. It does not make sense. We need to
reflect that this is about people. The internet industry
can be attacked for so long, but we have to get to
the point where we say, “How can law enforcement
do more?”
Chair: Thank you. We are going to come on to the
CEOP incorporation issue but before we do, I will
bring in Steve Rotheram.
Q114 Steve Rotheram: Before we do that, I
followed the line of questioning last week about the
victims in this. It is surprising that Jim has come up
with a very different explanation of that from Peter
last week, which is very concerning because one of
the priorities for me was what we do about trying to
prevent other people from becoming victims and what
we do to help those people who are the victims. How
far do we pursue that? Perhaps my line of questioning
did not go far enough, but you are saying that we do
not do enough in that respect?
Jim Gamble: We do not do anywhere near enough.
CEOP now have—I will estimate—two people
working on victim identification. Police forces are
dealing with 101 other issues out there. How many of
them do you think have dedicated resources right now
looking at these issues? I have read the statistics, the
26 million. That is five forces. You are looking at
millions upon millions.
In one investigation we did, Operation Cathedral,
there were 750,000 images, but they related to 1,232
or 1,233 individual children. On peer-to-peer, these
are being created. On the 50,000 Peter talks about,
those are peer-to-peer sites. Those are hard-core
paedophiles who do not stumble across anything on
Google. They nest in these places on the internet
where they can share secretly. You have to infiltrate
that. You have to infiltrate it. You have to identify who
they are and follow them offline and arrest them.
During Operation Ore there was massive controversy,
but you know what? In the UK, we identified, located
and arrested more than anyone did anywhere else in
the world. I am now more convinced than ever that
what we did was right because, for a short period,
people were afraid to go online and download these
images because they could be arrested and exposed to
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 38 Culture, Media and Sport Committee: Evidence
29 October 2013 Mr Jim Gamble
the full glare of daylight, and they do not like that.
You do not have an active deterrent. If you look at the
last two years’ CEOP reports, the arrests are lower
than they were for any year except our first year of
operation. That cannot be right. That is not a deterrent.
Q115 Chair: Just on that point, we came to visit
CEOP when you were running it and you ran a
programme over 24 hours, peer-to-peer file sharing,
and I think you came up with something like 20,000
sites making available hard-core images. Those are
20,000 IP addresses that you could identify and you
could go and raid them or arrest them, if you chose to
do so. Is it simply a lack of commitment or resources
that is preventing that?
Jim Gamble: It is not lack of commitment anywhere.
Peter Davies and everyone in CEOP are absolutely
committed to doing the right work, as are people in
all the police forces around the country. I think we
become seduced by all this and, because of the
volume, we say we can’t do anything. To use the CEO
of TalkTalk’s example, if we did not enforce speed
limits on the roads we would have hundreds of
thousands of people committing those offences, but
when they see a picture of a speed camera on the side
of the road they slow down because they realise that
the deterrent is real and they may in fact collect points
and lose their licence. We need to get to a position
where we stop dealing with some of the technological
issues that have seduced us for so long and attack the
issues, which are people.
To come back to your point about victims, we need to
get into the images to identify those children who are
new victims, those children based in a particular
geography, so we can share that with law enforcement
abroad or do more. Two people working all the hours
that God sends are not going to make that difference.
I am recognising the financial restrictions and
everything these days hides behind austerity, but for
about £1.4 million, giving the public, retired officers
and others the opportunity to engage you could make
a huge step-change and difference. That is about being
child focused and not being organisationally or
politically focused, in my opinion.
Q116 Steve Rotheram: Moving on to the other
aspects of it, we know that possession and publication
of child sex images is illegal. Are there any changes
in legislation that you think could help the police do
their job better?
Jim Gamble: You can always tweak legislation. I
heard your earlier comments on that and I would not
disagree that you might want to refresh the language.
My issue is that we do not need new legislation. We
need new thinking and we need people to apply the
law as it currently exists.
When I listened to the Member of Parliament giving
evidence earlier, Stella Creasy, I was struck by how
well she encapsulated it all. The Prevention of
Harassment Act is bullying. The Prevention of
Harassment Act is trolling. But we do not have
examples that we can take into the classrooms and say
to kids, “Don’t bully a friend to death. Don’t join in
this bullying because of these two 15 and 16-year-olds
who have been prosecuted.” When I say that at public
events and venues people, especially from the third
sector, will say to me, “That’s not right. You cannot
criminalise children.”
If a 15-year-old walked up to the bus stop and
punched you in the face, we would not say, “Well,
you’re only 15 or you’re only 16. That’s okay.” There
would be consequences. The internet is just like a
public place. We need to ensure that the laws as they
exist, when they can be applied, are applied. In my
experience you can track back. The thing is the child
who uses their phone or their laptop or if they are old
enough to have one of those old stationary computers
that three people would lift onto a desk in a bedroom,
they are leaving an evidence trail that is ultimate very,
very easy, to follow once you pull the first strand.
Educating children in the classroom is key. Where you
have two people at the minute—two people, I am
told—in CEOP who are currently qualified to deliver
the ambassador training, because focus is now
criminal justice and NCA rather than education and
social care, how do you get into the classroom? None
of the resources that are going out at the minute
include workshops on Snapchat, on Vimeo, on Vine,
on Ask.fm.
While you might have two or three lines on the
internet, getting into the classroom where you engage
in a virtual cycle: the teacher, the child, and critically,
their parent. Parents will complain and bang the table
for the three days after something terrible has
happened and then they go back to their normal lives
because they do not want to think of it happening to
their children. The psychology of it is you think,
“Well, that is terrible”. That is why I think we have
had all of this from those poor children that were
murdered by Stuart Hazell and—I am trying to
remember the other guy’s name, but after Tia Sharp
and April Jones’ murders.
What we have is this reaction, this overwhelming
urge, because we are all so emotionally engaged with
our children, to demand that something is done; but
rather than demand that the right thing is done and we
deal with people, we begin to deal with technology.
That is a huge problem about focus. We have to look
at how we get to the root cause, identify, locate and
hold to account as many of those people as you can,
which will deter the vast majority from coming online.
They do not want to be caught. They can walk into a
police station if they want to be caught. The only time
you ever hear a paedophile saying, “This was a cry
for help”, is after they have been caught.
We need to see more being arrested and you are not
going to do that with splash screens. If you are a hardcore paedophile and you search for an image and a
splash screen says, “You can get in trouble”, I do not
know—do you think these people grew up in a society
where they did not realise the stigma involved? Do
you think they kept the secret about the way they
engaged and abused children all those years because
they thought it would be somehow acceptable, that
they need a prompt to remind them that it isn’t? That
just does not make sense and it is a diversion of
attention and resource that does not work. We tried it.
Q117 Steve Rotheram: Again, I am for splash
screens, but not for the reason that you are say; not
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 39
29 October 2013 Mr Jim Gamble
for the hard-core paedophiles out there. They are far
more sophisticated than that anyway. That would not
make a blind bit of difference, but as a warning to
others perhaps and to try to change the culture online,
because I think that is very different to what happens
offline.
We all know these sorts of images are illegal
throughout the world. For me, I wonder why there is
such a prevalence of them if everyone knows they are
illegal. Is there a lack of resource from the police to
try to tackle it or, as I asked Peter last week, is there
a problem with international co-operation so that we
need to clamp down on those people who are the main
purveyors of these sorts of images?
Jim Gamble: In 2003–2004 we created the Virtual
Global Taskforce, which is an alliance across
international territory: the Royal Canadian Mounted
Police, the Australian Federal Police, Department of
Homeland Security, Interpol and others. I was its first
chair. In one operation, over 357 hours, I think it was
132 children that we rescued. You need to be online.
You need to be patrolling 24/7. I come back to
imaginative cost-effective solutions. If a police officer
in the UK is working an 8-hour shift, they hand the
baton to an officer in Canada, who hands the baton to
an officer in Australia, you have 24-hour coverage for
an 8-hour shift. We need to go back to those basics,
because it seems to me we have stood still. The
deterrent that existed during Operation Ore, the
deterrent that existed during Operation Chandler—to
put the splash screens in context, I would not be
against them if the splash screen was saying, “122
people have been arrested this year. Do you want to
be 123?” While I was in CEOP, we took an ad out at
Christmas in The Times saying, “Go online and search
for indecent images of children and we will present
you with a criminal record”. I am not against that per
se, but it becomes laughable if you do not have the
publically visible deterrents of people being held to
account in court.
Q118 Steve Rotheram: Is the real problem that if
there were 50,000 people who were accessing these
images, which are all illegal, and those people could
be charged and, if they go through the judicial
process, could be fined or imprisoned—hopefully
imprisoned—that we just do not have the
infrastructure anyway?
Jim Gamble: Here is the question. Can they all be
identified, located and held to account? How do you
know that Bridger was not online looking for those?
The issue for me is always this. In my role in
safeguarding now, in the work that I do as a consultant
and elsewhere, you see people popping up all the time
as teachers or in other roles who were downloading
these images six or seven years ago. The courts have
to take it more seriously. You have a situation now
where you will still see teachers, for example, getting
prosecuted for this and not going to prison.
At one stage, before my clash with the Home
Secretary, in an interview I did say that I recognise
that there was a role for cautions and managing some
people, but that is about the risk assessment of them.
There was a hue and cry that I should have resigned
then and many years earlier. I recognise that there is
a role to be played around that management, but the
more I have watched the situation develop the more I
believe we need to see substantial numbers of people
held to account. Courts need to apply meaningful
sentences. Courts need to avoid making statements
that diminish or trivialise the issue because it is an
image.
They need to be asking questions about whether these
images have been interrogated, whether we have
exhausted every opportunity in the images you have
to identify and locate the child, because I will bet if
you go out to the police forces who have millions of
images they have not all been viewed. If that was one
of your children, would you be more interested in that
being viewed so that they could be rescued from the
abuse that they are probably not telling you about or
would you rather that we had a splash screen? For me
it has to be children first. That is why this debate is
so important, because we have become seduced by
statistics and technology and they become images.
I have heard people like Claire Perry talk about the
fact that these are all crime scenes. Well, yes, great.
We coined that phrase in 2004–2005, but the fact of
the matter is we are still having the same
conversation. It is getting into that. Let us find the
kids. If you find the kids, you will find the people who
did it. If you hold those people to account, they will
not be doing it any more. We have seen in the Savile
case that there is never just one victim. There are
always multiple victims with these people. If you are
able to do that, you are going to create a more active
deterrence because the only way to manage
paedophiles is through fear, in my experience: fear of
capture, fear of exposure.
Q119 Mr Sutcliffe: At the risk of trying to reopen
old wounds, Jim, I asked Peter Davies last week if it
was a good thing that CEOP had been incorporated
into the National crime Agency. He said he thought it
was absolutely a good thing. His argument was that
CEOP had evolved and that the National Crime
Agency brought in wider resources. Clearly, I do not
think you see that as the case. Given the solution in
terms of how you see a potential solution in terms of
where we are, what are your concerns and worries and
do you still have those worries now about what has
happened to CEOP?
Jim Gamble: I came in from Victoria this morning. I
took a taxi across to the hotel where I am staying and
we drove past CEOP. As we drove past, there is a big
wall where the CEOP sign used to be and we used to
have all the photographs of kids who came in from
schools. There is now a big sign on the wall that says
“National Crime Agency”. The Home Secretary, of
course, said that the Government would invest in
CEOP and build on its success and that it would not
dilute or diminish its operational independence or
brand. I have done a lot of work in the last couple of
years as a consultant around brand, around social
media and everything else. Putting up a National
Crime Agency sign is not about not diluting the brand.
That is about the bigger brand suppressing the smaller
brand because it needs now to establish itself, which
is a good example of why it is not right. Child
protection in every other sphere has to fight its corner.
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 40 Culture, Media and Sport Committee: Evidence
29 October 2013 Mr Jim Gamble
The idea in CEOP under The Way Forward paper
presented by the previous Government was to build on
its early success as a consolidated centre, not around
criminal justice, but around child protection. I do not
disagree that CEOP could have been consolidated
somewhere else, but it would have been much better
consolidated as an agency of the Department for
Education than in the National Crime Agency. If you
want to consolidate like-for-like resources, then put
counterterrorism in there because, as a former Head
of Counterterrorism in Belfast and Deputy Director
General of the National Crime Squad, I do know a
little bit about organised crime and terrorism. There
are many things that are familiar.
To take a child protection agency and put it into a
national crime agency where the brand has now been
diluted, where they will continue to have protection
under Freedom of Information, which is
fundamentally wrong, how can you have a child
protection entity that is not subject to Freedom of
Information? You have a child protection entity that
is answerable through the NCA to one political
person, the Home Secretary. Where is the credible
reassurance around that? Where are the lessons of
serious case review around professional independence
and challenge? The fact of the matter is we had
planned the three pillars to be education, social care,
and criminal justice. We now have moved to one
pillar, which is criminal justice.
I read the evidence of last week with real interest to
see where the statistics are—let us drill into them and
see where it is—and I could not see any because it
was, “We are not sure about this number, we are not
sure about that number, but this is improving and that
is much better”. Where is the evidence? Two people
qualified to deliver ambassador training in the UK. I
do not know how many hundred thousand teachers
there are in the UK, but it is a lot.
I do know, from feedback I get at the conferences
where I speak and the other work that I do, that there
is a frustration that they are no longer able to get the
resources. Two fantastic people trying to do a difficult
role, but with no resource and no support. Why are
there no Ask.fm workshops coming out of what they
do? Why are there no Snapchat lessons for the
classroom? Why is there not a programme of work
that joins up with schools, children and parents
together so that you capture the digital footprint of the
family and you understand how you then engage, so
that the child learns to block, report, do privacy
settings with the parent?
Those were all the plans, the 24/7 command and
control centre that should have been in CEOP manned
by child protection professionals where you could
phone in if you had a child trafficking, child
exploitation or child abduction issue. Where is that?
Gone, subsumed into something else. CEOP’s brand
was important because it delivered reassurance and it
had a high profile brand. Do you think that brand is
still as high profile today? The places that I am going
to and the people I am speaking to are saying that it
is not. I was at the Police and Crime Commissioners
Conference in West Midlands, several hundred people
dealing with child protection, not three weeks ago.
Bob Jones was there. I asked, “Do you think it would
be easier for a child to approach the NCA or CEOP
as old?” Of course, maybe they just want to make me
happy and I accept that, but I do not think everybody
is that keen on making me happy; not a single person
in the audience, and there were some from the Serious
Organised Crime Agency as it was then.
I will give you one other example. CEOP’s academy,
partnership with Lancashire University, looking at the
psychology of offending, because we need to
understand why these people do what they do, so we
can interdict it earlier. Where is it? It has gone. It is
not happening. For the last three years it has been
suspended. Why? Why do we not have that learning
being distilled from those interviews and shared with
all of the police forces that are engaging and
interviewing these people on the ground? It is about
having a child protection focus and not having an
organisational one that is about facing drugs, facing
all other forms of organised crime.
One of the reasons I resigned was because 32
organisations made submissions about CEOP going
into the National Crime Agency. None of them, as far
as I am aware, supported it from the Association of
Directors for Children’s Services through the NSPCC
and everybody else. Today, with the gift of hindsight,
they may say, “Well, it is not too bad”. Do you know
why? The Government of the day is the Government
of the day and nobody likes to fight with them, but
why have those reports never been released? We tried
to get them several years ago and under Freedom of
Information they could not be released. It is not
plutonium poisoning. This is about child protection
and what experts’ opinions were at the time.
I think we suffered because of a bloody-minded
approach of the present Home Secretary at a time
when Government was new. They did not, as
everyone in child protection knows, do that critical
thing, which is pause, reflect and plan before you
change anything around child protection. I think what
we have seen is an increase in the number of images
available, an increase in confidence of offenders and
a decrease in deterrence.
Q120 Mr Sutcliffe: I am grateful for that and, not
surprisingly, would support all that you have said. The
issue for me when I was Probation Minister was the
multi-agency approach of dealing with offenders who
came out and had to be monitored under licence or
whatever. We seem to have lost that link in terms of
online safety with those perpetrators, who usually
have multiple convictions, who come back out into
society and have to be monitored by the different
agencies. Has that work diminished or is it still there
in terms of the relationships with the other parts of
the criminal justice system?
Jim Gamble: I do not know. All I can say is the plan
was, in the 24/7 command centre that was envisaged
under The Way Forward, that there would be, for
those critical high-risk offenders, a means for lifetime
monitoring them. We were looking at a number of
providers around that, so that you would have a
centre. Whether a family was in Portugal when a child
went missing or somewhere else, they could get expert
advice. Secondly, somewhere where there was a child
protection ethos; that if something was happening in
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 41
29 October 2013 Mr Jim Gamble
a particular Facebook environment or elsewhere,
someone could go and take an immediate look at it.
I am working with the Isle of Man police at the minute
and looking at how they can integrate their digital
strategy both on and offline that allows them, in a
recognised form, to patrol for their citizens online.
Rather than say, “Do not report crime to us”, say,
“You can engage this Twitter patrol and you can ask
it questions and it will give you professional answers
and help you, because it can be in your pocket”;
looking at how you do that. That opportunity has
been missed.
I think Keith Bristow is an outstanding police leader;
I think Peter Davies has had a very difficult job to do.
A lot of people left CEOP. I hear all the time that the
numbers are much bigger now and that is great if they
have been able to do that on a budget that is largely
the same, of £6.4 million in their annual review. If
they have done that by reallocating, fantastic, but
when I was there and I gave evidence to Home Affairs
Select Committee the week I resigned, we took a
headcount of who we had. It is about 120 now. I know
they say then it was 86. I am a bit concerned about
the difference.
I read with interest the phrases used, because I think
everything here is “more resources and number of
people available for CEOP’s work”. What does that
mean: number of resources and people available?
How many people do you think are in CEOP right
now? How many people do you think are working
right now for child protection in the country? If I were
you that is the question I would want to know. I also
understand that you have heard about the
psychological support that staff get. Let us look at
psychological. How many staff are in CEOP today,
working for CEOP, that have had psychological
support? That will give you an idea of the true number
who are working directly for CEOP; those people who
have had that psychological support who are working
in child protection today.
You have to get beneath the rhetoric of “there are
more”. Apparently, there is 145. Apparently, it is 175.
If “access to resources” means that they can use
Serious Organised or National Crime Agency
resources, are they the right resources? In my
experience in education, I came in as a hard-core
police officer. I did not come in thinking, “Well, we
will go down this education road”, but over the sixyear period what I saw was, “Educate and empower
young people. They can protect each other. Educate
and empower those who are in a role to protect them.
Support social care. Support teachers. You will have
a far greater opportunity of target-hardening those
people who are vulnerable while you do it”. I get no
sense that that has been captured.
When we talked about statistics of children rescued,
we defined a child protected as someone who—if you
are a paedophile and you are engaging two children,
we would count those two children because we would
know the name of the social worker, the name of the
paedophile and the law enforcement officer engaged.
In the broader safeguarding, we would not, if you
were a martial arts instructor that held a whole load
of classes, count 130 children because we had arrested
you. We would count those children that you may
potentially have engaged with. You have to look at
statistics and say, “Are we improving? Are we getting
better? Where have all these things gone?”
You can bring someone in from the National Crime
Agency who is a grade six analyst and put them on a
desk to look at child abuse images. Is that what you
want or do you want someone with a child protection
background who specialises in that and who has built
a history in doing it to be looking at that; who will
consider not just the issues about viewing the images
and the trauma and that, but some of the other
evidential issues and some of the safeguarding issues?
I said at the time I would love to be coming back to
say I had been proved wrong. I am not being
malicious or malign when I say that I think I have
been absolutely proved right. The brand has been
diminished
and
diluted.
The
operational
independence, do you get a sense that CEOP is
operationally independent of the National Crime
Agency? It is not even independent of politics and the
Home Secretary. That is not a good place for child
protection to be. If you were a safeguarding board, I
think you would be heading for trouble.
Q121 Tracey Crouch: I apologise for having to
whizz off straight after the question. My mother was
a social worker in Children & Families for many
years. I have to say I thought I was pretty immune
from stories about child abuse, but some of the
statistics that we have been hearing I have found quite
distressing. You talk about the role of linking up the
agencies for child protection. Do you think that our
current flock of social workers are being adequately
trained to deal with this kind of abuse on the internet
and how can we improve, not just awareness among
our enforcement agencies, but within our protection
agencies as well?
Jim Gamble: That is a great question. I am chair in
Hackney. What you have there is a blended staff. We
talk about the frustration and the problems. You have
young staff who grew up with Bebo, then MySpace,
then Facebook, who use Instagram every day. Those
young social workers bring a skillset with them that
the older ones do not have, but what the older ones
have is experience of dealing with families living in
complex situations, facing issues around alcohol
misuse, mental health and all the other safeguarding
issues that are there.
Do I think that the training is good enough? No, I
think it could be improved. Do I think we are in a dire
situation? No, I do not. My eldest daughter is a special
needs teacher. She is 28. She grew up with Bebo. So
in the classroom she understands all those issues. My
son who lives and works in London, when his younger
sister was coming of age he was the one who did not
want her to go onto social networking sites. Maybe
that was because he did not want her to see what he
was doing, but at the end of the day it was also about
protection. The youngest is training to be a social
worker as we speak and so from that point of view
I agree.
What I would say is we need to get this right again.
The NSPCC ran a survey—I think it probably links to
your question—where they said just about 50% of
social workers were not confident that they would be
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Ev 42 Culture, Media and Sport Committee: Evidence
29 October 2013 Mr Jim Gamble
able to identify a child whose abuse had stemmed
from the internet. The problem with that question was
it was put the wrong way around. The question was
about technology first and harm second. If you ask if
a social worker could identify a child who had been
the subject of neglect, abuse or mistreatment, then the
statistics would have been a lot higher.
Now, there are 85,000-odd social workers on the
general social work care consult register. There are
about 45,000 in local authority care and about 24,000
or 25,000 delivering children’s services. That survey
looked at 327 or 329. Some were from the NSPCC
themselves. Is it representative? I am not sure. Social
workers are like the rest of us. I would think you
would be au fait, if I am being stereotypical, with
social media whereas some of the rest of us would
not. Training needs to be updated and improved. It
needs to be in the workplace as well. We should be
delivering that training online. It has to be
contemporary, because if you develop a training
package now and you do not take account of the new
social media sites, the new trends, the new dos and
don’ts, then you lose sight of it.
It is like RIPA. I was chair of the UK Data
Communications Group. The problem with the
legislation is, when you make it too specific, you
cannot apply it generally enough. Training needs to
be about how you communicate, how you create risks.
Q122 Tracey Crouch: It goes back to your original
premise right at the start about professional curiosity.
Do you have any statistics or evidence or even just
any suspicions as to what proportion of children who
are victims of child abuse images that are posted on
the internet are also on the at-risk register in terms of
general protection?
Jim Gamble: No. If we go back to one of the Ofsted
reports—not the last one but before—I think there was
only about 12% of children who were on the Children
At Risk Register who were known to authorities in
any way. Those children who were dealt with, the
other percentage, the adults in the house were known
to create a risk.
One of the other problems is ContactPoint. How do
you identify and locate a child who is subject to a risk
that is being investigated by social care, by policing,
by health or problems with education? ContactPoint
was the mechanism by which we did that. It was
scrapped on 26 August by the present Government
and never replaced. That was technology that you
could turn all this on its head. The former Head of
Intelligence in CEOP had said to me at the time,
“Look, this has made it easier and quicker for us to
identify some of the children because we get partial
information. You can then look for the specifics.” It
was seen as a Big Brother database, which it was not.
Should there be better training? Yes, but I think the
training needs to always concentrate on the abuse. The
mechanism and means, whether it is child sexual
exploitation on the street, whether it is a child being
trafficked, or whether it is a child being groomed will
have two things in common: people first and
technology second.
Q123 Mr Leech: I do not think any of us can be in
any doubt on your views as to the change from CEOP
coming under the National Crime Agency but, given
that it is, are there any changes that you would like to
see that you think would improve the situation that
have not been introduced so far?
Jim Gamble: Yes. I accept the situation as it is. If it
was within my gift, I think there should be far greater
engagement with the Department for Education. The
education teams should be bolstered. CEOP should
have absolute operational independence from the
National Crime Agency, because children’s voices
must be heard and must come first. I am not sure how
you can do that, given the present format. They will
have an advisory board, but an advisory board is like
an ashtray on a motorbike. It does not make a lot of
difference once you start moving if people are not
using it. I think they need to re-engage education
much more. They need to invest in education.
They could still do what I have suggested around
creating special constables who would work
regionally with their local police forces for £1.4
million. Recruit, train, and accredit through CEOP,
attached to the National Crime Agency. None of that
would be stopped, but I think they also need to take a
step back and learn the lessons about brand. Taking
down signs, changing the CEOP website to put under
it “A command of the National Crime Agency”, that
is not the spirit of children first. That is evidence of
the organisational ego of the larger diminishing the
presence of the smaller simply because the larger is
now new and wants to create a profile.
I would put the CEOP media team back in CEOP, not
integrate it into the NCA. If you are going to be
“children first”, then you need to be advocating and
saying to the NCA, “No, you are wrong. We are going
to take on these cases. We are going to do it in this
way and we are going to invest this amount of our
budget in education”. I do not believe that that can
happen.
Q124 Mr Leech: Does the NCA offer anything to
CEOP?
Jim Gamble: Yes, it offers a capability around
technology when it comes to in-depth investigations
into the likes of the onion router, but the NCA should
have been a good and active partner of CEOP. I
believe the NCA as an entity is an excellent idea. As
far back as 2001, when I joined the National Crime
Squad as an assistant chief constable, the director
general said to me after three months, “What would
you change if you could?” I said, “I would drop the
‘s’ and put an ‘a’ on it and make it the National Crime
Agency and consolidate resources around organised
crime within it”. Now, the fact of the matter is it is a
good idea, but putting things like CEOP in it will
force CEOP, as it has done, to submit to its image, not
allow CEOP to thrive as an independent environment.
How many people are looking for missing children
now? That is a CEOP responsibility. How many
people do you think in CEOP are working on missing
children? The Prime Minister was asked on Channel
5, on the news programme, how many people he
thought were searching for images. I know when they
put to him “between two and five” he was shocked.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG02
Source: /MILES/PKU/INPUT/035254/035254_o002_steve_CMSC 29.10.2013.xml
Culture, Media and Sport Committee: Evidence Ev 43
29 October 2013 Mr Jim Gamble
The fact of the matter is, if you want to see the real
difference you need to get into CEOP and you need
to ask some questions because, as far as I am aware,
everything that I have said to you is absolutely true.
If you accept everything I have said to you as true and
you are not deeply concerned, then there is a huge
issue.
National Crime Agency is great. Organised crime,
terrorism, money laundering, fantastic. Child
protection in Hackney, child protection in
Northumbria, child protection in Humberside, that is
about local child protection teams, local safeguarding
boards, local police. It is about how you support them,
about how you create materials that go into their
schools. A national hub that had teachers in it, that
had social workers in it, that had those people that
create those imaginative products that children can
engage with in the classroom so that they learn the
importance of friendship and supporting a friend who
is being bullied online, who are able to work through
real life cases to say, “If that had been one of my
friends, I would have wanted them to report. Who
would you report to?”
You can be scared of your own shadow in this game,
because nobody wants to attack any of the big social
media providers. I will not name the charity, but I
heard someone from a charity who is on the advisory
body to one of the social media organisations that has
just changed its default settings for 13 to 18-year-olds
say, “We will need to learn more about it”. I think,
“Well, if you are learning more about it after the fact,
you are not an advisory body, if your advice was not
being taken before the fact”. What can I say? The
National Crime Entity is a great entity but subsuming
CEOP into it is a recipe for disaster. You can bring
however many people you want in, but while they are
still working there do you honestly think they are
going to come in and say, “No, I do not think it is a
good thing”? Do you think that will be careerenhancing or career-inhibiting?
Chair: It is 1pm. We had better draw this to a halt.
Jim, thank you very much.
Jim Gamble: Thank you.
cobber Pack: U PL: COE1 [SE]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 44 Culture, Media and Sport Committee: Evidence
Tuesday 19 November 2013
Members present:
Mr John Whittingdale (Chair)
Mr Ben Bradshaw
Angie Bray
Conor Burns
Tracey Crouch
Paul Farrelly
Jim Sheridan
Mr Gerry Sutcliffe
________________
Examination of Witnesses
Witnesses: Simon Milner, Policy Director, UK and Ireland, Facebook, and Sinéad McSweeney, Director,
Public Policy, EMEA, Twitter International Company, gave evidence.
Q125 Chair: This is the further session of the
Committee’s inquiry into online safety and we have
three panels this morning and I would like to welcome
our first witnesses, Simon Milner, the Policy Director
of Facebook, and Sinéad McSweeney, the Director of
Public Policy at Twitter.
Q126 Tracey Crouch: Good morning. I will start by
asking if you could outline what it is that you already
do to prevent harmful or illegal material online.
Sinéad McSweeney: In terms of illegal content we
have a zero tolerance for any child sexual exploitation
material on the platform. So when that is reported to
us, we review those accounts. They are suspended and
they are, in addition, reported to the National Center
for Missing and Exploited Children in the US and they
then have channels of communication with the
relevant law enforcement agencies, who can take
action both on the offenders but also ensure that if
there is a child at risk, that they can be identified,
located and rescued as well. We also have
relationships with law enforcement and police in
different countries. So we have good working
relationships with the police here in the UK not only
so that they are aware of our processes and procedures
in order to request information of us and work with
us, but also so that we would participate in some
education and awareness raising with them about how
the platform works and how to contact us when they
need to, whether that is in an emergency or when they
are investigating crime.
Q127 Tracey Crouch: What about harmful content,
before I move to Simon?
Sinéad McSweeney: Yes. Sorry, did you say Simon?
Tracey Crouch: No, no.
Sinéad McSweeney: Okay. In terms of content
generally, clearly, there are rules as to the type of
content that is allowed on the platform. So when
content is reported to us that is in breach of our rules,
we take action on those accounts. In addition, because
there is such a diverse range of discussions and
contact that take place on the platform, we also have
built into the product the ability for users to label the
media that they may be uploading as sensitive, and
the default setting for every user is that they do not
see sensitive media without a warning.
Q128 Tracey Crouch: You do not have a report
abuse button though, do you?
Sinéad McSweeney: We do, we do.
Tracey Crouch: Okay.
Sinéad McSweeney: We have it within each tweet
because Twitter is built around the small bursts of
information that are tweets. But right within each
tweet there is a report button and once the user hits
that, it takes them to the range of possible issues that
they may be reporting, whether that is abusive
behaviour, impersonation or illegal content.
Tracey Crouch: Okay. Simon?
Simon Milner: From our own perspective, we also
have a zero tolerance approach towards both illegal
and harmful material. On the illegal side, as well as
having clear policies about what people can and
cannot use our platform for, we also use technology.
So there was much talk yesterday in terms of the
announcements made at the PM Summit around
technology. We use one of the pieces of technology
that people are talking about, photo DNA, a piece of
technology created by Microsoft, and we use that to
scan every photo uploaded to Facebook. That are
some 350 million photos a day against a database of
known images of child exploitation imagery, and we
will not allow those to be uploaded and will take
action against those accounts, including notifying the
authorities about them. Then we work closely with
CEOP in respect of their access to that data via
Netmec in order to track down the offenders. When it
comes to harmful material, that is pretty much very
hard to spot by technology so that is where our
reporting processes come in, and we too have
extensive reporting processes via both our Help
Centre, which anybody can access—they do not have
to have a Facebook account for that—or from
individual pieces of content. They can report it to us
or indeed report it to somebody else who can help
that user take action so, particularly, in the areas, for
instance, of bullying and harassment, we do not
tolerate that. Those reporting processes are incredibly
useful to someone either who themselves is feeling a
risk of harm or if somebody sees somebody else on
the platform who they think is at risk, then they can
use those reporting processes.
Q129 Tracey Crouch: This week, as I am sure you
are aware, is anti-bullying week and there has been a
great deal of discussion about cyber bullying over the
past few months, especially as a consequence of some
rather tragic events. What are you doing to work with
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 45
19 November 2013 Simon Milner and Sinéad McSweeney
organisations to try to raise awareness about cyber
bullying in particular?
Simon Milner: So perhaps if I take that one first. We
work with a number of anti-bullying organisations.
Indeed, yesterday, I was at an event in London with
450 schoolchildren and their teachers with the Diana
Award. The Diana Award has an anti-bullying work
stream and they have a fantastic champion there in a
young man called Alex Holmes. He does great work
around the UK, working with schoolchildren, enabling
them to become anti-bullying champions so we
support that work. We help to fund what they do and
I was there yesterday to talk mainly to teachers to
help them understand how to deal with these issues on
Facebook, because it is something that we absolutely
recognise they need help with. But we also provide
lots of help via other people who reach out to schools.
So Childnet are another great charity for doing this,
because we recognise bullying typically does not just
happen on one platform and will happen across a
range of different media as well as in the school.
Therefore, working with people who can not only talk
to the schools about Facebook but also about Twitter,
about Snapchat, about Ask.fm and other services that
young people are using is much more effective than
having Facebook necessarily coming to the schools.
Sinéad McSweeney: I think we would have a similar
approach in terms of partnering with the people who
are talking to children all the time in this space, so the
South West Grid for Learning, the UK Safer Internet
Centre, participating not just in the Safer Internet Day
or the anti-bullying weeks but on an ongoing basis,
highlighting the resources that are available both
within the platform and more generally in terms of
breaking down the advice that we have to ensure it is
suitable for each different category of people who
need to know and understand what is happening. So,
for example, in our Safety Centre we would have tips
for teachers, tips for parents, tips for teens,
recognising that the language and the context may
need to be slightly different for each audience. That is
material that we have highlighted again, as I say, to
organisations who are working in this space on a daily
basis, whether it is Parent Zone or BeatBullying or
whatever, so that they can link to our material and
also so that we can link to their material so that there
is a constant shared learning. Also, as Simon says,
recognising that the context for the behaviour that you
may be seeing on one platform may be behaviour and
activity that is occurring in other platforms and also
in the offline world, it also about encouraging how to
deal with online conflict, encouraging empathy and
also trying to educate young people that the
consequences of their online behaviour are the same
as offline. If you say something online, it has an
impact on somebody in the real world although for
young people it is their real world, so sometimes we
have to be careful even in terms of that kind of
language. But it is a question of ensuring that we are
using our own communication channel effectively by
working with people like the Safer Internet Centre on
their Connect with Respect campaign but then also
getting into the other worlds in which young people
are accessing information and ensuring that those who
are guiding them through the world, whether it is their
teachers or parents, are as equipped to deal with the
questions and activities that they are involved in as
the young people themselves.
Q130 Tracey Crouch: You both mentioned that quite
often harmful behaviour is happening across multiple
platforms. Do you think that all social media providers
are equally good at dealing with this or do you think
some are better than others? I have in mind, obviously,
the controversy that came out over the summer about
Ask.fm and I just wondered what your views are on
that. Do you think that that, perhaps, poor or
lacklustre response from Ask.fm had a very negative
impact on the wider social media providers?
Sinéad McSweeney: I think we can only speak for our
own platforms but I know that our experience,
whether it is ourselves and Facebook who are here
today, is that it is an area that we work on together.
We are not territorial or proprietary about the way in
which we work with young people to ensure their
safety online. We all participate together in, for
example, the planning day for next year’s Safer
Internet Centre and there is another planning day
coming up in December. So, from our point of view,
it is about sharing the expertise and ensuring that
together we are giving young people the skills and
resources they need.
Simon Milner: In fact, we work so closely that when
Sinéad and her colleagues recruited our Head of
Safety for Europe, we were genuinely pleased. We
thought this was good for her in terms of her
development but also good that Twitter sees the value
in having somebody who has been working on safety
for four years come and become part of their team. I
think that we have to recognise one of the things about
young people is they love new services and they love
to try them out. Usually, they are the first ones to
adopt new services. It would be unrealistic to expect
that those new services have invested to the extent
that our two companies have in safety. Nonetheless, I
think we have a duty as citizens but also as part of
this industry to help them, so that is what we try to
do. I certainly do that around, for instance, the
UKCCIS Board that I sit on. We are always trying to
share intelligence around what are young people using
and how can we help them to be well-educated in
using those services.
Q131 Tracey Crouch: How many people do you
employ to combat illegal or harmful content?
Simon Milner: We have a user operations team of
several hundred people. It is not a public number. I
am afraid I cannot tell you. Our company as a whole
has 5,500 people. We have several hundred people
who work in user operations, including hundreds of
safety experts, so we have people who have seen
everything before and know how best to deal with
it. We are always also looking to work with outside
organisations to try to ensure that we are consistently
learning about what is the best way of helping our
users.
Sinéad McSweeney: Similarly, we are a much smaller
company obviously. We have 2,000 employees
worldwide. We have trust and safety teams based in
San Francisco and also here in Europe in Dublin, so
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 46 Culture, Media and Sport Committee: Evidence
19 November 2013 Simon Milner and Sinéad McSweeney
those teams are providing 24/7 coverage right across
the globe across the different time zones. The trust
and safety and user services teams are divided into
sub-teams to address the different issues that can arise
across the platform, the different rule breaches. There
is a team looking at abusive behaviour or user safety
behaviour through to impersonation through to illegal
content. So those teams are covering, as I say, the
different time zones on a 24/7 basis.
Q132 Tracey Crouch: Final question, Mr Chairman.
Clearly, although you might like to think so, neither of
your platforms are perfect in dealing with particularly
harmful content. What do you think you can do better
to try to stop or prevent future harmful content,
because I recognise that illegal is very different and
probably being dealt with quite well?
Simon Milner: That is a really tough question. I
would say we never rest on our laurels. We are always
looking to innovate and I think the moment when we
start to think we have this cracked, it would be an
unfortunate moment. It certainly would not fit with
the culture of the organisation, which is to constantly
innovate. So I really would be reluctant to predict
what is the next best thing we could do but I would
be very surprised if I came back here in a year’s time
and said, “We are just doing the same as we did last
year”. We are always looking to innovate.
Sinéad McSweeney: I think like anything in life, there
is always the potential to do something more
effectively, more efficiently and in a better way for
users. I spent 10 years working in police
communications and in that context, of course, you
also in some ways have the division between what
was illegal and things like antisocial behaviour, which
was harmful. So it is about constantly working with
the community, and in our case that is our users, to
know and understand how they want the platform to
be a safer and better place because at the end of the
day, it is there to enrich their lives. It is connecting
people to events and information and news and so on.
Q133 Tracey Crouch: But you both recognise that
you still have weaknesses?
Sinéad McSweeney: It is a question of understanding
that as the technology evolves, just like in the world
of crime or antisocial behaviour, people think of new
ways to breach your rules so you have to ensure that
you are one step ahead. So we have a dedicated
engineering team on the trust and safety side
constantly alive to the possibility that there are issues
within the product that we will need to address when
somebody finds a new way of using the platform. So
I do not think anybody would ever put their hand up
and say, “We are done on safety”, and particularly not
when it involves children.
Q134 Chair: You said you each have lots of people
sitting there who are safety experts. If I push the
report abuse button, how long before somebody looks
at what I am reporting?
Simon Milner: In Facebook, it would depend frankly
on who you are and what you are reporting. So if you
were reporting that somebody you know is suicidal,
we are going to look at that very, very quickly. It
could be within minutes. If you were reporting that
there is some spam that you are seeing, we may not
get to that for a week so it would depend on the
seriousness of the harm that you are reporting. Also,
frankly, how old you are. So teenage reports we are
going to look at more quickly than we do others. But
one of the things we always hesitate is to make those
hard-and-fast rules because, otherwise, we might find
that people think, “I have this problem. I will report
myself as being suicidal because that means I will get
looked at sooner”. So we always aim to get to reports
within 48 hours and most of them we do that. As I
say, for those that are most serious we will get to them
very quickly.
Sinéad McSweeney: Similarly, depending on the
nature of the abuse or harm that is being reported.
Again, threats of self-harm, suicide, threats of
violence and anything under child sexual exploitation
category is looked at within a very short period of
time. Again, other issues may take longer.
Chair: Are we talking minutes again?
Sinéad McSweeney: It can depend. It could be a
minute to an hour. Again, you cannot be hard-and-fast
in terms of the issues but we do respond to those first,
before other issues.
Chair: For instance, there is a choice. If I press the
report abuse button, I am given the choice of
different categories?
Simon Milner: Yes.
Sinéad McSweeney: Yes.
Q135 Chair: I see, and depending which I choose
will determine how long it takes to get back to me?
Simon Milner: It could be that but, also, it may be
dependent on what is going on in certain events. One
of the things about our teams is they are national
teams so they are interested in what is going on in the
news. So, for instance, around the time of the August
riots 2011, the team was on high alert for content
associated with those riots. So antisocial behaviour or
people organising that kind of thing might not have
been flagged otherwise but because we were
particularly on the lookout for that, then we will
deploy more resources. So it will depend on the
circumstances, the public circumstances as well as the
circumstances of the individual who is reporting.
Q136 Angie Bray: In terms of antisocial behaviour,
the debate has also moved on to whether you should
be showing certain footage of things like beheadings,
which I think, probably, just showing that does cross
into what I would consider to be fairly abusive. I
know that Facebook has taken quite a strong line on
that. Has it Facebook’s view changed as this debate
has raged or are you still adamant that this is freedom
of information and people should be allowed to see
these?
Simon Milner: I think the way I would put it is that
we have refined our approach to this kind of content.
I am sure we are grateful every day is that we live
in a country where beheadings are abnormal, where
random acts of violence do not happen to most people
we know, but there are lots of people who are using
our platform in countries where this is a normal part
of life. They want to use our platform to highlight
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 47
19 November 2013 Simon Milner and Sinéad McSweeney
what is going on in their neighbourhood, in their
country, in their city. Sometimes that means posting
graphic content because that is the only way to get the
world’s attention. We recognise there is a place for
graphic content when people are condemning it, so
absolutely not to glorify it and not to take pleasure in
it. When people are using our platform to condemn
this and to bring the world’s attention to it, then we
should be able to do that. However, we also recognise
that given that we are a platform that also has young
people, people under 18, we need to enable people to
get more prior warning and we are taking steps to
ensure that people share responsibly. There are
different ways in which we can do that on the platform
and we are developing those, frankly, in the light of
some recent events. So you are right to highlight it
and I am glad I have had the opportunity to address it,
but it is something where we are refining our approach
because we think there is a place for people to share
that kind of content in the right context and with
appropriate warnings.
Q137 Angie Bray: But do you set the context? It is
one thing to show footage but you are not suggesting,
are you, “We are showing this because it is
demonstrating what terrible things are happening”?
You are simply showing it, so are you setting an
adequate context?
Simon Milner: We are not setting any context.
Angie Bray: That may be the problem.
Simon Milner: No, just to be clear, because we do
not post anything to Facebook, it is our users that post
things to Facebook and it could be a government, they
could be a parliamentarian, they could be a
campaigning organisation and they could be an
individual. An example is after the terrible bombings
in Boston at the Boston Marathon, a number of people
who were witnesses posted footage to say, “Oh my
God, look at this horrendous thing that has happened”.
We enable people to do that and we enable people to
share that with the world, and we think that is part of
how we bring the world’s attention to some horrible
things that are going on. What we do is provide the
tools to enable them to share responsibly and we are
refining those tools as we speak to enable people to
put warnings, and also to share in a way that young
people are much less likely to see it. But just to be
clear, if you go on to Facebook and search for
beheadings, you would never have found them. Our
search facility does not allow you to do that whereas
if you go into many other parts of the internet and
search for things like that, you can very easily find it.
In the recent examples, we had no incidence of reports
to us from young people or others with responsibility
of looking after young people saying, “Oh, my
goodness, this young person has seen this video”,
because it was not being shared in that way.
Q138 Angie Bray: Twitter, of course, has been under
some kind of criticism for the fact that it was quite a
useful vehicle during the recent terrorist outrage in
Kenya where it would seem that some terrorists were
using Twitter to get all sorts of messages out. Again,
does that make you uncomfortable that Twitter can be
used in that way?
Sinéad McSweeney: As a platform, we do not
knowingly allow terrorist organisations to maintain
accounts on Twitter.
Q139 Angie Bray: But were you aware that it was
being used?
Sinéad McSweeney: Where we have reports to us
from law enforcement, where it is brought to our
attention that terrorist organisations are maintaining
accounts on Twitter, we take action on those accounts.
We do not allow accounts to be used for direct threats
of violence or for unlawful purposes. I think the flip
side of the events in Egypt, in Kenya and events right
across areas of conflict is that Twitter is used by
human rights activists and by the media to get
information out to the world about events like those
and ensure that people in real time can follow the
events. But to make it clear, we do not knowingly
allow terrorist organisations to maintain accounts on
Twitter.
Q140 Angie Bray: But it was over a period of two
or three days, not just a flash in the pan, that these
terrorists were able to send out to the world what they
were going to do, what they were not going to do, that
they were coming again, and they were basically using
it to send out their philosophical message. So at what
point would Twitter have woken up to the fact that
they were being used in this way? Really, the world
does not need to know about that, does it?
Sinéad McSweeney: Drawing on my experience from
policing and the way in which police approach the
investigation of certain types of crime, there are
different approaches to the existence of information in
the public domain and the way in which it can assist
investigation. So, as I say, we work closely with law
enforcement. Where we receive reports of terrorist
organisations maintaining reports on Twitter, we take
the action that is required of us.
Q141 Angie Bray: Did you get any reports on that,
because it was certainly all over the newspapers?
Sinéad McSweeney: I am not going to talk about
individual accounts. We do not talk about anybody’s
individual accounts or indeed individual engagements
that we have with law enforcement.
Angie Bray: So was action taken?
Sinéad McSweeney: Action was taken in those
circumstances but we do not comment in detail about
the interactions or communications we have with law
enforcement.
Q142 Mr Sutcliffe: If I can just turn to the antibullying theme, in the evidence that is given to us,
30% of youngsters under 16 expect to be bullied on
social media. What do you say about that? What is
your view of that?
Simon Milner: I think it is sad that young people feel
that. I expect if you ask them also, “Do you expect to
be bullied at school?” probably a similar number
would. Most bullying that we see on Facebook starts
for young people in the classroom, on the school bus,
in the canteen. I have spent a lot of time travelling
around the UK meeting with teachers and that is a
regular thing that they tell me. Indeed, because of that,
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 48 Culture, Media and Sport Committee: Evidence
19 November 2013 Simon Milner and Sinéad McSweeney
we have allowed a feature to enable teachers to
resolve issues on Facebook. They can both report
things to us without having an account but, also, a
young person who is feeling bullied by somebody else
in school can report that person to a teacher. All they
need is an email address and that teacher will get a
report saying, “Simon Milner feels he is being bullied
by—forgive me—Sinéad McSweeney and this is the
content that he feels is bullying”.
Q143 Mr Sutcliffe: But does that happen?
Simon Milner: Roughly half of all the young people
who report bullying use social reporting and that is a
global number. They find it very effective for
resolution. The reason we introduced this was often it
was very difficult for us to understand the context of
a relationship between young people and, typically,
bullies are friends of those who have been bullied.
Therefore, it was very hard to understand why that
person felt a particular piece of content was bullying
because we did not understand the wider context of
their relationship, and that is why we enabled this
feature for somebody who does understand that
context to be able to get involved in it.
Q144 Mr Sutcliffe: Is it just teachers? One of the
other things that concerns me is that children or young
people are on the internet without their parents being
involved. What do you do to help parents identify
potential problems?
Simon Milner: I am a parent of teenagers. I absolutely
understand where you are coming from although, like
fellow parents, one of the things you recognise as
children grow older is they do become more
independent. I do think part of their digital citizenship,
indeed their general citizenship, is about having an
ability to be independent online, particularly around
teenagers. One of the things we do provide is, firstly,
parents can report to us. So if they think there is a
problem on Facebook, they do not have to have a
Facebook account to report to us. They can do that
through the Facebook Help Centre. We provide
information through Parent Zone. You mentioned the
good work that Vicki Shotbolt and her team do earlier.
I have done a video for Parent Zone explaining to
parents about how they can deal with issues if they
have concerns about what is going on on Facebook.
But as with everything, including some of the things
that we have discussed yesterday in the Prime
Minister’s Summit, the key thing is getting young
people to talk to someone. That is where we found the
biggest problem in the terrible events of young people
committing suicide. Typically, they have not told
anybody about the problem and we cannot help
somebody who does not tell us, just the same way as
a teacher cannot help a pupil if they do not tell them.
So that is the main message we give to parents, “For
goodness sake, even if your child will not tell you, get
them to tell someone who can help”.
Q145 Mr Sutcliffe: I think the point I am trying to
get to, Simon, is that because we have social media,
it has not suddenly become a problem. That problem
has always been there, as you say, bullying at school,
that sort of thing. I think that is the issue that I am
trying to get across, that you have mechanisms in
place for people to get to.
Simon Milner: Also, I think one of the things that
distinguishes bullying online from bullying offline is
that there is a footprint. There is a record so this
business of being able to report to a teacher is not
saying, “Sinéad said this”. It is, “Here is what Sinéad
has said to me. Here is the photo she has posted”, and
that is physical evidence that can be used then to help,
not to necessarily punish the bully but to help educate
the bully about, “This is the consequence of what you
are saying and the words that you think may be funny
actually are very hurtful”.
Q146 Mr Bradshaw: Sorry, I am having trouble
finding your report abuse button. Could you help us?
Report abuse, where is it?
Sinéad McSweeney: It is within the tweet. If you are
in a tweet the three dots indicate more content. At the
bottom. It depends on which operating system you
are using.
Mr Bradshaw: Don’t worry. Perhaps you can show
us later.
Sinéad McSweeney: At the bottom, it says, “Report
tweet”. I am happy to talk you through it.
Q147 Mr Bradshaw: It did not seem to me to be in
an obvious place. You said earlier that you felt you
did respond in a timely and effective fashion to reports
of abusive tweets but that was not the experience of
our colleague Stella Creasy and the woman who was
involved in the save a woman’s face on the British
bank note campaign. They are quite high-profile
people so how can we have confidence—they were
very unhappy about how this particular issue was
responded to and the abuse they suffered—that
ordinary people who do not have that level of profile
and voice are having their reports taken seriously and
dealt with quickly enough?
Sinéad McSweeney: I think without getting into too
much detail of individual accounts—I appreciate that
while Ms Creasy has given evidence in public here
that Ms Criado-Perez, while she has spoken publicly,
has not given evidence in this forum—in some ways,
what was experienced in that number of days in
August is not the commonplace normal experience of
a Twitter user. The levels of activity that were directed
at those individuals were unprecedented in my time
with Twitter, so that is the first thing to mention. Since
then we have done a lot of work to highlight the
reporting resources that are there because part of the
issue was accessing the reporting forums. It was not a
case of non-responsiveness to our report. I think the
issue that they were talking about was the ability to
report so the in-tweet reporting, greater ease of access
around forums, better tools for users are things that
we have improved since and continue to improve,
going back to the other member’s question, that that
job is not done.
We have increased the resourcing around our trust and
safety team. Simon already made reference to the
person we have recruited to head up user safety for
Europe and she has started on our team since then.
In addition to that, again, we are working with the
organisations in this space like South West Grid for
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 49
19 November 2013 Simon Milner and Sinéad McSweeney
Learning, UK Safer Internet Centre, so that when
people contact them about issues, they immediately
are able to give them the advice and the resources that
they need. But publicly at the time there were a
number of statements from Twitter about what
happened and the fact that we took on board the
responses that we received. I have spoken personally
on numerous occasions in the time since to the people
involved in that and I continue to take on board the
feedback they have.
Q148 Mr Bradshaw: You mentioned the South West
Grid for Learning as part of the UK Safer Internet
Centre—in fact, both of you have more than once—
and in your evidence you have been full of praise for
this organisation. How important do you think their
role is in helping to mediate and report abusive
comments and getting them taken down?
Simon Milner: I think they are particularly helpful for
schools so there are professional hotlines. They are
there mainly for teachers and they can help people
understand problems because they have seen them
before and, particularly, Laura Higgins, who runs the
helpline there—it is a shame you are coming to the
end of your evidence, she would be a great person to
bring in to talk about it—she is an expert in all the
different services that young people are using and she
has seen every different kind of issue in the school
before. I was with her yesterday at this event with the
young people and their teachers and not one of those
teachers had heard of her helpline, so it was a great
opportunity for her to tell them about it and it is
wonderful for me to have this opportunity to tell
people about her helpline. They have a phone number
that teachers can call and say, “We have this problem.
Can you help us?” Then what she has is a hotline to
us. Often she can help people resolve their problems
without needing to come to Facebook but sometimes
she cannot. She has a hotline to our team in Dublin to
help resolve things where she cannot do it.
Sinéad McSweeney: That is the same approach that
we have, that we ensure that she has all the
information she requires and in the situations where
she cannot in the moment resolve the issue that the
school is bringing to her, she can contact us and we
give her the additional information she requires.
Q149 Mr Bradshaw: I should perhaps declare an
interest, Mr Chairman, because she is a constituent of
mine. Given the high praise that you both have for her
and her organisation, is it not surprising that it is run
on a shoestring funded by the European Union?
Simon Milner: Indeed, I was talking with her about
this yesterday. We have provided some resources to
help her and her colleague David Wright run a road
show around the country, so we helped to support that.
But I think you are right in saying this is something
that we as an industry should look at, as Laura is
looking to develop that helpline, as in, “Can we help
her further, including giving additional resources?” It
is a good point to raise.
Sinéad McSweeney: Similarly, we have assisted them
with training workshops for hotline managers and so
on, so we have been asked and have been able to
contribute to the way in which they provide their
service.
Q150 Mr Bradshaw: But this valuable service, I
think, is being done by one and a half people with
very little money. It is running out next year. You have
deep pockets. Is this not the kind of thing that you
should be funding as a matter of course?
Simon Milner: Laura has not actually asked us for
money to do that yet but I am sure when she does,
possibly as a result of this, we shall be listening with
a very open mind.
Q151 Paul Farrelly: You are both global businesses
so, presumably, you share with each other in your
organisations experiences from around the world.
Simon Milner: Yes.
Paul Farrelly: Presumably, if something happens in
one part of the world, as with any big global
organisation, if you are approached by the press or
committees like this, you will have notes on how to
deal with questions?
Simon Milner: I will not put it as strongly as that. We
tend to talk about what are the issues we think might
come up and, frankly, who is going to answer the
question first.
Q152 Paul Farrelly: Sinéad, if you were asked about
things that are happening in one part of the world,
would you expect to be briefed by an organisation?
Sinéad McSweeney: It depends. My responsibility
goes beyond the UK so I would be familiar with issues
that are going on in the EMEA region, which is
Europe/Middle East basically.
Q153 Paul Farrelly: Okay. I was following the
build-up to the All Blacks match at the weekend and,
as we were doing this inquiry, I was asked whether
we have come across the case that is causing huge
controversy in New Zealand called the Roast Busters.
Have you come across that case?
Simon Milner: No, I was at the recent Family Online
Safety Institute conference in Washington DC with
NetSafe, the New Zealand equivalent of the UK Safer
Internet Centre, and they did not mention it, so I am
not aware of that, I am afraid.
Q154 Paul Farrelly: Have you come across it?
Sinéad McSweeney: I am not familiar with it.
Q155 Paul Farrelly: No. I am just reading from a
CNN report and reports have been all over New
Zealand press over the last two weeks so this is from
a couple of weeks ago. Just a synopsis of this case, it
is about a group of teenage boys who were boasting
online about raping drunk and underage girls, getting
so incoherent they did not know what they were
doing. What came to light, which was exposed by a
Channel 3 television programme in New Zealand, was
that for some months they have had a Facebook site
on which they were posting these boasts. Days after
the exposure of the site, it was shut down and there
was apparently another site that had started on
Facebook. Facebook is not the only social medium to
be used by this gang. Just again from the CNN reports,
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 50 Culture, Media and Sport Committee: Evidence
19 November 2013 Simon Milner and Sinéad McSweeney
Twitter was also a conduit for their boasts, Ask.fm
and also YouTube. I clearly cannot ask you what
lessons you are currently learning from that
experience because you do not know about it.
Simon Milner: Mr Farrelly, I am not aware of that
case. I am happy to ask my colleague, who does my
equivalent role in that part of the world, about it and
see whether there are some lessons we might learn. I
do not really want to comment further. It is clearly not
appropriate because I do not know any more about it.
Q156 Paul Farrelly: Would you both drop us a note
on it?
Simon Milner: Sure.
Q157 Conor Burns: Can I just ask in relation to this
report abuse facility within Twitter, does it alarm you
that you have three relatively sophisticated and regular
tweeters on the Committee and we cannot find it? I
was not aware that it existed. When did you launch it?
Sinéad McSweeney: It was rolled out at the end of
July and the plan was to roll it out across all the
platforms by Christmas but, in fact, we achieved that
by the end of September. If you are saying to me that
we need to highlight it more, that is something that I
can take away. We have done blog posts. We have
tweeted about it. We have spoken to various safety
organisations who work in this space. I think every
time that I have been at a meeting, I have mentioned
it but as I know only too well, having spent 10 years
in Comms, that you can never tell people something
often enough so we will keep telling people about it.
Q158 Conor Burns: I think the message from us is
that with the nature of what we do, we are probably
slightly more prone to abuse sometimes than your
average user and I was not aware that it existed.
Sinéad McSweeney: It is not the only means by which
you can report abuse. You can also report abuse
through the Help Centre.
Conor Burns: I know. I have just googled that. I have
seen that there is another mechanism to do that.
Sinéad McSweeney: Yes.
Q159 Conor Burns: Facebook. When we were in
California, I raised an instance where someone had
posted something that was profoundly libellous
against someone else. It was reported and absolutely
nothing happened. Facebook came back and said they
deemed that it was legitimate comment. Would you
like to comment on that?
Simon Milner: I do not know the details of the
individual case but, as you can see, we looked into it
and we do not always agree with reports that we get.
Frankly, a lot of the reports we get are frivolous. The
classic one is somebody says they like Justin Bieber
and other people report them as being suicidal so I am
not belittling it but we have looked at it.
Q160 Conor Burns: What about somebody saying
that they believe somebody else is a child abuser?
Simon Milner: Again, we will look at the context and
because one of the things we have learnt is just
because somebody has said it, that this is suicidal or
child abuse, does not mean it is. We have to look at
the context and the details.
Q161 Conor Burns: No, I am putting to you that
when somebody posts that they believe somebody else
to be a child abuser and that was not taken off.
Simon Milner: I do not know the circumstances to
the individual case, and perhaps you and I can
correspond on it afterwards, but one of the key things
that is open to people if they think we have made the
wrong decision is they can go to the courts.
Conor Burns: Which is out of reach for a lot of
people.
Simon Milner: That is a matter for Parliament as to
how the defamation process works in the United
Kingdom. I do not think that is a matter for me.
Q162 Conor Burns: But there are often instances
where there are lovely, warm words that come forward
about procedures, hundreds of people on help desk,
but we encounter real people who are finding that
there are no options open to them and you hide
behind, “That is a matter for Parliament”. But you are
the publishers. You are allowing these things to be
published and said about other people.
Simon Milner: We are a platform for free expression.
We have a mechanism for somebody to report to us
something that they think is defamatory or libellous.
We look at those very carefully, we are risk-averse as
a company, so I am sure that the team that have made
that decision have done so with full knowledge that
having made that decision to leave that content up that
things are open to an individual to not only take action
against the person who posted it but also against us.
But on the matter of, “These are warm words”, these
are not just warm words. We see examples week in
week out of particularly young people and people in
distress being helped as a result of reports by their
friends—young people who are reporting that they are
feeling suicidal or indeed that they have taken some
pills or cut themselves. In situations where we see that
somebody who thinks somebody is in imminent risk
of harm, we will alert the local law enforcement who
would go and find that person. Quite typically, they
would get to them before they do the deed or in order
to help them so we help people in the real world
enormously week in week out.
Q163 Conor Burns: Yet, as a publishing platform,
you allow comments to remain up that a newspaper
would not in 100 years dream of publishing.
Simon Milner: I do not think that is fair in the context
of I do not know the details of this case but I am
absolutely certain that a team would have properly
looked at that and made the decision based on the
available evidence in the context. It is open to that
individual to take action if they wish.
Q164 Chair: Can I pursue with you some of the
measures that you have in place? First of all, what is
the minimum age for somebody to subscribe either to
Facebook or Twitter?
Simon Milner: It is 13 for Facebook.
Chair: 13.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 51
19 November 2013 Simon Milner and Sinéad McSweeney
Sinéad McSweeney: Yes, our services are not directed
at under-13s.
Q165 Chair: 13, and how do you verify that
somebody is 13?
Simon Milner: With our platform, when people sign
up for Facebook, there is quite an extensive new user
process and one of the things we require people to do
is provide their date of birth. If somebody puts in a
date of birth that is under 13, we will not allow them
to open an account and we will also use technology
to prevent them from trying to put in a new age from
the same device, trying to change their age, so that is
one mechanism. We recognise that people,
unfortunately, lie about their age. Did you see this
recent report from the NSPCC, which came out on
Friday, about this? They reported a pattern that we
have seen in other studies that in the majority of cases
of 11 and 12-year-olds joining Facebook, their parents
have helped them set up the account so the parent
knows that their child has a Facebook account. One
imagines typically the deal is, “You can have a
Facebook account but you will be my friend on
Facebook so I can look after you”. In those situations,
it is very difficult for Facebook to spot that that person
is an underage user. We still also encourage and
indeed find reports from our users of underage
accounts incredibly helpful and indeed when teachers
ask me about this, I say, “You can report it. You can
let us know that somebody is underage”.
Q166 Chair: Okay and what about on Twitter?
Sinéad McSweeney: We do not collect age
information on sign-up. I think Twitter has established
a reputation in the area of privacy. We minimise the
amount of information that we require from users to
sign up so we do not collect age or gender or other
details about our users. Where it comes to our
attention that somebody under the age of 13 is using
the platform, their accounts are removed.
Q167 Chair: You say you have to be 13 to use
Twitter but you are also saying you do not ask
anybody if they are 13 or over 13?
Sinéad McSweeney: Yes. That is a balancing between
the premium that we put on our users’ privacy and the
issues that you are raising.
Q168 Chair: How does somebody know they have
to be 13?
Sinéad McSweeney: Sorry?
Chair: If I joined Twitter, am I told that, “You have
to be 13”? How do you make people aware?
Sinéad McSweeney: It is in our privacy policy that
this service is not directed at people who are under 13.
Q169 Chair: So you have to go into your privacy
policy and read it and then you discover that you are
supposed to be 13?
Sinéad McSweeney: Yes.
Q170 Chair: Right, okay. What about fake identities,
people who open accounts in false names?
Simon Milner: With Facebook, that is against our
rules. You have to use your real name on Facebook
and, indeed, we think that is the core of Facebook’s
success. We would not have reached 1.2 billion
monthly active users had we not have that policy
because Facebook is all about connecting with people
you know in the real world. So it is against our terms
and we very much encourage people if they see
somebody who they think is not using their real name,
they are an imposter, to tell us about it and we will
then look into those accounts and remove them if we
conclude that is the case. We may require people to
prove their identity if they claim that this name that
appears very fake is real. That is something that we
take seriously and indeed is at the heart of our
mission.
Sinéad McSweeney: We do not require users to use
their real name. Most people do because they want
other people to be able to find them and connect with
them on the platform. We allow people to use
pseudonyms. We allow people to set up parody
accounts that range from emergency cute puppy
accounts through to political satire. More importantly,
we have seen the platform used in areas of conflict by
human rights activists, journalists, people who are at
risk of harm from the information that they are
attempting to get out into the world publically. That
is very important to us and that is why we facilitate
that kind of political activism and content on the
platform.
Chair: As I understand it if I set up as Donald Duck
there is no problem with Twitter.
Tracey Crouch: That has already been taken.
Chair: Donald Duck the 52nd. Twitter, you are
perfectly content. Facebook, it is only if somebody
tells you that I am not Donald Duck and then you will
do something.
Simon Milner: No, it is not if somebody tells us they
are not Donald Duck, it is if somebody else tells us.
Chair: Yes, I understand.
Simon Milner: There are also certain names that we
will look out for so you cannot set yourself up as
David Beckham, for instance. We also, like Twitter,
provide a process for people to have their accounts
verified. This is particularly useful for people in
Government, for instance, so you know that when you
look at David Cameron’s page on Facebook it is
David Cameron and not somebody pretending to be
David Cameron. That is something that we are rolling
out to help people who are public figures to ensure
that people understand yes, this is the real person.
Q171 Chair: Let us assume Donald Duck52 indulges
in highly offensive behaviour, trolling or even posting
child sex images, how do you identify them? Do you
keep records and ISP addresses?
Simon Milner: If somebody is using our platform to
try and distribute child abuse imagery then that is
going to be spotted through photo DNA assuming it
is a known image. As you may be aware, we are
working together with other technology companies to
try and ensure that we share even more intelligence
around these known images.
Chair: I am asking how you identify whose account
it is.
Simon Milner: Reports are incredibly powerful. We
kind of have a neighbourhood watch community of
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 52 Culture, Media and Sport Committee: Evidence
19 November 2013 Simon Milner and Sinéad McSweeney
scale, and our users really care about their safety and
the safety of others in the platform. They will be
reported to us and then we will deal with it.
Chair: You retain IP addresses?
Sinéad McSweeney: Yes, we retain communications
data.
Q172 Chair: That does not address my question. It
may be important to you, you may find it but if it is
being posted by somebody who claims to be Donald
Duck52, how would you identify it in that instance?
Simon Milner: I see, I am sorry. In the situations
where, for instance, law enforcement are on to
somebody like that and they want to find out who they
are and where they are living and what have you, we
have a process for them to request communications
data from us to enable them to do that. That would
include the IP address.
Chair: You will be storing the IP addresses and the
messages for how long?
Simon Milner: We store the IP logs for a period. I
would have to come back to you to tell you exactly
how long. The messages, it will depend. Typically it
depends on how long a person wants to retain those
messages but most of the time law enforcement will
have the messages because somebody has come to
them saying, “Look at this thing on my account”.
They have the messages, they have the content, what
they need to know is who is this person and that is
when they can come to us to ask for the
communications data.
Chair: Twitter, do you keep that same data?
Sinéad McSweeney: Yes, we also retain
communications data. We have protocols and
guidelines for law enforcement on how they can work
with us to request information, and they can also
request that we preserve information pending receipt
of further requests from them.
Q174 Chair: You would allow law enforcement
access to that with a court order, would you?
Sinéad McSweeney: We receive law enforcement
requests. They range from emergency situations where
if there is an immediate threat of life we respond to it
straight away, and also in other areas of serious crime
we deal with requests from law enforcement on a case
by case basis.
Chair: Civil actions?
Sinéad McSweeney: Civil actions, again it would
depend on circumstances. It is difficult to talk about
hypotheticals but on receipt of valid legal process we
deal with those requests.
Chair: I would need to get a court order.
Sinéad McSweeney: Yes.
Q173 Chair: But you are keeping records of tweets
and IP addresses from where they originate.
Sinéad McSweeney: The tweets are public anyway.
Chair: They might get removed or deleted.
Sinéad McSweeney: In the main most of the activity
is public and remains public, and we will retain a
certain amount of communications information.
Q175 Chair: A court order from the UK, would you
accept that?
Sinéad McSweeney: Again it is not my exact area of
expertise the actual legal process that is required. I
would need to come to you on that but we would need
valid legal process.
Chair: It was suggested to me you had to get a court
order in California, which is not always terribly easy
for a civil action for say privacy invasion or
defamation.
Sinéad McSweeney: Yes, but there are various treaties
and so on that cover civil litigation and co-operation
between jurisdictions also that would be invoked in
those circumstances.
Chair: I think my colleagues have exhausted their
questions so thank you very much.
Simon Milner: Thank you very much.
Sinéad McSweeney: Thank you.
Examination of Witnesses
Witnesses: Tony Close, Director of Content Standards, Licensing and Enforcement, Ofcom, Claudio Pollack,
Group Director, Content Consumer and External Affairs Group, Ofcom, gave evidence.
Chair: For our second session this morning may I
welcome Tony Close and Claudio Pollack both
representing Ofcom?
Q176 Tracey Crouch: Could you perhaps just start
by telling us what regulatory powers you do have in
relation to the internet, please?
Claudio Pollack: Yes, our role is limited to television
and television like content broadcast over the internet
in the UK, originating from the UK. We do not have
any powers that go beyond that, the regulation of
content.
Q177 Tracey Crouch: That would include video
content for example, child abuse videos or rape
pornography, or something like that? That would
include that?
Tony Close: If that content was included in a regulated
service. If it was included in a linear television service
provided over the internet or if it was included in an
on demand service provided over the internet and
established in the UK, yes, it would include any
content.
Tracey Crouch: But if somebody posted up a
homemade video, for example, this would not be
included within your regulatory powers?
Tony Close: Probably not, no. Not if it is not part of
a broader regulated offering, a broader service is
regulated.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 53
19 November 2013 Tony Close and Claudio Pollack
Q178 Tracey Crouch: Given the extent of Ofcom’s
powers and the obvious expansion over time as
technology has changed, do you think there will be a
greater role for Ofcom in the future regulating
internet content?
Claudio Pollack: Shall I take this one? I think at the
moment there are a mix of instruments around making
the internet safer for people. There may or may not
be a different role for us in the future, and that is
obviously a matter for Parliament. What I would say
though is that the mechanism that exists today, the
long established mechanism that exists today for the
regulation of television services, which has been very
effective, I do not think it would be appropriate or
possible to extend that form of regulation across all of
the content on the open internet. I think a different
mix of instruments will be required over time to make
the internet safer for those people that use it.
Q179 Tracey Crouch: You do have people working
at Ofcom on harmful and illegal content on the linear
platforms that you were talking about?
Tony Close: Yes.
Tracey Crouch: How many people do you have
working on it?
Tony Close: Within the standards enforcement team
within Ofcom we have around 40 people but I think
if you included some of the other teams within Ofcom
that have a role in monitoring, researching or
developing policy around potentially harmful content,
it would be around 50 people. They, of course, work
on a range of different content related issues at any
one time.
Q180 Tracey Crouch: If your powers are extended
into wider internet use, perhaps like draft
Communications Bill at the moment in front of the
House, do you foresee any problems with Ofcom
having further powers or wider powers to do with
this issue?
Claudio Pollack: It really depends what it is that that
extension means.
Q181 Tracey Crouch: Given obviously this inquiry
has been looking at harmful and illegal content on the
internet, and yet you by far the most powerful
regulator on these matters in general terms for
television and radio do not have any regulatory
powers, do you think that it is time that you did and
that you were able to help in terms of regulating
content in a way that is obviously very much needed?
Claudio Pollack: I wonder if it is worth saying a little
bit about how we carry out protection on television
and radio services and then maybe we can explore
whether there are aspects of that that would be
appropriate or relevant in the open internet.
Tony Close: Yes, I think it is worthwhile reflecting
just quickly on that. We do have powers in this area.
They are limited by statute. They do require us to
regulate linear television or on demand services when
provided over the internet as well as via other
mediums or platforms.
Chair: This is what you delegate to ATVOD?
Tony Close: In relation to video on demand, yes.
Across those services that we currently regulate on the
internet we have a comprehensive set of rules. We are
required in relation to video on demand with ATVOD
to implement those rules. We continuously monitor
compliance with those rules through complaints or
proactive monitoring. We refresh those rules where
we think that the industry has moved on or public
policy concerns have moved on. We have very
significant powers currently to either fine providers
who have broken the rules or to stop or to suspend
their services online or not online. So we have a
comprehensive set of powers and a comprehensive set
of rules that apply to these services even when
provided over the internet already. I think the
challenge is identifying the best aspects of the way
Ofcom currently operates and identifying the best
aspects of the existing framework and seeing how
they could be applied to a broader set of services
potentially online.
They are obviously limited by geography, our powers,
and there is a significant challenge around regulating
content over the internet that comes from outside the
EU of course.
Claudio Pollack: For example it is not possible at the
moment, as I understand it, to broadcast in the United
Kingdom without a licence that is granted across
Europe. On the open internet content can come from
anywhere in the world. of course.
Q182 Tracey Crouch: The on demand services, so
the XXX channels for example that are broadcast on
Freeview, presumably if, for example, they have
specific licensed agreement that they can only provide
programmes after midnight or something, if they were
to bring that forward to 10.00pm that is where you
would step in and remind them of their duties?
Tony Close: Yes, in relation to both linear and on
demand
adult
services,
services
providing
pornography that are regulated by Ofcom and by
ATVOD as well there is a broad range of interventions
and measures based on time, based on preventing
access to children, double, triple PINs and obstacles
for gaining access. Obviously there are limits on the
content as well that are already in place, limiting the
explicitness in particular environments. For example,
you cannot have hardcore pornography on Freeview
even behind considerable PIN and other access
restrictions.
Q183 Chair: In terms of what is and what is not
possible, there has been a lot of press speculation
about the need to have greater control over the internet
and I suspect a lot of it simply is not possible but one
of the things that particularly you have looked at in a
different context is website blocking, URL blocking
particularly. When you last looked at that you threw
up a whole lot of technical reasons why it was very
difficult to do. Is that still the case in your view?
Claudio Pollack: We looked at it in the context of
online copyright infringement and whether website
blocking could play a role in that. We absolutely did
highlight a number of technical challenges and
limitations but for me the report was very clear in
saying that, to making it more generic and applicable,
if this about lessening the availability of content that
is considered harmful, harmful in different ways, that
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 54 Culture, Media and Sport Committee: Evidence
19 November 2013 Tony Close and Claudio Pollack
website blocking can play an important in that. It is
not the answer to everything. It is not 100%. It is not
watertight. It can be bypassed and it is not perfect in
the sense that, given the amount of content out there,
it can over block. It could block things that you did
not mean to block and it can under block, it can allow
certain things through.
In both of these examples, public policy examples, if
you take blocking or filtering as part of a suite of
measures then it absolutely can have an impact and
can play a role.
Q184 Chair: It can play a role but if I am a
technologically literate child abuser seeking to
overcome technical measures, can I do so?
Claudio Pollack: I am hesitating because the
conversation I was having was focused more on
content that is generally lawful but accessed by
children that we do not want children to access. If we
are talking about illegal content, for example child
abuse content access on the internet that is slightly
further outside our area of expertise. The simple
answer is at the moment, yes, my understanding is
that there are technologies in place that allow those
that wish to access these images to bypass controls
that are put in place.
Chair: We are told that a lot of this material is now
lurking in the dark net as they call it, Tor is
particularly indentified. Is it your understanding that
is something that it is possible to intercept and identify
or is it so hidden that it cannot be reached?
Claudio Pollack: I think we are reaching the limits of
our expertise in terms of what Ofcom does. We are
far more focused, as I say, on understanding how
people can be protected or protect their children from
content that they do not want to be viewed in terms
of this aspect, these two aspects of child protection.
Tony, your turn.
Tony Close: The only thing I would add—and Claudio
is right, this is outside our remit—there have been
some limited but notable examples recently of
prosecutions or investigations of people using Tor or
other dark web browsers and routers. It is possible and
it has happened on very limited occasions.
Claudio Pollack: It is obviously a big, big area of
focus for CEOP, as was, and the IWF. When I have
met with them we have explored and also assisted
with some of the technology expertise that exists
within Ofcom in understanding the mechanisms
around identification and blocking. There will be,
given the current state of technology, imperfect
mechanisms but I suspect they have a very strong role
in law enforcement as well.
Q185 Chair: In terms of the role Ofcom plays
therefore in encouraging the provision of child
protection filters and that kind of thing, how much
power do you have to lean on the ISPs for instance?
Claudio Pollack: Beyond the areas of broadcast and
on demand content on the internet, our role is not
about powers. We do not have any powers to mandate
or in particular we do not even have a function, I
would say, to engage with internet service providers.
The real legal route and where our remit fits here is
around our media literacy duties. It is very much
about understanding and encouraging media literacy
among, in this case, children and parents, and we will
carry out a lot of research and therefore be the keeper
of the facts. I would add that as part of that role, our
main role on online safety, we have a seat on the
executive board of UKCCIS, which is the main
institution that leads activity in this area. As part of
theirs and the Prime Minister’s negotiation or
agreement with the main internet service providers to
provide network level filtering by the end of the year,
they have asked us to carry out a particular role in
understanding how that implementation goes.
Tony Close: Shall I add something to that? Before
talking about the specific reports that we are carrying
out into parental views of online behaviour and safety,
it is worthwhile just making absolutely clear that in
our role as a board member of UKCCIS we provide
all of the data that we harvest as part of our media
literacy programmes so that public policy can be
developed, public debate can be had about the risks
and benefits of the internet and of child internet use.
Claudio touched upon some reports that we will be
doing over the next year for the Government around
child internet safety.
Next month we hope we will be reporting to
Government on parental awareness of, take up of and
confidence in parental controls and other strategies,
mediation strategies, available for parents to protect
their children in an online environment. We will be
following that up in one year’s time with a follow
up report to look at how things have progressed. Just
quickly in between the two reports we will be
producing a third report for the Government that looks
at progress made by the four major ISPs in
implementing network level filtering.
Q186 Chair: You say you do not really have much
control over the ISPs, that your main area is TV
companies and radio, but the Government has said
that it wants to ensure that internet enabled devices
are supplied with filters and that you are going to be
involved in overseeing this. What does that mean?
Tony Close: I think it is what I was just talking about.
I do not think we have a specific role.
Chair: But this suggests that you are going to have
one in the future.
Tony Close: I think that that is just another way of
saying that at least for the next year we will provide
a number of reports that provides the Government
with our insights into how network level filters have
been implemented, whether or not ISPs have delivered
on the promises. More than that they will provide our
insights on how parents feel about the level of
protections that are available to them and whether or
not they are confident or not confident in protecting
their children in an online environment.
Chair: You see Ofcom’s role essentially as one of
research and opinion surveys.
Claudio Pollack: Within the current statute that is
exactly right. Beyond the areas that we have identified
and described, there is no regulation of the internet in
this country in the statute and we do not have a role
in that nor do we have a role, a formal role in policing
or mandating filtering. The conversations that
happened were all about a voluntary commitment to
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 55
19 November 2013 Tony Close and Claudio Pollack
deliver a form of network level filtering and Ofcom,
because of its media literacy duties and because of its
expertise in networks, has been called upon by the
Government to help in a degree of evaluation of the
extent to which the filters have been properly
introduced and also the impact that that is having on
the way that parents and children use the internet.
Q187 Chair: But you do have a role in licensing
telecoms companies, do you not?
Claudio Pollack: We have. I was going to say that
when you said we only do television and radio that is
in terms of protection of UK citizens from harmful
content on radio, television and TV like content on
the internet. Separately, of course, we have a role
around the regulation of telecommunications
companies as providers of electronic communications,
networks and services. It is not through a licensing
regime, it is through a general authorisation regime
but that explicitly does not cover regulation of the
content that is delivered by those networks.
Q188 Chair: It does not but if the Government chose
to say, “We are now going to make it a requirement
that you also have child protection measures”, you
could do that?
Claudio Pollack: Yes. To be very clear, what we are
describing is what our role is within the existing
statute. Of course there could be a different role if the
statute was changed but that is obviously a matter for
Parliament, not for us.
Chair: That would require primary legislation to
change the terms of the regulatory regime for
telecoms.
Claudio Pollack: I believe so, yes.
Q189 Mr Sutcliffe: Just on that theme, I think you
are pointing out you are a regulator without powers in
terms of this area and clearly that is a decision for us
as Parliamentarians to make, but I think your point us
to a direction. What concerns me then is if the ISPs
know that you do not have any regulatory powers
what is your relationship like with them in terms of
how are you seen by the ISPs in this sort of area? Do
they work with you in a proactive way or do you have
to react to what the agenda is? I am trying to get a
feel really in terms of what the relationship is like
without you having the powers to be able to say to
them, “You must do this, that or the other”.
Tony Close: I wonder if I might answer first in
relation to the specific task we have ahead of us in
the next year, and whether Claudio may want to add
anything about our broader relationship. We are
conscious that because we are limited by statute that
we do not have specific powers to require ISPs to do
certain things for us in order for us to carry out a
process of reporting to Government on network and
filtering. As part of the agreement with Government
on doing this work, we secured agreement that the
Government would contact the major ISPs themselves
and ask them to co-operate with us in order for us to
deliver this set of reports. I think we all agree that it
is important and we cannot do it without the
co-operation of the ISPs. We have no broader role
with ISPs but, Claudio, I wonder whether you want to
add anything?
Claudio Pollack: I have a lot of experience in dealing
with the internet service providers in those areas
where we do have power and that is a particular type
of relationship as you would expect. In this area it is
quite straightforward. We do not have any powers but
nor do we have any duties or functions so there is not
a gap if you see what I mean. Where we would be
coming in and saying, “There is a problem here in
terms of what we are expected to deliver”, would be
if there is a gap between our duties and our powers to
deliver that.
To be very clear, the provision by the four largest ISPs
of network level filters to be used by parents by the
end of this year was an agreement between the Prime
Minister and the internet service providers. That
agreement is about what they deliver but also as part
of that agreement and in the exchange that we have it
must be that they are required to co-operate in any
form of evaluation of whether they have delivered to
what is essentially a voluntary deal. If there is not
co-operation, for example if we find that it has not
been implemented properly or substantively or there
is not co-operation in our ability to carry out that task
then of course it is an agreement between the
Government and the ISPs and it is for the Government
to decide, if this voluntary arrangement has not
delivered what it hoped to deliver, whether it wants to
consider other avenues as well. At the moment it is a
voluntary agreement in which we are assisting.
Q190 Mr Sutcliffe: Tony, to go back to the survey, I
think, you are doing of parents, this report on parental
involvement, which I am pretty keen on because I
think that part of the problem is that parents do not
know what is going on in their own households. How
do you go about it? What are the terms of reference?
What is the background to it?
Tony Close: You are going to have to forgive me if I
cannot provide you with full details of this because I
am the guy at Ofcom that looks after setting rules and
enforcing those rules. I am not the guy that looks after
the quite complex and comprehensive programme of
research that we do but I will have a stab at answering.
Claudio, you might want to follow up.
We carry out a range of different quantitative and
qualitative pieces of work. We are doing them on an
ongoing basis so that we can benchmark and track
changes in parental attitudes. The qualitative pieces
are obviously limited in terms of numbers of people
that we talk to but really gets under the skin of
parents’ and teens’ attitudes to internet use, online
safety, what concerns them or worries or distresses
them, what they like about it, what they understand
about parental controls. The quantitative piece is
obviously different. It is bigger, it is less detailed but
tries to canvass a significant range of views and
attitudes to what the risks are online, what steps
parents take to mediate or mitigate those risks, what
strategies that they develop. It is quite comprehensive.
I cannot go through all of it.
Q191 Mr Sutcliffe: I think somewhere in the report
it talks about there is evidence now that people are
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 56 Culture, Media and Sport Committee: Evidence
19 November 2013 Tony Close and Claudio Pollack
more aware of what the issues are because of
experience, if you like, online. How often are these
reports commissioned? How often are the surveys
changed?
Tony Close: Okay, one of the reasons the Government
were probably keen for us to do the work is because
they are aware that we already do this on an annual
basis as part of our media literacy programme. Those
two reports, and there are other reports, programmes
of work that we carry out as well that might feed into
this are a rolling programme of work. I cannot tell
you how long we have been carrying them out for but
I think it is a considerable period of time.
Claudio Pollack: I think it is since 2005, so annual
surveys since 2005.
Mr Sutcliffe: There is an opportunity now to see a
pattern of things happening.
Tony Close: Yes. It enables us to track people’s
behaviours and attitudes as their internet experience
changes, as they become more confident or as they
become less confident, as they become more aware of
risk or less aware of risk.
Mr Sutcliffe: I do not know whether it is feasible,
Chairman, but it might be worthwhile getting a copy
of the sort of questions that are asked of parents, if
that is possible.
Claudio Pollack: Under the media literacy duty every
year we have a research report, a very, very
comprehensive research report one on adult media use
and attitude and one on children and parent media use
and attitude. It is the primary source, as Tony said,
used by UKCCIS to understand changes and it is
absolutely vital in understanding how things are
moving in this area, with limitations. There are a
number of things that it shows around attitudes. One
of them is that most parents are doing something,
there is a number of things they could do, but most
parents are also reporting some difficulty in using
certain tools. No majority is using, for example,
filtering but a combination of using filters, being in
the room when the child is using the internet, talking
to their children once a month, the combination of
those three mechanisms is 85%. So 15% are doing
nothing, but within that 85% there are a number of
people that are doing just one or just two of those
things.
The other thing that I think is fascinating—and, yes,
we will absolutely share the questions of what we are
doing annually—is the specific targeted report that we
have been commissioned to do using that data. What
makes this area so difficult is just how quickly it
moves. One thing that I am going to quote to is if
parents were saying and taking comfort from the fact
that the PC was in the room that they were in and they
could observe what was happening, in just 12 months
tablet computer use at home among five to 15-yearolds between 2012 and 2013 has gone up from 14%
to 42%, and that gives you a real sense of the
challenge of protecting children. Even when parents
are investing the time and effort to do so, it is very
difficult because things move so quickly. There is a
lot in the research that will help understand that better.
Tony Close: Just very quickly, Chairman, of course
we will share the questions with you and we will share
the report once we have finished it as well.
Q192 Chair: Just to be clear, the Government has
waved sticks at Google, at Microsoft and at the ISPs.
The Prime Minister has basically said, “Report back
to me within so many months” with the implicit threat
that unless they do these things action is going to be
taken against them. The position is that at the moment
there are not any powers to take action.
Claudio Pollack: Sorry, say that again?
Chair: The powers do not exist at present to take
action.
Claudio Pollack: In the area of regulating internet
service providers in having a role in policing content
the powers do not exist today. Of course if there are
individuals who have the capacity to introduce
legislation then they have a different response.
Chair: Of course, but therefore the Prime Minister
and Ministers’ implicit threats are that they will have
to introduce new legislation to take powers to require
internet companies to do these.
Claudio Pollack: Yes, that is right.
Q193 Chair: You do not have at the moment really
any provision within the licensing requirements that
would cover this kind of area?
Claudio Pollack: There is nothing. It is the general
conditions of entitlement that is through legislation
that derives from Europe around electronic
communications. Not only is there nothing in the
general conditions nor do we have the capacity to
introduce regulation that is around content using that
instrument.
Q194 Chair: Television like material is covered by
ATVOD?
Tony Close: Yes. The powers are vested in Ofcom.
Ofcom contract out the function of regulating
television like video on demand services to our coregulator, ATVOD.
Chair: Would that cover YouTube?
Tony Close: It would cover parts of YouTube.
YouTube is a mix of services. It is in large part and
most notable for user generated content and hosting
user generated content but increasingly it is a platform
for professional services provided from different
corporate entities. There are a number of professional
YouTube channels that are established in the UK that
meet the definition of a television like service and are
regulated by Ofcom and ATVOD.
Claudio Pollack: To be clear, in order to be regulated
they both have to be TV like and they have to be
established in the UK. Content that is visible on the
internet that is neither TV like nor established in the
UK, we would not have jurisdiction.
Q195 Chair: But if I post a video on YouTube that I
filmed on my handheld camcorder or whatever, is that
covered by ATVOD?
Tony Close: No, probably not, not if it is user
generated content and it is just a part of broader
offering. It will be covered by YouTube’s own
community standards.
Q196 Chair: Sure, so how do you draw a line
between YouTube content that is covered and
YouTube content that is not covered?
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 57
19 November 2013 Tony Close and Claudio Pollack
Tony Close: Partly on terms of jurisdiction, to put
them on to the site you have to be established in the
UK.
Chair: All right, so assume we are in the UK.
Tony Close: Then you have to identify a service that
exists, something that you can see, feel, something
that is whole, and then you have to run through a
series of potential criteria or characteristics to
determine whether or not it is TV like or whether or
not its principal purpose is the provision of TV like
material to consumers. This statute, I think, contains
a number of conditions, all of which need to be passed
in order for us to determine, or ATVOD in the first
instance, if something should be subject to regulation.
Q197 Chair: Do you feel that this whole area needs
further Parliamentary examination, because I was
involved in the Bill that set up Ofcom and none of
these services existed at that time?
Tony Close: I think it is fair to say that there is a
challenge. There are different degrees of protection
offered for different types of service online, very
comprehensive rules set for linear services online, less
comprehensive rules set for television like on demand
services, and no statutory rule set or intervention for
certain other types of material. There is a challenge
there. It is a legitimate debate to be had about whether
that is an ideal scenario for consumers or whether or
not we should be thinking about a more coherent
framework for media standards that spans a greater
set of audio visual media. I think there are certainly
some benefits in thinking about that.
Q198 Chair: Will this feature in the Communications
Bill, do you think?
Claudio Pollack: You have someone giving evidence
shortly that I think would be better able to answer.
Chair: Indeed, and that is a question we might put to
him. I think that is all we have, thank you.
Claudio Pollack: Thank you very much.
Tony Close: Thank you.
Examination of Witnesses
Witnesses: Rt Hon Damian Green MP, Minister of State for Policing, Criminal Justice and Victims, Mr
Edward Vaizey MP, Parliamentary Under-Secretary of State for Culture, Communications and Creative
Industries, and Claire Perry MP, Special Adviser to the Prime Minister on Preventing the Commercialisation
and Sexualisation of Childhood, gave evidence.
Q199 Chair: Good morning, thank you for attending
our third panel this morning. It is afternoon now. May
I welcome Damian Green, the Minister of State for
Policing and Criminal Justice from the Home Office,
Ed Vaizey, Minister for Creative Industries and
Communications, and Claire Perry, the Special
Adviser to the Prime Minister on Preventing the
Commercialisation and Sexualisation of Childhood?
We have just been having a discussion with Ofcom
about what powers are available to them should ISPs
and other internet involved companies not voluntarily
adopt some of the measures that we have been
encouraging them to take, to which the answer was
there were very little powers. Do you think we need
to give Ofcom more powers in this area?
Mr Vaizey: We await your guidance, Chairman, on
who you would like to answer.
Chair: I think possibly, Mr Vaizey, you are best
placed.
Mr Vaizey: I knew I should not have said anything. I
think Ofcom obviously has some powers in this area.
It is clearly a regulator of broadcast services and
co-regulates what we call video on demand with the
Association of Television and Video On Demand.
Anything that is a broadcast like service that is
broadcast over the web, Ofcom has the powers to
regulate. As you know, it has also been asked by the
Secretary of State to report on parental awareness and
how comfortable parents are with the safeguards that
are now being put in place. They were also behind
creating ParentPort as a one stop shop to allow parents
to get the kind of guidance they need. We are working,
related back to video on demand, on common media
standards because we take the view that as television
and internet services converge anyone, parents,
consumers of any kind, will assume that the same
standards apply to what appears on their living room
screen, which I think you can still call a television for
these purposes.
Ofcom also sits on the UKCCIS Board so they have
a great amount of input into what we are doing in this
area. I would be open to any discussion on an
extension of their powers. I have taken the view that
we can achieve a lot without necessarily imposing
legislation or widely extending Ofcom’s powers at the
moment, for example the whole issue of filters where
we, I think, have made some rapid progress, has been
by engaging in a dialogue with the ISPs. Self
regulation, if I can put it that way, seems to me a way
of making the kind of rapid progress one needs, not
only just rapid progress but also keeping the flexibility
so that we can react quickly to events. Some of the
changes that Google have made, for example, would
not necessarily have succeeded, I think, as quickly if
we had had endless debates about the appropriate
legislation and so on.
But it is quite clear that Ofcom has a role in this space.
The key issue, I think, will be always to be open to
ideas and suggestions about whether or not Ofcom
should have its powers extended and if so how, what
its remit will be, particularly considering you are
dealing with global companies who effectively deal
with different regulations all across not just Europe
but across the globe. I do not know whether Claire
wants to add anything to that?
Claire Perry: Thank you. The independent
parliamentary inquiry that I chaired raised the issue
that there are numerous regulators in this space and I,
like the Minister, think it is an interesting and valid
question as to whether Ofcom’s remit should be
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 58 Culture, Media and Sport Committee: Evidence
19 November 2013 Rt Hon Damian Green MP, Edward Vaizey MP and Claire Perry MP
extended. The two areas where, I think, the issue most
arises is one around to whom does one complain if
one has problems around content or treatment, and of
course ParentPort is a good attempt to try to centralise
the complainant process. Perhaps more work could
and should be done around publicising that. Indeed
we will move onto the internet safety campaign that
is being funded by the ISPs. One of the asks is to
direct parents where to go, users where to go, when
there is a problem.
I think it is right and good that Ofcom has been given
an implementation role in the roll out of the policy
around filter adoption. One of the questions on the
filter process is what does success look like? With a
sense of about 30% of the households in the country
being those with children one might argue that is a
good, soft target though there are not, of course,
official targets for this. Asking Ofcom to look at the
implementation of that and also perhaps to distinguish
between ISPs and the success of their implementation
is a very valid thing to do.
Like the Minister I am very minded to think that if
we were trying to regulate such a fast moving space
in any form we would still be sitting round looking
for parliamentary time on the first debate on filtering
implementation. Instead of which we have
extraordinarily effective and dynamic filter roll out,
and indeed companies are competing to offer their
customers the best filters and the most family friendly
packages. I think an extended role is possible, and I
welcome the Minister’s commitment to looking
potentially at how that could look.
Q200 Chair: The whole thrust of the pressure that
has been placed on internet companies to date has
been rather, “Unless you agree to do this then the
Government will make sure you have to do it”. Do
you see any advantage perhaps, and we have a
Communications Bill coming, in the Government
taking powers rather as the Government took powers
in the Digital Economy Act to require ISPs to do
something if they failed to do it voluntarily?
Mr Vaizey: As I say I am open to any suggestions and
I am looking forward to the Select Committee’s
report. I think one has to be careful. If one took too
wide a reserve power, if I can put it that way, to
“regulate the internet” that would create an element of
uncertainty. Businesses need certainty and need to
know what regulations they would be required to
comply with. We need to know what problem needs
to be fixed and why legislation rather than working
with organisations and encouraging them would be the
answer. As you say, Mr Chairman, I do think that the
potential of legislation, and the Prime Minister made
that clear on a number of issues in terms of his speech
in July and continuing dialogue he has with the
industry, is always there.
My message to all the companies in this space is that
there is potentially an appetite in Parliament for
regulation and that they should not be blind to that
and therefore co-operation—and to use a terrible
American expression—getting ahead of the curve is a
much better place for them to be but they must
understand that is, as it were, the mood in Parliament
that I don’t think MPs would shy away from
legislation should they feel that they were not getting
the adequate results that they needed.
Claire Perry: Mr Chairman, just on that point, if I
may challenge you slightly, sir. The idea that
companies are doing this or else I think perhaps is
something that we felt a couple of years ago. I may
just be starry-eyed through lack of sleep after
yesterday but I genuinely think that the growth of
corporate social responsibility among those who are
making their living from anything to do with internet
activity is really quite impressive. Certainly the
British ISPs are well ahead of the ask in terms of
delivering and, with yesterday announcements, we can
see that some of the global internet companies who
perhaps dance to a very different legislative tune have
also woken up to that. So again perhaps we are in an
awakening of a new dawn of social responsibility on
the web, which is to be welcomed, but I would
perhaps slightly challenge that the companies are
living in fear of regulation. I think we are beyond
that point.
Q201 Chair: That is encouraging. Can I turn to the
illegal material because that is particularly going to be
I suspect more for the Home Office. The area of data
retention in order to identify perpetrators of
distribution of child abuse images and so on, that
strays into what has been a difficult territory for the
Government in recent months. Are you satisfied that
enough data is being retained and is available to the
police to identify people?
Damian Green: I think in this instance it is not a
problem of data retention because the civil liberties
aspect of the debate does not pertain here. If people
are producing, exchanging, storing child abuse images
then they are committing a serious criminal act. So,
as I say, both legally and morally it does not seem to
me that whatever civil libertarian arguments one may
wish to adduce about data retention in other spheres
apply in this case. I think the key is identifying the
images, blocking them, trying to get into some of the
more hidden networks, moving on peer-to-peer
exchange and so on. Those are the areas where
activity both by the industry and by this Government,
and indeed Governments overseas with whom we act
in concert, is concentrated.
Q202 Chair: Are you getting complete co-operation
from every single internet service provider for
instance?
Damian Green: In a sense this is more a matter for
search engines than ISPs and obviously it is both
because the ISPs will host the sites. One of the parts of
yesterday’s announcement after the Prime Minister’s
summit, was that Google and Microsoft have changed
their algorithms so that they can now both identify
and block child abuse images. People are using search
terms, 100,000 search terms and more, so they now
have a dynamic algorithm that will be much more
effective than original ideas of just producing search
terms and saying block all of those. People who both
produce and search for child abuse images are not
only evil, they are cunning and it does not take much
cunning to decide to change a search term slightly. So
you do need the full extent of the technological
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 59
19 November 2013 Rt Hon Damian Green MP, Edward Vaizey MP and Claire Perry MP
expertise that is available to companies like Google
and Microsoft to ensure that the algorithms can cope
with that. The tests that the Child Exploitation Online
Protection Centre and other bodies have been running
show that these are very effective, what is coming in.
Perhaps most encouragingly we learnt from Google
yesterday that instead of getting back a child abuse
image, getting back a warning page saying, “What
you are seeking may well be illegal and do not do
this” essentially, that has already reduced the amount
of searching by 20%. So it has had a very significant
immediate effect.
Q203 Chair: Obviously the progress that has been
made through search engines is welcome but a lot of
the distribution of material that is being done by the
hard core is not going to be done using search engines.
Do you think enough has been done and to what
extent can you assess the scale of the distribution that
is occurring either through peer-to-peer networks or
through the dark net?
Damian Green: You are absolutely right, Mr
Chairman, there are three elements to it. There is the
blocking of the initial search, which we should not
dismiss because what it will do is stop people
embarking on this particularly horrible journey so it
is worth doing in its own right. But you are right,
people who are hardened may well be using peer-topeer or may well indeed, for the worst type of abuse
images, now be using now the dark net, in the current
jargon. The answer to your question is yes in both
those areas. One of the other things we announced
yesterday was a pilot project that CEOP, along with
the Internet Watch Foundation and the industry
companies will be working on to see how we can deal
with peer-to-peer. That is obviously different from
using a search engine algorithm for reasons the
Committee will understand. So that is one project.
One of the key areas where the UK/US taskforce, that
I will coach here with an Assistant Attorney General
of the American Government, will be to bring together
the best industry technologists along with the best
people, GCHQ and their American equivalents who
are actually very good at encrypting and therefore
decrypting secret networks, to use all those skills
precisely to get in to the dark net and so on. So we
are working on all three of those areas.
Chair: I think that probably leads neatly in to Gerry
Sutcliffe’s area.
Q204 Mr Sutcliffe: I do not know if Mr Green has
had the opportunity to read the evidence that was
given to us by the existing Chair of CEOP and the
former Director of CEOP, quite opposing views about
where CEOP sits. An interesting perspective from the
Minister about how CEOP’s role is developing and
what are the strengths and what are the weaknesses?
Damian Green: The big change, I agree, having read
the evidence, there was a difference in emphasis, if I
can put it like that, between the former view and the
current view of CEOP, is that CEOP has changed very
radically over the past couple of years and has had its
most radical change in the last few months by
becoming an arm of the National Crime Agency. That
is a game-changer for CEOP because it is not now a
small very effective, very good outfit, it is now a very
effective good outfit that is part of a very large outfit.
For example, the NCA will have 4,000 officers, all of
whom will be trained in child protection because it
has its national links, it has a border arm as well as
taking over many of the international policing
functions that we need in this country. It can, if you
like, look the FBI in the eye and say we can now work
with you as an equal, and since this is clearly a global
problem it requires global solutions. The fact that
CEOP itself is now part of a global size police
organisation, the NCA, I think gives it opportunities
and access to expertise that it did not have in the past.
Q205 Mr Sutcliffe: Regardless of others’ personal
views, what they were unanimous on was that there
were about 50,000 potential areas where people could
be convicted on internet abuse. What are the resources
going to be like for CEOP in terms, and I accept what
the Minister is saying in terms of being part of that
bigger organisation, there may be some possibilities
of it being not downgraded but diluted in terms of its
ambition and in terms of where it needs to go?
Damian Green: Absolutely not. You are quite right
that the CEOP threat assessment, which it inevitably
has to be—it is the best guess we have that is what it
is—was up to 50,000 individuals involved in
exchanging child abuse images. That was this year’s
threat assessment. But, no, I think the exact opposite
of dilution is happening. It is being amplified that
CEOP now has access to people, expertise, resources,
international access in a way that it did not have by
itself.
Claire Perry: Just to add some colour to the point
the Minister made about exchanging, the interesting
behavioural factor of those engaged in this traffic is
that they connect. They seek reaffirmation for their
desires from connection and in a way that makes the
problem easier in that if you are going after a
particular node or interrupting a particular site, and
we have seen this in several busts recently, you have
potentially access to a large group who have
connected through, even in the dark internet, a node.
Mr Chairman, if I may just throw back a point about
the issue of the dark net, I mean things like Tor, The
Onion Router, these are often set up by Governments
and indeed funded by Governments as a way of
providing anonymity to people for various reasons.
One thing that would be very interesting to flag up to
those Government departments who are setting up
these organisations is what the unintended
consequences may be. Internet anonymity is both a
blessing and a curse, particularly if people are using
it for illegal purposes. I don’t think we think enough
about that when we are looking schematically at the
internet.
Q206 Mr Sutcliffe: If I can just return to the
regulation point with Ed and Ofcom. What was
interesting this morning, Ofcom’s evidence was that
the regulations were not there but one can understand
that for the rate of progression and rate of change that
takes place. I think the negotiated route is probably
the best route to be taken and it is heartening to hear
that the ISPs are dealing with that. One of the things
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 60 Culture, Media and Sport Committee: Evidence
19 November 2013 Rt Hon Damian Green MP, Edward Vaizey MP and Claire Perry MP
that I think that we need to look at is parental
involvement in all of this. Again, Ofcom this morning
said that they are conducting surveys of parents going
back perhaps to 2005. What can the Department do or
what can Government do in general terms to try and
raise awareness of the potential of these issues
affecting vulnerable young children?
Mr Vaizey: As I say, Ofcom was behind setting up
ParentPort because it felt you needed a one-stop shop
for parents to go to get all the advice they needed.
The Secretary of State has asked Ofcom to report
regularly, three times over the next year or so, on
parental awareness. The other thing that I think has
been of benefit from the negotiated route is that we
have got the ISPs to agree on a marketing campaign.
I took the view these are organisations that know their
customers, they have to sell them services, they will
know them back to front and, dare I say, each
company will have pretty well paid marketing
directors who should know their trade. So I think it is
a massive opportunity to have the four of them work
together. I know what Claire says about internet
service providers competing in terms of providing
services to parents, that is good because it will mean
innovation, but there needs to be a strong degree of
co-operation and to have them come together for a
marketing campaign is a good thing. My
understanding is that it is about £25 million worth of
advertising that they are planning to put in place and
I think next Spring is the target date. Hopefully there
will be a strong and high profile campaign to make
parents aware of the filters and the kind of things they
can do to protect their children from inappropriate
content. It is also important to say that should not be
a one-off, you cannot just have a blitz once and then
forget about it and assume that people will know
about it for ever and a day. One has to keep refreshing
and reminding parents of what issues are out there.
Again Claire, who has been involved in talking to
them, might have something to contribute.
Claire Perry: I think it is interesting that this question
comes up as to why is this not a Government sort of
public health campaign. Indeed the last Government I
believe spent £3 million on the same campaign, which
was called, “Zip it, block it, something it” and frankly
nobody can remember what it was. It almost sank
without trace I would submit, and part of the problem
was it was written by us, by people who are not out
there every day working at how to get marketing
messages to their users. There is a collective
commitment of the ISPs to do a £25 million campaign
over three years to alert parents to the new filtering
technology and what they can do but also the broader
issue of conduct questions. Technology only solves
some problems, it does solve cyber bullying, it doesn’t
solve sexting. There is a lot of stuff out there. I have
always felt that parents on the one hand tell us that
they are technologically overwhelmed and they do not
understand but on the other hand are paying for these
contracts. It is like, “If you do not want your children
to be sexting then take their mobile phone off them
and stop paying for the contract or switch the router
off at night”. We somehow have abrogated this space
to our kids and I think it is time to remind parents that
they are part of the solution as well.
I will say that one of the big breakthroughs on
technology—forgive me for talking so much—is
moving device level filters to the home network level
is a massive step forward in making life easier for
parents. The ask now for parents to install a filter on
every single internet enabled device in their home is
a joke. The fact that we have persuaded the ISPs to
implement a home network level filter, which we were
told initially could not be done for technical reasons,
until TalkTalk did it, is massive because it means that
parents will only have to set the safety settings once
and know that everything in the home is protected.
That is an enormous step forward in helping parents
in my view.
Q207 Chair: Where are we on your suggestion that
there should be a default setting that is a strong filter
and that people should then be required to opt in to
receive stronger material?
Claire Perry: Having proposed that, we of course had
the Government consultation on that, which was very
welcome. We heard that people were all over the place
on this default on. What we have arrived at is that we
are going to ask every account holder in the country,
which is 20 million accounts, so this is a very
significant ask, whether or not they would like to
activate parental filters. Of course two thirds of them
will not be parents, so they are going to be asked for
something that perhaps is not relevant. We will have
the box, “Would you like the filters installed?” preticked. To me, in a way, that satisfies the desire to
have people make a choice. We basically present them
with an unavoidable choice.
Chair: Who is presenting them?
Claire Perry: I do not know who your internet
account is with, but you will have an interrupt at some
point in the next year and they will say, “You can now
install fantastic family friendly filters would you like
to come and install them?” The box will be pre-ticked,
“Yes”. You can’t just do nothing, you have to choose
to get off that page. That to me satisfied this
requirement to ask people to choose, which is the
classic tension. Do you ask people to choose or do
you make the choice for them in quite an elegant way.
It was a proposal from the industry that I felt satisfied
where we wanted to go. We ask everybody and if you
do nothing the filters are automatically installed or
you cannot move on.
Chair: Just to be clear, one day over the course of the
next year or perhaps—
Claire Perry: One year.
Chair: Right. Every single customer of every single
ISP is going to turn on their computer and suddenly
get a page come up saying, “Do you want filters?”
Claire Perry: To be clear, the top four ISPs who have
around 85% of the market, we would like the others
to commit to the same thing. They will ask you in
different ways. Some might ask you when you go to
query your bill online, some may interrupt your
browsing session, which is a first for the UK to do
that. This is the commitment that Ofcom will be
monitoring. Every household will be contacted and
asked whether or not they would like to switch on the
filters, and the box, “Yes, please” will be pre-ticked.
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 61
19 November 2013 Rt Hon Damian Green MP, Edward Vaizey MP and Claire Perry MP
Q208 Chair: You have conducted your inquiry, you
have talked about the change in attitudes of the
companies who are now much more willing to pursue
these kinds of things, are you pretty much satisfied
that everything you wanted to see happen has
happened?
Claire Perry: Well, yes and more in some ways. For
example, the Google commitment yesterday,
hopefully to be matched by Microsoft, to block illegal
returns for search terms globally was huge. So we will
basically be protecting the global internet with those
changes. However, I think one can never declare
victory because the technology moves on at pace,
which is why regulation of the space is so difficult. I
think we have a solution that is going to be what
parents have been asking for, what the country’s been
asking for. There may be a new way of accessing the
net or a new form of conduct or a new device that
pops up in the next couple of years that changes again,
which is why having the company self- awareness and
involvement is so important. You want them to be
thinking about these dangers rather than relying on
politicians to constantly be jumping up and saying,
“What about this?”
Q209 Chair: Given we expect more and more
internet browsing to be done on mobile devices, can I
check, are the mobile providers also going to be
putting up this question of saying, “Do you wish to
fit filters?”
Claire Perry: Interestingly, they already do that. The
mobile operators had an opt in system where you had
to prove you were over 18 in order to have that filter
dropped. They have both expanded that, so Three has
become involved and also they have moved to BBFC
filtering technology, so their filters are even better and
some of the issues like false positives have been
removed. In a way we have layers of protection
around young people now with mobile devices in the
home, and of course with public wi-fi.
Q210 Chair: Damian, can I come back to the
hardcore that you are concentrating on, the figure that
Gerry quoted and you confirmed, up to 50,000, when
we took evidence on this earlier it was suggested to
us that a number of those 50,000 who can expect a
knock on their door is pretty small. Is that satisfactory
or should we not be deploying more resources into
this?
Damian Green: Yes, and we are is the brisk answer
to that question. That is one of the points of having
CEOP as part of the NCA is precisely to improve our
enforcement capability. You understand that that
50,000 figure is a pretty broad brush. One of the
interesting things, as Claire just alluded to in another
context, is that you do tend to break open groups of
people, it is not a question of 50,000 isolated
individuals. The successful law enforcement
operations that take place, both here and overseas,
have two characteristics: firstly, lots of people are
involved, and, secondly, there may well be lots of
people in lots of different countries. You break open
networks.
One of the points made at the summit yesterday was
that the big sites that are effectively child sex abuse
image multiple sites, there are probably not that many
of them. There might be 10 or 12 of them based
around the world and when you get the law
enforcement capability to break into them you may
well find very large numbers of people who are
attached to them. That is in a sense the way you will
get the greater number that I accept we want to see.
Chair: Jim Gamble suggested to us that you could
perhaps have special constables who would receive
training and then be sent out to look for this material
and identify it.
Damian Green: To some extent a lot of that work, in
terms of identifying where the images are and from
then to the individuals is now being done more and
more by the Internet Watch Foundation, and the IWF
has a significant increase in its capacity coming with
the extra money from the industry. The big change
that will happen in the coming months is that instead
of being passive and receiving information from the
public the IWF will be able to go out there and
actively search. The key from that will be to make
those searches active, in terms of the individuals as
well as the individual images. I think we are meeting
this desire to have more pairs of eyeballs looking at
these images and from that will flow through the
information that hopefully enables us to go and finger
the individuals.
Q211 Mr Bradshaw: Damian, do you know how
many prosecutions there have been for threatening and
abusive behaviour online?
Damian Green: I don’t off the top of my head, I am
sorry.
Mr Bradshaw: Would it be possible for you to let
us know?
Damian Green: Certainly, yes. We will find out
about that.
Q212 Mr Bradshaw: Do you think it is a serious
issue?
Damian Green: Yes. It is a different part. I suppose it
develops out of cyber bullying in a way, so in a sense
it will be networks. Facebook is obviously the biggest
one, but we have seen a lot of the other ones where
this kind of behaviour has become—prevalence is the
wrong word, but we have all become more aware of
it simply because that is the arena in which people
exchange views and information. Therefore those who
are predisposed to be threatening and abusive can now
be threatening and abusive online.
Q213 Mr Bradshaw: Do you think it is important
that the criminal law is used as a deterrent, in terms
of our learning process and how we view this sort
of behaviour?
Damian Green: The principle one has to operate is
that if something is illegal in the real world then it is
illegal online as well. You can cause as much distress
to someone by threatening behaviour online as you
can offline. To that extent, yes, absolutely. There are
various stages one can go through online, and just as
criminals can use technology, therefore law
enforcement can use technology against criminals as
well. Similarly, the social networks know their
audience very well, they have the capability at the first
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 62 Culture, Media and Sport Committee: Evidence
19 November 2013 Rt Hon Damian Green MP, Edward Vaizey MP and Claire Perry MP
sign of this of enabling people to switch individuals
off, to report bullying and those kind of instant buttons
on screen that you can press. They seem to me to be
very good trip wires. At the end you need the criminal
law, but I think you need all those trip wires on the
way up as well.
Q214 Mr Bradshaw: Do you think the CPS shares
the seriousness with which you take this problem?
Damian Green: Yes. I think across law enforcement.
The CPS will be the recipient, that is what it is for. It
receives reports from the police. Again, you have to
go through the procedure, the police have to know it
is happening, you have you be able to collect the
evidence, they have to receive complaints, they have
to have people who are prepared to give evidence, all
those kinds of things. If they have cases that are best
resolved, as it were, at the end, as a criminal law case
then they will approach the CPS.
Mr Bradshaw: I have a question for Mr Vaizey, I do
not know if anyone else wants to come in before that?
Chair: Conor, do you want to come in now?
Q215 Conor Burns: I just want to ask very quickly
on the social media, do you think the social media
companies—Facebook, Twitter in particular—could
do more than they are currently doing?
Damian Green: They are now much more aware than
they were even a few months ago of the importance
of that. I have been dealing with this portfolio for
about four or five months and I am struck by the
change throughout the industry of some of these very
big companies—based and with their origins in the
USA with its different tradition—where freedom of
speech is such an absolute right that any kind of
reduction of it is regarded with extreme suspicion. The
journey that the companies, including Facebook and
Twitter, have taken over the past few months is to
realise that in these areas, particularly with child sex
abuse where there is no argument, but also in terms
of bullying and abuse that we have seen can lead to
tragic incidents of teenage suicide and so on, they do
now get it in a way that I suspect, until it was brought
to their attention, they didn’t quite.
Q216 Conor Burns: Earlier this morning when we
had evidence given from Twitter we were told that
there is a report abuse function now on Twitter that
was brought in in the end of July. None of us on the
Committee who use Twitter were aware that that had
been brought in. Do you think there is more they can
do to promote how to report abuse, because the
reporting of abuse means it is taken more seriously if
it is in volume?
Damian Green: Yes, I think that is a very good point.
The timing is interesting, it was in July Twitter had
got the point. All these companies have the
technological capacity to solve these problems, the
trick is to get them to recognise them as a problem.
There is then a further problem of letting users know.
I suspect in that world, once you have established
something new, so much of it does get out by osmosis
that people just know it. It may reflect our age that we
are not instantly aware that something new is
available on Twitter now, but you can guarantee that
around the world hundreds of billions of Twitter users
probably do know about it now, just because that is
the way things operate. But, yes, a bit of marketing to
the elderly might be useful.
Claire Perry: Could I come in with a couple of points
of reassurance and perhaps a suggestion? One is I had
a meeting with the Solicitor General on this very issue
because obviously I have attracted quite a lot of third
party cases on particularly bullying or stalking and
harassment issues. He may be worth taking a
statement from. I was very reassured that the law has
been looked at and expanded wherever necessary to
include online conduct as well as offline conduct.
Having been on the receiving end of quite a storm of
Twitter abuse over the summer relating to my
campaign, I have to say I don’t think companies do
do enough. I think part of the problem is the
anonymity of usage.
I was encouraged with Ask.fm—having spent a lot of
time with Hanna Smith’s father, who was one of the
young women who did indeed commit suicide—that
company did set up a facility where users could
choose to be anonymous, but you would know if the
user was anonymous when you were exchanging
information with them. People Tweet, people post
abuse about how they would like to rape you and kill
you because they think you do not know who they
are. If there was some way of the company knowing
and being prepared to verify that identity and to show
you that that identity was verified I think it would lead
to a diminution in that sort of behaviour. I don’t think
the companies do enough and there is a great concern
around it, as the Minister said, given the US legal
framework within which much of the global
companies operate.
Q217 Mr Bradshaw: Do you think it is important
that we see some prosecutions so that there is a
deterrent out there?
Claire Perry: Yes, I do. I think it is deeply
misogynistic as well and it was interesting over the
summer because there was a bit of flurry of push back
from a number of people in the public over this. I
don’t know that we saw prosecutions, we saw cautions
and we saw apologies but prosecuting, for what is
pretty vile behaviour, would be helpful.
Q218 Chair: On the point you raised, we explored
this point with both Facebook and Twitter, Twitter
basically told us they are perfectly happy for you to
use any alias you choose and there is no requirement
for you to tell them who you really are. Facebook said
to us that it was their policy that you should use your
real name and if somebody told them that somebody
was not doing so then they might take action about it,
but they would not take any action themselves to
check. Are you saying that you think that those two
companies should be doing some kind of verification
process?
Claire Perry: I think a way around squaring the circle
would be for them to indicate to other users those
users that have chosen to stay anonymous. I have no
issue, as I am sure many users of Facebook and
Twitter do, with my identity being public because I
am obviously there as a public person and I think most
cobber Pack: U PL: COE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Culture, Media and Sport Committee: Evidence Ev 63
19 November 2013 Rt Hon Damian Green MP, Edward Vaizey MP and Claire Perry MP
people would feel the same way. If you are aware that
so and so has chosen to remain anonymous and sit
behind their funny little avatar picture then it is your
choice as to whether you engage with them or indeed
accept communications from them. Perhaps the
companies could give us the choice of who we engage
with, but certainly flag up who is choosing to remain
anonymous. My sense is that might start to drive
down some of this abhorrent behaviour across all the
social communities.
Chair: I think it was John Carr who suggested to us
that Facebook should be required to verify the
identity. I am not sure that is practical, but is this an
area where the Government feels more could be done?
Mr Vaizey: Again, this is an area where you can work
with the organisations. I certainly think there is
mileage in sitting down with the social media
companies.
You made the point earlier, Chairman, about the
different approach of Twitter and Facebook in terms
of verifying identities, and I think there are a whole
range of issues that the layman would take as
important in a straightforward way, “How do I report
abuse? How quickly will they respond to me about
the abuse? How can I challenge a decision?” If they
decide that the neighbour’s Facebook page saying that
you are a such and such and a so and so is fair
comment rather than abuse, “How do I appeal that?”
Also, do the same principles apply across different
social media sites? Obviously individual social media
sites should have the freedom to have their own
approaches, but I think there is a way of joining things
up. Neelie Kroes, the Digital Commissioner, put
together some safer social networking principles a few
years ago. She asked some of the CEOs that work
with her on those to look at them again and update
them and my understanding is that they are going to
come out with a statement in early 2014. It will be
interesting to see that. I certainly do think there is an
opportunity for the Government to broker a dialogue
between different social media sites and also to
examine how effective their protocols are.
Q219 Chair: As I understand it, in terms of
identifying criminality the police would have to first
of all obtain an order to obtain an IP address off
Facebook or Twitter, you then have to go and probably
get a second order in order to get the ISP to reveal
whose the IP address is. Are you content that that
system is sufficient?
Damian Green: It does work.
Chair: It does work?
Damian Green: As Claire said, that there were at the
height of the particularly horrible trolling that is going
on over the summer, people were visited and the
police found out who they were and went and knocked
on their door. I just observe as a fact that it is possible
to identify individuals, even when they are trying to
be anonymous and saying repellent things using things
like Twitter.
Q220 Mr Bradshaw: Mr Vaizey, in your evidence
you singled out for praise the South West Grid for
Learning, part of the UK Safer Internet Centre that
happens to be based in Exeter, for the work it does in
mediating reports of abuse and threatening behaviour
and getting stuff taken down. I do not know whether
you are aware that they run on a shoestring and the
only money they get is from the European Union. I
just wondered whether you thought there might be
grounds for doing the same kind of thing with abusive
and threatening behaviour as already happens with
child abuse and there needs to be some sort of industry
scheme to fund an organisation like that, and perhaps
some national recognition from Government?
Mr Vaizey: Yes, I have met with the South West Grid
for Learning and I am aware that they run as a lean
machine and the vast majority of their funding comes
from Europe. I would certainly happily look at that.
That is a very good suggestion of the wider principle
that the ISPs and social media sites support the work
that is being done to mitigate some of the harms. The
mantra is there is a huge amount of good that comes
from the internet, but there is a significant degree of
harm. In terms of supporting the Internet Watch
Foundation and also working with CEOP, particularly
into supplying skilled engineers and technicians, I will
certainly look at that on your behalf.
Q221 Conor Burns: I would just like to go back to
what we were talking about earlier, about the filtering.
Claire, you say that people will have in the course of
the next year, possibly their browsing session
interrupted and you conveyed that as though that was
a very good thing. In pursuing the aim of clamping
down on illegal abhorrent content on line that is a
good thing, but would you agree that we have to be
slightly cautious that we do not step over the line into
seeming to be moralising people who do wish to
continue to access material that is perfectly legal
whatever one’s personal moral judgment about it
may be?
Claire Perry: I completely agree. A big question for
the ISPs was how do you get to people who most need
the filters. To be clear, we know that only four out of
10 families currently employ any sort of filtering
device at home. This is Ofcom data. It may have gone
up slightly in the last six months, but we have a real
problem in that the companies have worked quite hard
to generate filters and we know that the take up is
very poor. Browser interrupts is one of several ways,
other companies who have a very substantial amount
of online billing say that they can do it in the “my
account” space. But, of course, one question is if there
is a kid surfing at that time when the interrupt comes
up how does that satisfy the requirement, the adults
are the ones who should be putting on the filter. This
is for the companies to deliver.
One of the questions, and we did test this quite heavily
in focus groups, was do people mind being interrupted
or being asked about filters and is it too much to ask
them to unclick the “Yes, please” box, which is
effectively what you are asking. The response was
actually no. The analogy given was that it is a bit like
slowing down for a lollypop lady, that you may not
have school age children and you might be in a hurry,
but because of the general question of child safety you
are prepared to slow down and potentially even stop
for that school crocodile to pass. Again, there has been
a little bit of push back since the announcement was
cobber Pack: U PL: COE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG03
Source: /MILES/PKU/INPUT/035254/035254_o003_steve_CMSC.19.11.13.xml
Ev 64 Culture, Media and Sport Committee: Evidence
19 November 2013 Rt Hon Damian Green MP, Edward Vaizey MP and Claire Perry MP
made, but not as much as one might think. I think
people think it is reasonable and unobtrusive. One of
the challenges was people then said, “Oh, but people
will know if I have said I don’t want the filters”,
which is sort of mad, because in a way they know
now. The ISPs know now because you do not request
to download them. There has been this sort of air of
are you invading people’s privacy. Well, of course
Google knows every single keystroke and every single
search that everybody is conducting, so it is slightly
bizarre that people are prepared to give up their
anonymity to private companies who sell it and are
concerned that somehow they will end up on some
sort of list. There is no list, there is no Government
mandate of this, this is the companies basically
interacting with their customers and saying, “Would
you like to download the filters? Yes, please or no,
thanks”. I think it is a very valid point.
By the way, the tone of this has always been about
among consenting adults not an issue, this is
absolutely fine. Although of course with closing the
rape pornography loophole there are some rather
unpleasant areas of currently legal material that the
Government rightly has decided should be illegal.
Again, it does not seem to be problem so far, and I
think we need to see with implementation and
Ofcom’s feedback if people do think it is
unreasonable, because it may be companies have to
revisit how they market their products.
Q222 Chair: Can I just pursue this point, which may
be best for Damian, of material that is just about legal
for adults, so it is pretty hardcore but it is legal. It was
pointed out to us that a successful prosecution had
been brought on the basis that the material could be
viewed by children. They did not have to prove that
it had been, just the fact that children could view it
had allowed a successful prosecution. Is that a sort of
route you think could be used again?
Damian Green: As you say, it is extremely unusual
and Claire has made the point about pornography
depicting rape, which I think one can regard in the
same way as hate speech if you like, speech that will
incite violence or something like that, because that
might incite criminal behaviour. The issue that I know
DCMS have been dealing with is pornographic
content that is entirely legal and that nobody would
be disturbed, or most people would not be disturbed
with adults viewing. But it is age inappropriate and
the ability to stop normally young boys repeatedly and
incessantly watching that kind of material is both a
technical issue and a policy issue of what you should
stop and where you draw the lines. It is not criminal
material so it does not fall under the criminal law, but
it is clearly age inappropriate for normally young
boys.
Chair: I think that was the point that was put to us,
which was that it could be deemed criminal under the
Obscene Publications Act if insufficient safeguards
have not been put in place to prevent young people
accessing it.
Mr Vaizey: It is an interesting issue. Obviously these
are the points put forward by ATVOD, the Association
for Television on Demand, and I think they have made
some significant progress on this. As was indicated
earlier by Claire, we are moving to outlaw R18
material. At the moment ATVOD can step in where
sites are offering R18 material, but we want to remove
any ambiguity because there is some discussion about
whether one has to show that there has been harm,
and obviously rape porn as well.
The question arises about the Obscene Publications
Act and whether or not a website that effectively
doesn’t have any mechanism for identifying the age
of the user and no kind of pay wall that you have to
get behind to access the material breaches the
Obscene Publications Act. I am not sure it is for me
to say in terms of whether sufficient prosecutions have
been conducted, because that would obviously be
something for the CPS. I do know that ATVOD, in
a rather imaginative way, is looking at working with
financial services providers and to say to them—
because obviously a lot of these websites offer you
free access and then hope to make you pay later—that
they should not provide payment facilities for sites
that do not offer robust age verification. That is issue
number one.
Obviously nothing should stop us doing the right
thing, in terms of prosecutions or clarifying the law.
Nevertheless we do have to be aware that a lot of
these sites do not provide suitable identification as to
who owns them, and again ATVOD is suggesting that
we clamp down on those sites by denying them their
financial support. That is something we need to
continue to work closely with ATVOD on, they do
have a serious point. We are also in UKCCIS—the
United Kingdom Council for Child Internet Safety—
looking at robust age verification measures, technical
measures, whether it is the unique pupil identifier that
is used in schools to enable children to get access to
the web, or the digital wallet that I think is a line
pursued in Demark to ensure that you can identify the
age of someone by effectively giving them a digital
wallet to use on the internet. I also think, and ATVOD
say it in their evidence submitted to you, that there is
a role for the European Union—sorry to keep coming
back to Europe—of working across all these
territories, because again one of the issues here is
what is acceptable in this country or not acceptable in
this country, in terms of culture, if you like, is
acceptable in other countries across the EU. It is
something we need to continue to look at and I think
it is an important issue.
Chair: Thank you. I think we have exhausted the
questions, thank you very much.
cobber Pack: U PL: CWE1 [SO]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 65
Written evidence
Written evidence submitted by the Children’s Charities’ Coalition on Internet Safety
Thank you
—
—
—
for giving us the opportunity to submit evidence to your Committee’s enquiry into:
How best to protect minors from accessing adult content.
Filtering out.....images of child abuse.....
Preventing abusive or threatening comments on social media.
Our fuller views are set out in the following pages and summarised as follows:
1. Protecting children from accessing adult content on the internet can be done most effectively through
educating and empowering children to look after themselves though with very young children there are obvious
limits. Filtering programmes have an important supplemental role to play in supporting good practice.
2. Every internet enabled device or internet based service sold or supplied into the consumer market and
likely to be owned or used by children and young people should, by default, come with protective software
preinstalled and operational.
3. Law enforcement should step up its efforts to enforce the decision in R v Perrin.
4. Nominet should make compliance with R v Perrin a condition of registration. UK-based web hosting
companies should be required to ensure pornography sites serving the UK have age verification.
5. CHIS supports ATVOD’s call for banks and credit card companies to refuse to process payments to sites
that do not comply with R v Perrin.
6. The BBFC should establish transparent industry standards in relation to blocking online adult content.
7. The scale on which child abuse images are now circulating on the internet, and the number of people
involved in downloading or exchanging them, have outstripped the capacity of law enforcement to deal with
them adequately. It is vitally important that the internet industry is enlisted to help find new or better technical
solutions to combat the problem.
8. There needs to be a specific push on eliminating or reducing the number of child abuse images circulating
on Peer2Peer networks and we need to invest more in victim identification work.
9. Social media sites where significant numbers of children and young people congregate should be expected
to employ an appropriate number of human moderators and to deploy sophisticated software which can help
moderators spot and stop bullying and grooming.
10. There is a strong case for an major enquiry into the use of anonymity or unverified identities in social
media where children and young people are major users.
How Best to Protect Minors from Accessing Adult Content
1. With the spread of WiFi and laptops, tablets or other mobile devices which are internet enabled it is
wholly unrealistic to expect parents, teachers and carers to be able to supervise everything their children do
when they go online.
2. Protecting children from accessing adult content on the internet can be done most effectively through
educating and empowering children to look after themselves although with very young children there are
obvious limitations to this approach. Filtering and blocking programmes can play an important supplemental
role underpinning parental guidance and good practice. Such software may be especially useful supporting
younger or particular groups of vulnerable children.
3. Every internet enabled device or internet based service sold or supplied into the consumer market and
likely to be owned or used by children or young people should, by default, come with filtering and blocking
software preinstalled and operational to provide protection against exposure to adult content. An age-verified
adult ought to be able to modify the preinstalled protective programmes’ settings or abandon them altogether.
4. Filtering and blocking software is still far from perfect but it continues to improve. However, any and all
technically-based safety measures should only ever be seen as an adjunct to and not as a replacement for
educational and awareness initiatives. Parents and children need to understand the nature of potential online
hazards and appreciate both what safety software can do and what its limitations are.
5. The decision in the case of R v Perrin is honoured more in the breach than in the observance although it
is recognised that most if not all of the offending web sites are owned by publishers based overseas.
6. The police should be more vigorous in applying R v Perrin, perhaps seeking to extradite overseas web
site owners who make no attempt to shield minors from adult content.
7. Nominet should make compliance with R v Perrin a condition of operating a .uk domain name eg if a
site is to publish pornography the operator must give a binding undertaking to put an effective age verification
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 66 Culture, Media and Sport Committee: Evidence
process in place before the site goes live or within a reasonable timeframe. It should be noted that the UK’s
online gambling industry has used age verification with great success.
8. UK-based web hosting companies should ensure publishers making pornography available within the UK
have an effective age verification process in place.
9. CHIS supports ATVOD’s proposal for banks and credit card companies to refuse to process payments to
any pornography sites that do not have an effective age verification process in place.
10. CHIS congratulates the UK’s mobile phone networks for sustaining their policy of, by default, putting
adult content behind a bar which can only be lifted by the user completing an age verification process. CHIS
welcomes the recent engagement of the BBFC to oversee the administration of this scheme including setting
standards governing which content should go behind the adult bar.
11. CHIS also commends the UK’s largest WiFi companies for deciding that, when they are asked to provide
WiFi access in a public space where children and young people will normally be present, by default they will
put pornographic content behind an (immovable) adult bar. Several of the WiFi companies have already
implemented this decision. CHIS calls on the remainder to make a statement making clear when they will have
done the same. Smaller WiFi suppliers should follow a similar path within a reasonable timeframe.
12. The BBFC should be encouraged to develop a kitemark scheme and associated standards in respect of
public WiFi and WiFi providers ought to adopt and advertise their compliance with it.
13. A child should not be prevented from accessing certain types of adult content while they are using their
mobile phone company’s network only to find they are able to access identical material via the same device
simply by switching to WiFi. There needs to be a high degree of consistency as between the standards set by
mobile operators and those being applied by WiFi providers. The BBFC would be well placed to help establish
such consistency and also ensure transparency in relation to the content standards and processes being used.
14. CHIS welcomes the announcement by the UK’s major ISPs of their intention to upgrade the level of
protection against adult content offered to new customers and their support for the “one click” approach. The
ISPs have pledged to have their new offerings in place by the end of 2013. Existing customers will be put in
an equivalent position by the end of 2014.
15. Since none of the ISPs have yet disclosed what their final offerings will be CHIS does not propose to
comment further at this stage other than to say CHIS believes ISPs should, as closely as possible, implement
a system similar to that which exists on the mobile networks.
16. The logic of this approach points towards the need for individual accounts for each household member.
By default adult content would therefore be inaccessible to the whole household and remain inaccessible unless
and until a responsible adult has authorised a change, account by account. The worry otherwise is that in
households with people of widely differing ages it will prove unworkable for everyone’s internet access to be
configured to be suitable only for a child. There are routers on the market which have been built precisely to
allow for this type of arrangement. Alternatively it could be achieved on the network.
17. The BBFC once more could play a useful role in helping ISPs roll out a solution while providing
consistency with other platforms and transparency as to the processes.
18. In relation to adult content not directly accessed from the internet but obtained in other ways eg via
Bluetooth, USB sticks, memory card exchanges, emails, disc swaps or downloads CHIS looks to the wider
deployment of technical tools to deter or deflect such activities and thereby help protect minors from age
inappropriate content.
Filtering out.....Images of Child Abuse.....
19. The UK has done extremely well in more or less eliminating the hosting of child abuse images on UKbased web servers. The deployment of the IWF’s url blocking list has also been important in limiting web
access to that kind of material. Images found in Newsgroups are swiftly dealt with. However, whilst it is
important to retain a strong focus on the web and Newsgroups, technology has moved on and we are now a
long way from coping with more modern manifestations of the problem.
20. In 2012 the NSPCC issued FOI requests to every local police force in England and Wales asking them
to state how many child abuse images they had seized in arrests made in the two years ending April, 2012.
Within the NSPCC’s timeframe only five forces replied but it emerged that between them they had seized over
26 million. On one calculation that would imply that over 300 million illegal images may have been seized by
all forces over the same period. Numbers like these are unprecedented and while numbers do not by any means
tell the full story, they most certainly tell a story.
21. On ITN News on 28th May, 2013, Peter Davies, the Head of CEOP, acknowledged that the UK police
had identified between 50,000 and 60,000 individuals who appeared to have been exchanging or downloading
child abuse images, principally over Peer2Peer networks. Davies said the police do not have the capacity to
arrest all of these people although he said “I wish we could”.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 67
22. In no year since records began, including at the height of Operation Ore, has UK policing arrested more
than 2,500 individuals for child abuse image related offences. Thus, even if there were no new offences from
now on, and assuming the maximum rate of arrests was sustained year on year, conservatively the last person
off the current list would not be picked up before 2032. This has worrying implications both for the abused
children depicted in the images and children who may yet become victims of individuals whom the police
have identified as being engaged in downloading.
23. The technology has outstripped the current capacity of UK law enforcement to cope with the volumes
of images in circulation and with the numbers of offenders involved in downloading or distributing them. Most
police forces around the world are in the same position.
24. However, even if the UK was living through times of super abundance, as opposed to times of austerity,
it is hard to imagine how we would ever be able to manage criminal behaviour on the sort of scale indicated.
Society therefore has a stark choice. Either we settle back and accept that substantial numbers of people living
among us are routinely accessing child abuse images, that it has become, so to speak, part of the background
music of 21st Century Britain, or we look for new and better ways to enlist the internet industry’s support in
finding technical measures to address the problem. CHIS does not think anyone in a position of responsibility
is ready to go with the first option. We strongly favour the second.
25. There is no single measure which will get rid of child abuse images from the internet. A range of tactics
are needed. CHIS puts forward the following for consideration:
(a) Greater use of splash pages and warning messages to deter a certain class of person with a low
level, opportunist or early interest in child abuse images.
(b) Greater use of algorithms to prevent search engines being used to locate child abuse material
or locate information helpful to paedophiles.
(c) Greater use of tools capable of comparing hashes of known illegal images with images in
remote storage locations and, wherever possible, in transit.
(d) The development of botnets or other crawler technologies capable of locating images or sites
of possible interest which have not yet been reported.
(e) Establish a national initiative to give a specific focus to eliminating or at any rate hugely
reducing the volume of images being exchanged over Peer2Peer networks and increasing the
numbers of individuals arrested for this type of activity.
(f) British companies should strengthen their engagement with measures designed to address the
traffic in child abuse images eg by stepping up the work they do with their employees.
(g) Ask British policing to strengthen its engagement with victim identification work and
investigate if there is a case for helping to establish a strong internationally based victim
identification resource.
h. Ask British policing to construct a national database of images which will be used by all UK
forces and will also integrate into Interpol’s international initiative.
(i) Mount a campaign to heighten awareness of the harms associated with this type of offending
and inform people how to report online child abuse images.
(j) Broaden participation in the UK-US Taskforce announced by the Prime Minister on 22nd July.
Preventing Abusive or Threatening Comments on Social Media
26. Clearly this is an area where educating people about the importance of behaving in a civilized and
responsible way when using social media, and explaining the potential consequences of not doing so, will have
an important part to play in combatting some of the worst excesses which have attracted the media’s attention
in the recent past.
27. Peer-based support networks which develop a sense of social solidarity among the users of social media,
which encourage people to intervene to support someone being bullied or victimized and to bring an end to
another person’s bad behaviour are the sorts of initiatives which all social media sites should support.
28. However, unless social media services decide to pre-moderate every post, be it of images or text, it is
difficult to imagine how they could ever “prevent” abusive or threatening comments being made.
29. Many online companies, including some small or niche social media sites, do pre-moderate everything
or almost everything that goes up. They do so for a range of reasons at least one of which is a concern for
their own reputation but also they are keen to minimise any potential legal liability for libel. In some instances
a concern to protect younger users from possibly harmful self-disclosures has been a motivation for using premoderation. There may be some situations where pre-moderation is essential eg on services specifically directed
at young or vulnerable children.
30. That said, the scale on which sites like Twitter and Facebook operate probably renders pre-moderation
impracticable even if it was thought desirable. However, it is a myth to assume that all pre-moderation systems
inevitably slow down chat or interactive environments to a point where it is impossible to maintain a sense of
real time or swift action.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 68 Culture, Media and Sport Committee: Evidence
31. Nonetheless technological solutions are available which can analyse text streams and help identify “hot
spots” connected with bullying or other types of abusive behaviour including grooming, some of which may
lead on to the creation of child abuse images or sexual assaults of a different kind. The software ought to flag
up potential problem areas to human moderators who should be working 24/7 and employed in sufficient
numbers to be able to intervene rapidly if necessary. Measures of this type should be in place from the start of
a site’s operations. Companies should not wait for a tragedy before doing the right thing. Someone in the
company should sign off confirming that full consideration has been given to all child safety aspects before
any new online product or service is launched, especially on to the so-called “free” internet where it is known
children and young people will have ready access.
32. There seems little doubt that the ability to hide behind an apparent cloak of anonymity, in particular the
ability to manufacture an entirely bogus or opaque online identity which is then used on social media sites,
lies at the root of much of the problem, even on sites which ostensibly specify a “real names” policy.
33. In principle CHIS has no problem with individuals signing in as “Donald Duck III” or “Diana Dors”.
There may be many situations where not using your real name will be positively helpful or beneficial. What
matters is traceability. In an environment where everyone’s real world identify had been robustly verified the
log in they used would be less important but one would expect people’s behaviour to improve because they
would know that if they crossed particular lines police or the civil courts would be able to identify and locate
them extremely quickly if required.
34. CHIS accepts that the implications of the view expressed here are major and radical. For that reason
CHIS would like separate and specific detailed consideration to be given to the issue of online anonymity,
perhaps focusing specifically on social media sites which are known to be particularly popular with children
and young people.
35. CHIS has no desire to make it harder for whistleblowers to continue to perform an important public
service. Nor does CHIS wish to require political dissidents or persons dealing with sensitive issues to disclose
their true identities before they can log on to any online service.
36. However, equally, CHIS finds it difficult to accept this is a zero sum game where advances in online
child protection are forever seen as being made at the price of imperilling political dissent, whistleblowing or
the position of others with a genuine need for anonymity.
37. The internet is now many different things to many different people. Perhaps it is simply expecting too
much for all of it to be governed by a single perspective, a single set of principles or priorities. In other words
not all social media sites or online services need to be governed by the same rules. Perhaps those that are used
by large numbers of children and young people could reasonably be expected to conform to different standards.
September 2013
Written evidence submitted by the NSPCC
Introduction and Summary
1. The NSPCC applauds the Culture, Media and Sport (CMS) Select Committee for launching an inquiry
into Online Safety and we welcome the opportunity to respond. The NSPCC recognises the internet has brought
considerable benefits to society. However, alongside these benefits, the popular use of the internet has brought
unique challenges to young people and can present some significant risks to the safety of children.
2. In 2011–12 ChildLine dealt with 3,745 counselling sessions where the young person mentioned either a
mobile phone or internet issue.1 We also know from the ChildLine messaging board that negative experiences
online can have a harmful impact on young people, from affecting a child’s school work to self-harm. In some
of the most extreme cases, cyber-bullying in particular has resulted in young people taking their own lives.
3. Additional to these challenges, the internet has provided some adults with a new vehicle for sexually
abusing children and young people—most notably, the use of some social media sites and platforms as a tool
for sexual grooming and the extremely worrying proliferation of child abuse images.
4. The NSPCC believes children have a right to an abuse free experience online and that Government,
industry, the voluntary sector and the public all have a role to play in achieving this aim. As such, this
submission responds to the three topics set out by the CMS Select Committee with a focus on how the internet
is used by and could be made safer for children and young people. Most notably, this document includes
recommendations for:
Age-Inappropriate Material:
— All Internet Service Providers to create simple ways for all internet customers to regularly make
clear and meaningful choices around content controls for internet-enabled devices in the home,
and for the effectiveness of these filters to be evaluated frequently.
1
ChildLine “online issues” statistic 2011/12
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 69
—
Digital judgement and e-safety to be a component of sex and relationship education that is a
compulsory part of the national curriculum.
Child abuse images:
— Industry to act swiftly to find and put into implementation a technological solution to block
more searches for child abuse images.
— Greater police prioritisation and resources to bring offenders to justice.
— International collaboration in combatting the making, viewing and sharing of abhorrent child
abuse images across the internet.
Harmful contact online:
— A dual response from the education system and the internet industry to prevent and respond to
harmful online behaviour like cyber bulling and online sexual grooming.
Topic One: How to Best Protect Minors From Accessing Adult Content
5. The NSPCC would like to encourage the CMS Committee to consider how best to ensure children and
young people are accessing content that is appropriate for their developmental age rather than simply looking
at this challenge as a binary under 18/over 18 problem.
6. For instance, research the NSPCC is about to launch shows that while many of the biggest social
networking sites, such as Facebook, Twitter and YouTube, require users to be over the age of 13 to have a
profile, large numbers of children under the age of 13 are active users of these sites and, therefore, potentially
exposed to content that is not appropriate for their developmental age.
What we know from ChildLine
7. During 2011–12, there were 641 counselling sessions where the young person specifically mentioned
being exposed to sexually indecent images.2 While these incidents will not exclusively relate to online
content, a large proportion of this sexually explicit material will have been accessed through internet enabled
devices. Young people often told ChildLine that they felt guilty and disgusted about what they had seen and
were extremely worried about getting into trouble for accessing these sites. ChildLine has also seen a growing
trend of young people talking about being addicted to online pornography.
Filters and parental controls
8. In June 2011 the government published the Bailey Review of the Commercialisation and Sexualisation of
Childhood. This Review recommended that it should be made easier for parents to block adult and agerestricted material from the internet. In response, the UK’s four biggest internet service providers (ISPs)
announced a new voluntary code of practice and the adoption of “active choice” systems for new customers
which was rolled out in October 2012. The NSPCC was broadly supportive of this initiative, but at the same
time felt that this progress did not represent a panacea, especially as the companies’ existing customers would
not be encouraged to make a choice about filters and parental controls.
9. Following an independent Parliamentary inquiry into “Online Child Protection” led by Claire Perry MP
and a Government consultation on parental controls, both in 2012, most recently the Prime Minister came to
the NSPCC main London office to announce a package of measures to make the internet safer for children and
young people. One key call of the Prime Minister’s speech was for Internet Service Providers (ISPs) to ensure
that not only new but also existing customers are asked to make a choice about applying content filters, with
the “filters on” button pre-checked (“active choice plus”) by the end of 2014. The NSPCC welcomes this
proposal and that a clear deadline for delivery has been set.
10. The ISPs are expected to report back to the Government at the end of October on progress towards this
goal. The NSPCC considers it is important that the major service providers are developing uncomplicated and
meaningful mechanisms for ensuring customers are making active and informed choices about the filters they
apply in their homes. We are, however, keen to see that after the initial decision is made by the customer about
the filter package they want in place, the ISPs will ask them to reconsider their choices on a regular basis and
that the effectiveness of the filters is evaluated frequently. We believe this would help to ensure the filters
remain relevant for children’s digital needs and habits over a sustained period of time. We also want to see a
commitment from the smaller ISPs to implementation of similar processes to ensure their customers have
access to the same level of meaningful choice
Education
11. We believe building children’s e-literacy and digital judgement is essential for helping children and
young people to understand the challenges they may face when accessing content that is inappropriate for their
developmental age and encourage them to make an active choice about avoiding material that might upset
2
Ibid
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 70 Culture, Media and Sport Committee: Evidence
them. The responsibility for this falls on both parents and the schools. However, the NSPCC understands that
not all parents, carers and teachers feel equipped to deliver these messages. As such, we believe there is a role
for Government to encourage the delivery of this education by:
— Making e-safety an aspect of sex and relationship education which should be a compulsory part
of the national curriculum. Teachers should be given suitable training and teaching aids to
deliver these lessons.
— Supporting the launch of a national awareness raising campaign aimed at parents to help them
understand the risks the internet poses in relation to their children accessing material online
that is inappropriate for their developmental age, and to direct them to information to support
them to have conversations with their children about this content.
12. We are disappointed that the Government does not plan to update the sex and relationship guidance for
schools, which remains unchanged since 2001. We support the Daily Telegraph campaign asking Government
to reconsider this position.
Topic Two: Filtering out Extremist Material, Including Images of Child Abuse and Material
Intended to Promote Terrorism or Other Acts of Violence
13. The NSPCC would like to challenge the CMS Committee’s use of the word “filtering” in the context of
tackling the making, sharing and viewing of child abuse images online. Filters are tools which should be
applied to manage access to content which is legal but possibly not appropriate for younger audiences, not as
mechanisms for hiding content which is wholly illegal and should not exist on the internet. Instead we believe
the Committee should be talking about “combatting” and “eradicating” child abuse images from the internet
as there should be no place on the web for this abhorrent material.
An overview of child abuse images and prevalence
14. Child abuse images are a visual record of the sexual abuse of a child. They can include pseudophotographs, animations, drawings, tracings, videos and films which are being streamed live. In the UK images
are graded on a 1–5 scale. Level 5 images involve sadism or bestiality, Level 4 will portray a child engaged
in penetrative sexual activity and so on to Level 1, where the images will depict erotic posing with no visible
sexual activity. Please note that the sentencing council is currently reconsidering the definition of level 3.
15. The number of separate child abuse images in circulation is unknown. In 2010, an NSPCC Freedom of
Information request revealed that nearly 26 million child sexual abuse images had been confiscated in the
preceding two years from just five of the 43 police forces in England and Wales which were able to check
their records. Most forces said it would take too long to interrogate files to see how many pictures of children
being sexually abused they had accumulated during investigations. However, estimates of the number of online
child abuse images confiscated by police in England and Wales each year in the UK range from 100–360
million.3
16. In the past producers of child abuse images were limited in the extent to which they could obtain and
disseminate images. However, advances in technology, including the popularisation of the internet has made
child abuse images available to a far greater number of people than ever before. Now, once an image has been
uploaded to the internet it may be replicated and downloaded an unlimited number of times. The knowledge
that images of abuse cannot ever be fully deleted has a profound effect on the victims involved and this may
follow them into their adult lives. Victims may also become aware that the material involved can be used to
groom and to abuse other children. Perpetrators of sexual abuse have also been known to use sex abuse images
as part of the grooming process, in order to desensitise children and young people to the idea of sexual contact.
A proactive response to the challenge
17. The NSPCC was pleased when, in June 2013, it was announced that for the first time the Internet Watch
Foundation (IWF) will be working with Child Exploitation and Online Protection Centre (CEOP) to proactively
seek and block online child sexual abuse images. We hope this will help to significantly reduce the number of
images available on the web. The commitment shown by the UK’s leading internet providers, who donated to
the project £1 million of extra funding and a signed a “zero tolerance” pledge to help tackle the creation and
distribution of child sexual abuse images, was also welcome.
18. While we are pleased that more resource and capacity is being dedicated to the search and take down of
child abuse images, the NSPCC does have concerns that there is currently insufficient police resourcing to
proactively identify, arrest and seek prosecution of people who are making, sharing and viewing child abuse
images. We believe, however, that there is a unique opportunity with the launch of the National Crime Agency
(NCA), within which CEOP will now sit, for the tackling of child abuse images to become a key UK
policing priority
19. The NSPCC understands Her Majesty’s Inspectorate of Constabulary (HMIC) intends to conduct a
programme of Child Abuse related inspections in 2013–14, the first of which will focus on the police service’s
3
NSPCC, https://www.nspcc.org.uk/Inform/research/questions/people_convicted_of_child_abuse_images_2010_wda83579.html
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 71
response to child sexual exploitation facilitated or caused by use of the internet. We hope that the inspection
will address the capacity of forces to investigate crimes relating to child abuse images. More specifically, it is
vital that the NCA and local forces have both the necessary expertise and sufficient resources to find and bring
to justice people who make, share or view child abuse images.
Deterring potential perpetrators
20. The NSPCC is pleased that the major search engines are implementing “splash pages” warning internet
users of the crime they are trying to commit and its implications when they try to view a page already know
to the IWF to contain illegal or indecent images. We are also pleased to understand that there is commitment
from some part of the “search industry” to also developing advice pages offering people clear routes to
legitimate sites on the web when terms are used that could be interpreted as a potential searches for images of
child abuse.
21. The Prime Minister’s online safety speech in July also included a challenge to search companies to take
one further step and deliver “no returns” in certain instances were searchers are unequivocally looking for child
abuses images. The NSPCC fully supports the Prime Minister’s call to action and wants search companies to
pledge to deliver swiftly meaningful and effective technological solutions to this challenge at the Autumn
summit on online safety the Government has committed to host. We also call on the search companies and
Government to agree together a robust monitoring process to scrutinise the efficacy of the solution and an
approach to tracking and making public progress achieved through these technical solutions.
22. In addition, we welcome the resource the Government has committed to creating a new National Database
of Child Abuse Images to work as a central repository for online child abuse images that have been identified
and confiscated. We are keen that the Government now sets out a clear timeline for delivery of the database
and make public more information about how it will be used in the fight against child abuse images.
23. Lastly, we ask the Government to explore the potential of creating a public awareness campaign aimed
at people who might consider making, viewing or sharing child abuse images informing them on the possible
consequences of their actions and directing them to support services, like Stop It Now!
The international dimension
24. The UK Government has shown leadership in the battle against child abuse images. However, the
NSPCC, like others, regards the issue to be a global problem. We believe to make significant inroads to tackling
the dissemination of child abuse images online, greater international cooperation and activity is required with
more countries buying-in to the need for effective search and take down mechanisms. We are pleased that a
UK/US taskforce has been established to look at the international dimensions of this challenge and we urge
the taskforce to look into the feasibility of:
— The new UK National Database of Child Abuse Images playing a role as part of a bigger global
network of image databases and from this network a global process for proactively searching
for, reporting and removing child abuse images being established
— Greater prioritisation of technological solutions to disrupt peer-to-peer networks and sharing
across the darknet
— Relevant internet companies agreeing to an accord of information sharing in the name of
disseminating best practice in the field of search and take down technology
Topic three: Preventing Abusive or Threatening Comments on Social Media
25. As well as issues around the content that is available on the internet, either inappropriate or illegal, a
significant challenge young people face is learning how to behave in a responsible and socially acceptable way
while online and knowing how to respond to unwanted contact or engagement via the internet.
26. The main issues about which young people talked to ChildLine counsellors when discussing online
safety were cyber-bullying, online sexual abuse, sexting and social networking issues. The NSPCC has found
that almost a quarter (23%) of 11 and 12 year olds who have a profile on a social networking site say that they
have been upset by something on it over the last year. We respond to this section of the inquiry by focusing
on two distinct aspects of abusive and threatening behaviour on social media that affects children and young
people: a) peer to peer cyber bullying and b) online sexual grooming. We then explore the policy solutions
that could be used to tackle both online dangers.
Cyber bullying—the extent and impact
27. The ChildLine website describes online bullying to young people as: “when a person or a group of
people uses the internet, mobile phones, online games or any other kind of digital technology to threaten, tease,
upset or humiliate someone else…” The perpetrator can be known to the victim or they can make themselves
anonymous by setting up fake accounts or by using public computers so that their own personal IP address
can’t be traced.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 72 Culture, Media and Sport Committee: Evidence
28. It can happen at all times of the day and inside the young person’s home or somewhere they may have
previously felt safe. As a result the young person may feel trapped and that they are unable to escape.
Traditionally guidance given to young people on how to handle bullying would be “to walk away”; however
this advice has now become redundant. In a world dominated with technology a young person is no longer
able to walk away from the situation.4
29. In 2012–13 Childline saw an 87% increase in counselling session about online bullying. We’ve also seen
from the ChildLine messaging board just how significant an impact this experience can have on young people.
“she gets all her friends against me…every time I look at my messages it’s something about
the way I look or my personality- swearing at me and telling me to kill myself. I’m tempted to
commit suicide”. ChildLine message board user
“The lad who bullies me has made a new Facebook account and was spreading rumours about
me which everyone believes. Everyone began writing status’s about me/calling me ‘twisted’
‘sick in the head’ ‘waste of space’. I was even told to kill myself…I’m scared to go to school
tomorrow. I’ve been crying all weekend…I don’t know what to do”. ChildLine message board
user
Online grooming—the extent and impact
30. Grooming can be defined as actions that deliberately establish an emotional connection and trust with a
child or young person, with the aim of engaging them in sexual behaviour or exploitation.5 Online grooming
is exactly the same, but done through the internet, and using technology such as smart phones, laptops, tablets
and computer game consoles. The exact numbers of children who are subjected to online grooming is unknown.
However, we do know that in 2011–12 The NSPCC’s ChildLine service had 413 contacts from children where
the primary concern was grooming. Sixty% of these specifically mentioned online grooming.6
31. Many young people will not disclose the offences against them online because they feel shame and guilt
about the way they have been made to feel, and mistakenly, that they are somehow responsible for their own
abuse, for example when they have posted explicit images or had conversations online that can never be
erased.7 While other victims deny the abuse happened or are unaware that they were being abused, thinking
that they were chatting or meeting up with their “boyfriend”.
A solution—education
32. As mentioned above, online bullying and online sexual exploitation are very distinct and separate
challenges. However, education has a role to play in helping to prevent both problems.
33. In the case of online grooming children and young people need to be taught about how to be resilient—
keeping personal information safe, blocking unwanted sexual approaches, talking to a trusted adult about
contact they have received online that they are uncomfortable with. While with online bullying, children need
to understand more about pro-social and respectful behaviour both in their face-to-face contact with their peers,
but also in their virtual interactions.
34. Re-emphasising what we called for previously in this submission, the NSPCC wants to see sex and
relationship education being taught as a compulsory part of the national curriculum and for online safety
featuring the components listed above to be included as a key part of these lessons.
A solution—monitoring and reporting
35. The NSPCC also believes social networking sites and others who provide online services and products
to children have a corporate social responsibility/duty of care for protecting children and young people from
harmful contact with both peers and adult sexual predators. More specifically we would like to see:
4
5
6
7
—
Providers should ensure that there are a range of options for how site users report illegal,
offensive or abusive content, and these are easy to find and use
—
Social networking sites in the UK to provide a simple route for younger users to seek support
and help via ChildLine.
—
Default privacy settings should be set to the highest levels possible and any avenues for
anonymous contact set to “off” for children under 18.
Digizen (2013) “Understanding Cyberbullying”, available from:
http://www.digizen.org/resources/cyberbullying/understanding/how-is.aspx [Accessed on 27 August 2013]
Turnbull, M (2012). Caught in a Trap: Impact of Grooming in 2012. London: ChildLine
Ibid
Palmer, von Weller and Loof. (2010) The impact of internet abuse on children and how best to intervene; The Link—The
Official Newsletter of the International Society for the Prevention of Child Abuse in Neglect (ISPCAN) Colorado USA
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 73
—
Adequate use of human moderators on social networking sites so that appropriate and timely
action can be taken in response to a) breaches in the sites’ codes of conduct b) abusive or
threatening behaviour c) potential illegal activity.
September 2013
Written evidence submitted by BeatBullying
1. Introduction to BeatBullying
1.1 BeatBullying is the UK’s leading bullying prevention charity, creating a world where bullying, violence
and harassment are unacceptable. BeatBullying empowers people to understand, recognise, and say no to
bullying and cyberbullying, violence and harassment, by giving them the tools to transform their lives and the
lives of their peers.
1.2 As an organisation with children’s and young people’s safety (both online and offline) at its heart,
BeatBullying’s online programmes have embedded user protection mechanisms at several levels from the
beginning. Users have access to all the functions of a social networking site—email-style private messaging,
instant messaging chat in real time, a customisable profile page, the chance to upload content including blogs,
and a chatroom—but with the safety and wellbeing of its users built in as a core priority across every aspect.
The BeatBullying site is fully moderated for language and content both by software and by human moderators;
all content is checked and approved before being uploaded; the site has clearly stated rules for safety and
behaviour which users are expected to follow, with a system of sanctions for breaking them; staff and
moderators can eject, ban and block any user whose behaviour is unacceptable; there are extensive reporting
systems in place both for users to report issues such as bullying by others on the site and for staff to share
information, including the reporting of users who appear to be adults. BeatBullying.org is the only e-mentoring
and social networking site to be endorsed by CEOP. We strongly believe that our approach to online safety
must be adopted by all internet providers if children and young people are to be safe online.
2 Summary
2.1 BeatBullying’s online programmes all utilize a combination of approaches to keeping children and young
people safe online, of which the technical safeguards outlined above are a fundamental part. However, these
cannot be considered in isolation from the other aspects of our work—education, support, attitude and
behaviour change, and empowerment—which are just as fundamental.
2.2 BeatBullying is completely in agreement with the principle of preselecting to block genuinely harmful
content such as pornography but could not endorse any system that might deny access to legitimate information,
advice and support around sensitive issues due to over-zealous and inadequately nuanced filtering.
2.3 The most useful “parental control” is parental responsibility and, inseparable from this, the commitment
to developing their children’s own sense of responsibility as they grow up.
2.4 Blocks and filters can only block and filter; they cannot educate children about how to deal with the
kind of content and behaviour they might encounter online, how to manage their own behaviour responsibly
or why certain material is harmful.
2.5 Everyone involved with children’s and young people’s use of the internet—parents, schools, service
providers, organisations and children themselves—has a shared responsibility for online safety. That is why in
April 2013 BeatBullying launched a campaign for better anti-bullying protections called Ayden’s Law.8 The
campaign calls for a national strategy to tackle cyberbullying and would set out how the voluntary and
community sector, parents and schools would be equipped to (a) protect the children in their care from harm
online and (b) educate and equip children about internet safety and responsible digital citizenship so that they
understand the issues for themselves.
2.6 Any approach to online safety must ultimately be about shaping attitudes and changing behaviors as
much as it is about teaching techniques for staying safe or for anything else.
3 How Best to Protect Minors from Accessing Adult Content
3.1Findings from our recent survey of children and parents9 show that the majority of parents and children
(78% and 48%) agree that parents should mainly be responsible for controlling what children access online
with two thirds (66%) of parents feeling confident enough to advise their children where to go online.
3.2 Whilst we understand that parents have a responsibility to protect minors from accessing adult content
online, we would argue that parental controls are only part of the wider issue of responsibility. The most useful
“parental control” is parental responsibility and, inseparable from this, the commitment to developing their
children’s own sense of responsibility as they grow up. It is imperative that children and young people are
8
9
Launched May 2013 with The Sun Newspaper following the tragic death of Ayden Olsen Aged 14
The Online Behavioural Habits of Young People Today, which questioned a 1000 8–16-year-olds and 1500 parents
from Parentdish and BeatBullying, July 2013
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 74 Culture, Media and Sport Committee: Evidence
encouraged to take appropriate responsibility for themselves and their own behaviour increasingly, so as they
grow up but beginning in small ways at an early age. In this way, while parents and other adults will still be
highly involved, children can be supported to grow up with the idea of responsibility as something integral to,
rather than entirely outside of, themselves. Our peer mentoring system works on the premise that children and
young people can and do rise to the challenge of appropriate responsibility, properly supported at every stage
by older peers and trained adults. Of course online safety cannot be the responsibility of children in isolation,
but neither can it be the responsibility of parents alone.
3.4 Over half (57%) of the 11—18 year old respondents surveyed in 201210 told us their parents talk to
them about online safety and what they should or should not look at on the internet as a way to keep them
safe online. As the largest category of response this indicates that among our respondents a majority of parents
are actively engaged in their children’s use of online technology. Over a quarter (27%) say that their parents
do not use any methods (technological or otherwise) of keeping them safe online and a fifth (20%) say that
they use technological filters. Almost the same proportion (17%) report that their parents encourage them to
use websites about internet safety, and 12% say that their parents sit with them while they are online and watch
what they do. Further written responses included “offering non-intrusive advice” and “trust”, which illustrate
an aspect of the online safety debate which is often overlooked: the relationship between parent and child and
its bearing on the development of trustworthiness. This reinforces the argument for increased education for
children/young people and parents, ideally helping them to develop a continued dialogue about the use of
the internet.
3.6 Under 18 is a broad category; while we agree that there is some material that no-one under this age
should be able to access, there are also many grey areas: some things that would be completely off-limits to a
primary-school child might be fine for a 14 year-old. Just as in the offline world, older teenagers are expected
to have much more independence than young children; what would be appropriate for a seven year-old and for
a 17 year-old to view or engage with online are going to be very different. CHIS (Children’s Charities Coalition
on Internet Safety) suggests the possibility of allowing different user accounts to be set up for users of different
ages in a single household, via routers that are built for this specific purpose; BeatBullying supports this, with
the caveat that age verification systems need to be more advanced in order to ensure that it works as it should.
4 Filtering out extremist material, including images of child abuse and material intended to promote
terrorism or other acts of violence
4.1 Leading on from parental responsibility, it is important to recognise that the technical safety features of
these sites are integral to the sites themselves. It does not matter where a user accesses them from, on what
device, or through which network, the safety features remain consistent.
4.2 BeatBullying is completely in agreement with the principle of blocking genuinely harmful content such
as pornography but could not endorse any system that might deny access to legitimate information, advice and
support around sensitive issues due to over-zealous and inadequately nuanced filtering. It would be valuable if
categories subject to filtering were broken down into more meaningful neutral, positive and negative subgroups, such as: “Eating disorders—general information” (neutral), “Eating disorders—advice and support for
prevention or recovery, such as the charity Beat” (positive) versus “Eating disorders—pro-anorexia (negative)”
and parents could make a separate decision in each of these distinct categories.
4.3 BeatBullying is in favour of the Government’s “pre-select to on” approach to filtering. This option has
the virtue of simplicity for parents, setting up blocks as the default and requiring a conscious and deliberate
opt-in for certain content. However, as explained above, we would welcome more detailed scrutiny of what is
or is not acceptable: categories such as “violence” or “anorexia/bulimia” in themselves tell us nothing.
4.4 Implicit in having certain things ready-blocked is a “social norm” of what is and is not acceptable. The
assumption inherent in pre-blocked content is that parents will not want their children to access pornography
etc., and the idea is that this acts as a prompt to conform to this expectation. While we believe that the majority
of parents are responsible, there is a minority who might lack either the capacity or the will to make informed
decisions in this area. That is why we take the view that having automatic blocks on some content, which can
be removed by adults if they choose, would be the best option to ensure a balance between guidance and
freedom, thereby protecting children more effectively.
4.5 It is important that in a time where technology has outstripped the capacity of law enforcement, filtering
and blocking is not just at the feet of parents; it is imperative that search engines and ISP’s actively challenge
themselves to work towards an advancing solution to an on-going problem. In July this year the Prime Minister
threatened legislation if internet providers failed to commit to a blacklist of CEOP search engine terms which
we support and hope that progress is swift.
4.6 Since 2008 the UK Council for Child Internet Safety has tried but failed to introduce meaningful selfregulation. BeatBullying would therefore introduce regulations on internet safety that would:
— Ensure clearer and simpler mechanisms for children to report cyberbullying or abuse.
— Agree protocols in response to reports of cyberbullying or abuse. This includes initial response
times and timescales on the removal of harmful content.
10
Parental Internet Controls Survey: Children and Young People’s (2012)
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 75
—
—
—
Ensure robust moderation of user generated content.
Request prominent signposting to sources of expertise, advice, support and help.
Require independent monitoring of a code of practice for industry, with results promoted to
parents and children.
5. Preventing Abusive or Threatening Comments on Social Media
5.1 Our latest survey11 of children and parents found:
— One in five 12–16 year-olds have interacted with strangers online.
— More than a third of 12–16 year-olds go online most often in their own bedroom.
— One in five 12–16 year-olds think being bullied online is part of life.
— More than a quarter of 12–16 year-olds admitted to witnessing bullying online, but only half
did something about it.
— The primary reasons young people gave for not doing anything about the online bullying was
being worried about being bullied themselves or not knowing who to speak to about it.
— Almost a quarter (23%) of 12–16 year-olds spend more than five hours a day online during
school holidays. More than double the number during term time (10%).
— The majority (80%) of 12–16 year-olds said they feel safe online, compared to only 60% of the
younger age group (8–11 year-olds). But worryingly, one in five (22%) of 12–16 year-olds said
they think being bullied online is part of life.
— For those 12–16 year-olds who did do something about the cyber bullying, most went to their
parents for advice, however only 38% of parents think their children are at risk of being
bullied online.
5.2 Whilst there are technological advances that allow streams of text to be analysed and flagged up for
bullying behaviour, BeatBullying works to educate and empower children and young people to take
responsibility for their actions and understand that bullying in any form is unacceptable.
5.3 Among primary-school children who participated in our 2012 research,12 6% of those who had
experienced cyberbullying had been targeted on a hate site or hate group (often referred to as Trolling), 9%
had had embarrassing images circulated and 10% had had a photo edited to make fun of them. There are also
qualitative accounts of children this age experiencing death threats and sexual harassment online. Among 11—
16 year-olds who had been cyberbullied, 6% reported receiving a message or image on the subject of sex
which had made them feel uncomfortable or upset. Of those who knew who had sent it, the vast majority
(80%) said that it had come from someone around the same age.
5.4 This research found that one in 13 secondary-school and one in 10 primary-school-aged children had
been subjected to an intensive, relentless campaign or online attack over a period of weeks, months or even
years by other children or young people. Many online phenomenon’s such as “sexting”—sending sexual
pictures and messages via mobile phone—is overwhelmingly peer-to-peer; it is an instance of children and
young people victimising each other and unknowingly victimising themselves, by sharing photos that can be
used as weapons for bullying or end up in the hands of adult predators. This perfectly illustrates the point that
a significant proportion of negative online content is generated by young people themselves. Given this, it is a
concern that any discussion of technological blocking and filtering systems in isolation does not take into
account the interactive and user-generated nature of some harmful material, which is likely not to come within
the scope of a filter anyway. To our knowledge, it is not possible completely to block interactions such as
bullying or “sexting”, although, as outlined above, there are both technological and human ways of curtailing
these and limiting the harm done. Blocks and filters can only block and filter; they cannot educate children
about how to deal with the kind of content and behaviour they might encounter online, how to manage their
own behaviour responsibly or why certain material is harmful.
5.5 That is why BeatBullying has been calling for several years for education programmes focusing on
behaviour and addressing issues such as cyberbullying and e-safety, to be introduced into all schools. We
support and welcome the proposed changes on e-safety to be introduced in the 2014 national curriculum across
all key stages. It is clear that education for safety and good behaviour online needs to begin as early as possible,
before cyberbullying even becomes an issue.
5.6 Everyone involved with children’s and young people’s use of the internet—parents, schools, service
providers, organisations and children themselves—needs to be seen as sharing responsibility for online safety.
In April 2013 BeatBullying launched a campaign for better anti-bullying protections called Ayden’s Law.13
This campaign called on the Government to produce a national anti-bullying/cyber bullying strategy which
would, for the first time, set out the role of teachers, parents, local authorities, the public, the voluntary &
community sector and children and young people themselves in preventing and intervening when bullying and
11
12
13
The Online Behavioural Habits of Young People Today, which questioned a 1000 8–16-year-olds and 1500 parents
from Parentdish and BeatBullying, July 2013
Virtual Violence II Part II: (2012) BeatBullying
Launched May 2013 with The Sun Newspaper following the tragic death of Ayden Olsen Aged 14
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 76 Culture, Media and Sport Committee: Evidence
cyber-bullying takes place. Without the introduction of this strategy, progress in tackling online bullying and
other safety threats will always be frustrated by the lack of understanding of roles and responsibilities by those
charged with keeping children safe online.
5.7 A national strategy to tackle cyber-bullying would set out how the voluntary and community sector,
parents, schools, LAs and the police would be equipped to (a) protect the children in their care from harm
online and (b) educate and equip children about internet safety and responsible digital citizenship so that they
understand the issues for themselves. The implementation of this can be greatly facilitated by a solid technical
framework of support from business, such as Claire Perry’s work with internet providers to introduce
preselected parental controls or the adoption of the measures listed in para 4.6 above.
5.8 More than 1,700 cases involving abusive messages sent online or via text message reached English and
Welsh courts in 2012. However, cyberbullying is not a specific criminal offence in the UK. Some types of
harassing or threatening behaviour—or communications—could be a criminal offence. These laws were
introduced many years before Twitter, Facebook and Ask.FM, and they have failed to keep pace with the
demands of modern technology. Unfortunately, serious cases of cyberbullying, which have often resulted in
suicide, have dominated our headlines in recent months. That is why BeatBullying have been calling on the
Government to review current legislation and make bullying and cyberbullying a criminal offence so that
children and young people have the protection they need and deserve, at the earliest opportunity, to avoid this
escalation. This year we have worked with Tracey Crouch MP to introduce bullying and cyberbullying as a
form of anti-social behaviour, which will result in victims and perpetrators benefiting from earlier interventions
and support outlined in the current Anti-Social Behaviour, Crime and Policing Act Bill.
5.9 For BeatBullying, the approach to online safety must ultimately be about shaping attitudes and changing
behaviors as much as it is about teaching techniques for staying safe or for anything else; it is about equipping
young people with the inner resources that will keep them safe, and responsible, long-term. Some negative
online behaviors such as cyberbullying and sexting are most likely to be a form of child-on-child aggression
rather than adult-generated; in order to tackle this successfully, it is crucial to work with perpetrators as well
as victims to address the attitudes that underlie the behaviour.
September 2013
Written evidence submitted by the Internet Watch Foundation
—
“Online Safety” is a very broad concept that covers a wide and diverse range of issues. When
addressing this topic, it is important to clearly identify the issue at stake as different problems
will require different solutions, from a technical as well as a procedural perspective.
—
The Internet Watch Foundation is concerned that the issues relating to legal adult content
filtering and the fight against criminal child sexual abuse images become conflated.
—
The Internet Watch Foundation has been very successful in fighting child sexual abuse images
and videos on the internet and has become one of the most successful hotlines in the world.
—
The success of the Internet Watch Foundation is mainly due to its specific focus on child sexual
abuse content and the continuing support from the online industry.
The Importance of Clearly Defining the Problem
1. Different problems require different solutions and it is therefore important to clearly distinguish between
the different issues under consideration. Combatting illegal online content (which is content that by definition
should not be uploaded or viewed by anybody) is different from protecting minors from accessing harmful
content (for instance adult content), which is content that in itself is not necessarily illegal. Preventing abusive
or threatening comments on social media might be considered a problem related more to conduct rather than
content, notwithstanding that the result of this conduct leads to abusive or threatening online content (ie the
actual comments).
2. The merit of taking into account the various aspects of online safety lies in the fact that each of these
specific issues opens up a different range of possible solutions, both on a technical as well as a procedural
level. Other actors can be involved, other safeguards and processes developed. The most effective solution in
one area of online safety could therefore be very different from the preferred solution in another. Equally, an
effective solution in one area might have negative implications for the effectiveness in another area.
Considering the various different challenges individually will protect against developing general, however wellintentioned, solutions that might have an unforeseen negative impact in particular areas of online safety.
Combating Illegal Content on the Internet
3. Even in the area of combating illegal online content, it is important to keep in mind the differences
between various types of illegal content, for instance the difference between child sexual abuse content and
material intended to promote terrorism or violence.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 77
4. For different types of illegal content different questions have to be considered and depending on the
answers different solutions can be formulated. Issues to consider include:
(i) “What is the nature of the content?” (What is the “level” of illegality? What is the impact on
the people viewing the content? What is the impact on the victims of the original crime?);
(ii) “How is the content available on the internet?” (Where can this content be found? Where is the
content hosted? What is the pervasiveness of the content on the internet? How easy is it to find/
stumble upon the content?);
(iii) “What is the preferred solution?” (How important is it to make the content unavailable as
quickly as possible? How can this content be found and identified? Who can decide whether
the content is potentially illegal? Which involvement of law enforcement/the judiciary is needed
at each stage of the solution—identification, investigation, removal?).
5. The answers to these questions will be different when considering material intended to promote terrorism,
material constituting a copyright infringement or child sexual abuse material. Depending on the answers,
different solutions will be available to effectively deal with the issue and different procedures, actors and
safeguards might have to be considered.
Fighting Online Child Sexual Abuse Content—The Internet Watch Foundation
6. The UK has currently one of the most effective systems in the world for dealing with child sexual abuse
content on the Internet. This success is based on the efficiency and experience of the Internet Watch Foundation
(IWF), the support it receives from the online industry and its close cooperation with law enforcement, in
particular with CEOP, the UK’s Child Exploitation and Online Protection Centre.
7. IWF Background: The IWF is the UK hotline where the public can report child sexual abuse content,
criminally obscene adult content and non-photographic images of child sexual abuse. The IWF is a charity
founded in 1996 by the UK Internet industry and it is therefore independent from Government and Law
Enforcement. However, the IWF has very close working relationships with law enforcement and its operations
are covered by a Memorandum of Understanding with the Crown Prosecution Service and the Association of
Chief Police Officers and a Service Level Agreement with the Association of Chief Police Officers.
8. The IWF is a membership organisation and is funded by over 100 global members, including internet
service providers (ISPs), mobile operators, content providers, hosting providers, filtering companies, search
providers, trade associations and the financial sector, as well as by the EU as part of the UK Safer Internet
Centre together with Childnet International and South West Grid for Learning. As an independent body,
transparency and accountability is of the utmost importance and the IWF is governed by an independent Board
and the Hotline is independently audited by a highly qualified team every two years. The IWF also operates a
thorough complaints procedure and has commissioned an independent Human Rights Review.
9. IWF processes: When the IWF receives a report from the public, its analysts assess the content to confirm
whether it is within its remit and potentially in breach of UK legislation. If the content is considered to be
potentially criminal, the IWF can take action against the child sexual abuse images and videos hosted anywhere
in the world. It can also act against criminally obscene adult content and non-photographic child sexual abuse
content hosted in the UK.
10. Depending on where the content is hosted, the process for dealing with potentially criminal images and
videos differs. When child sexual abuse content is found to be hosted in the UK, the IWF will inform CEOP.
After confirmation from CEOP that action can be taken, the IWF will notify the hosting provider who will
remove the content from its servers, typically within 60 minutes after receiving the notification from the IWF.
This process is commonly referred to as “Notice and Takedown”.
11. When child sexual abuse content is found to be hosted outside the UK, the IWF will inform its
counterpart hotline in the hosting country through INHOPE, the international association of hotlines or link in
directly with local law enforcement. As other countries take significantly longer to remove child sexual abuse
content—50% of the content the IWF passes on internationally is still available after 10 days—the IWF adds
the links (URLs) to the content to its URL list (or “blocking list”). IWF members can use this list to voluntarily
block access to these URLs to protect their customers from stumbling upon the images and videos. The URLs
on the list are as targeted as possible, often comprising a single image or video. This is done in order to prevent
the blocking of anything outside the criminal images or videos.
12. In addition to “Notice and Takedown” and the URL list, the IWF also compiles a keyword list of terms
that specifically refer to child sexual abuse content. This list is used, for instance, by search engines to prevent
people from finding images and videos of child sexual abuse content. The keywords are very specific—or very
specific combinations of words—that carry no meaning besides the specific reference to child sexual abuse
content. This means the keywords will not prevent access to legitimate websites such as, academic research
papers into the area of child sexual abuse or websites aimed to help or inform people in relation to child
sexual abuse.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 78 Culture, Media and Sport Committee: Evidence
13. Finally, the IWF also monitors newsgroups and can issue notices for removal of individual postings
where child sexual abuse content is identified or entire newsgroups if these are being used to specifically
distribute child sexual abuse content.
14. IWF achievements: In 2012, the IWF processed 39,211 reports and assisted with the removal of 9,696
URLs containing potentially criminal child sexual abuse content. A URL can be as specific as a single image
or could refer to an entire website containing potentially thousands of child sexual abuse images or videos.
The majority of victims (81%) appeared to be 10 years old or younger (with 4% 2 years old or under) and
53% of the images and videos depicted sexual activity between adults and children, including rape and sexual
torture. Over the past 17 years, the IWF has assessed over 400,000 webpages and assisted with the removal of
100,000 URLs containing potential child sexual abuse content. By sharing intelligence with the police, the
IWF aided with the identification and rescue of 12 children in the past three years.
15. IWF success: In the past decade, the UK has become the world leader for dealing with child sexual
abuse content and the IWF is considered as one of the most successful hotlines in the world. Both the amount
of reports received/processed by the IWF and the speed with which action is taken is amongst the best within
Europe and beyond. As a result, the share of child sexual abuse content that is hosted in the UK is less than
1% of the total amount of known content, down from 18% in 1996 when the IWF first started. This means
that—because of the efficiency of the current system—the UK has become an extremely hostile territory to
host child sexual abuse content.
16. The success of the IWF is foremost due to its very clear and specific remit. Because of the strong focus
on child sexual abuse content, the IWF has accumulated a very thorough and extensive knowledge on how to
deal with this content effectively. On the one hand, this means the IWF’s analysts have become experts in this
area, often recognising images and even victims/perpetrators within the images. They can pass on this
information to law enforcement to aid the investigation into the sexual abuse. On the other hand, the IWF has
been able to develop a strong network to fight the availability of child sexual abuse content. This multistakeholder approach is crucial for successfully dealing with child sexual abuse content and the IWF has built
a very strong network, both with law enforcement, the online industry and international partners, for specifically
dealing with child sexual abuse images. In addition, the success of the IWF is also a result of its self-regulatory
model which allows the IWF to adapt quickly to new developments and of the strong support it has received
from the online industry over the past 17 years.
17. IWF future: The online environment is ever changing and the IWF is committed to continue developing
in order to fight child sexual abuse content effectively in the future. Following a donation from Google and
additional funding from other Members, the IWF will increase its number of analysts from 4.5 (full-time
equivalents) to 11.5 (full-time equivalents). In addition, the IWF was asked by Government to start proactively
searching for child sexual abuse content rather than working only from public leads. The IWF will continue
its role as the UK hotline but will be able to add-on the proactive work in order to increase the amount of
child sexual abuse images it can take action on. An ongoing membership review is addressing the need to
improve IWF’s resources in order to ensure a sustainable future for the IWF and its increased activities.
18. A final aspect of IWF’s future development concerns its international activities. As 99% of the known
child sexual abuse content is hosted abroad, it is of the utmost importance that international cooperation is
further developed to remove the child sexual abuse images and videos at source. The IWF is actively engaging
with international partners to share its expertise. It has developed “OCSARP”, its “Online Child Sexual Abuse
Reporting Portal”, which countries without child sexual abuse content reporting facilities can implement, while
the IWF provides assistance with the assessments of the reports.
Conclusion
19. The area of “Online Safety” is very broad and very volatile, with risks as well as possible solutions
developing constantly. It would be worthwhile to clearly frame the debate and break up the overall topic in
clearly defined aspects of online safety. Online conduct is different from online content. Criminal or illegal
content is different from legal but potentially harmful content and an image of a child being sexually abused
is different from the promotion of terrorism. This is not to say there is necessarily a hierarchy between the
“severeness” of different types of illegal content, but different types of content will require different procedures,
different checks and balances and the involvement of different actors.
20. The IWF—and the online industry—has been very effective in fighting child sexual abuse content over
the past 17 years. The IWF’s very specific self-regulatory and multi-stakeholder model is extremely effective
for dealing with child sexual abuse content and the IWF remains committed as ever to continue developing
and improving the fight against child sexual abuse content in the years to come.
September 2013
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 79
Written evidence submitted by Internet Services Providers’ Association (ISPA)
Introduction
The Internet Services Providers’ Association is the trade association for the internet industry in the UK.
ISPA has over 200 members from across the sector, including a large number of access provider ISPs from
small to large, content platforms, hosting providers and others. ISPA therefore works across the areas in which
the Committee is looking and we welcome the opportunity to provide input into the Committee’s inquiry.
We believe that the Committee’s terms of reference for the inquiry identify the key issues that are currently
of relevance in relation to online safety. However, we would like to emphasise that the issues that have been
identified should not be conflated. The issue of tackling child abuse content, which is clearly illegal, requires
a different response from industry and Government than the availability of extremist material which may or
may not be illegal. Protecting children from accessing potentially harmful content again requires a different
response as it may cover content that is clearly legal but simply not appropriate for children and young people
under 18.
We further welcome that the Committee considers that any potential dangers of the internet are a “correlation
of the immense benefits provided by unimpeded communication and free speech” and that “any attempts to
mitigate harms have to be proportionate and, where possible, avoid disadvantageous consequences.” We believe
that the recognition of this correlation is of vital importance but are concerned that policy-makers sometimes
disregard it which often leads to disconnected and potentially harmful policy-making.
Variety of Internet Companies
It is important that the Committee understands that there is a considerable diversity of companies that operate
internet services. When considering the steps that industry can take, it is important to consider that each type
of company may be playing a different role, and they will have varying degrees of ability to deal with
potentially illegal or harmful content. The below description provides a rough guide14 to the various kinds of
companies that are involved in making the Internet work. If it were felt to be helpful, we would be happy to
brief the Committee in more detail about the position of each company type in the internet value chain.
Access Providers
Access providers are commonly referred to as Internet Service Providers. They connect customers to the
Internet—either through fixed or wireless connectivity. As the ISP does not initiate or modify their users’
communications and is only passing traffic across a network, they are deemed “mere conduits” under the ECommerce Regulation 17 which grants limited liability.
Hosting Providers
Hosting providers store others’ content online, often for a charge. Traditionally hosting providers have hosted
complete websites of individuals and companies and even Government hosts some of its websites with these
private hosting providers.
More recently, new types of hosting provider have emerged. These providers, eg social networks, generally
provide a platform on which users can upload content (videos, blog posts, images etc.) which they themselves
have created. These kinds of hosting provider do not have editorial control over what is posted on their services,
but may have active or passive moderating policies that allow them to remove content or restrict its availability.
Under Regulation 19 of the e-Commerce Regulations both traditional and modern hosting providers are not
considered to be liable for the content they host as long as they do not have actual knowledge of unlawful
activity or information. However, upon obtaining such knowledge, hosting providers become liable if they do
not act expeditiously to remove or to disable access to the information.
Websites Where Operators Have Editorial Control
Individuals, companies and other organisations that run their own websites can be regarded as having
editorial control over content that is available via their website and so are considered to have more direct
responsibility. However, it is important to point out that websites can contain both content where the operator
of a website has editorial control, eg a news article, and user generated content, eg comments about that
news article.
Search Engines
Search engines index web pages by scanning the Internet. They use algorithms to display relevant results
based on what search terms users input but generally do not exercise editorial control over the links that they
present to users. Search engines can be considered as “caches” under Regulation 18 of the e-Commerce
14
We merely consider the e-Commerce Regulations and online companies may have duties and defences through other legislation.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 80 Culture, Media and Sport Committee: Evidence
Regulations and act expeditiously to remove or to disable access to any information if they are made aware of
that the fact that this information may be illegal.15
How does this apply in the real world?
It is worth considering three different examples:
1. A website hosting child abuse images
2. A person who posts a potentially illegal message on a website operated by the same person
3. A person who posts a potentially illegal message on a forum operated by a third party
In relation to the first example, ideally the hosting providers who provides the space for the website should
be notified that illegal material is being hosted on a website on one of its servers. This notification is being
done on a regular and effective basis by the IWF. If the operator is based outside of the UK and responds
slowly or not at all to a notice from the IWF or its international partners, the IWF can add this page to its list
of illegal websites. Access providers accept the judgment of the IWF, which has great expertise in this area,
and use the IWF’s list to filter out the relevant page (ie they make the URL of that website inaccessible, and a
user would see an error message if they attempted to access it).
In relation to the second, the person should be approached directly as they have editorial control of the
comment and the website on which it can be found. If the person does not respond then it may be necessary
to contact the hosting provider who provides the space for the website who may then need to make an
expeditious assessment of the content and take it down if appropriate. The access provider would theoretically
be able to block access to the website but this would be less timely and cost efficient than approaching the
hosting provider and generally requires a valid takedown notice.
In relation to the final example, again, the person who posted the content should be approached directly.
However, if the person does not respond, or cannot be identified, the third party who operates the forum should
be approached who will then need to make an expeditious assessment of the content and take it down if
appropriate. If the third party does not react then it may be necessary to approach the hosting provider, however,
this should be a matter of last resort as the provider would generally only be able to remove the whole forum,
thereby curtailing the service and rights other forum users. The same would be true if an access provider would
block access to the forum.
In all these examples it is important to note that it is often not clear cut whether content is illegal or not and
online companies cannot be expected to be the judge and jury of others’ content.
Industry has a Role to Play
We strongly believe that the industry has a vital role to play in protecting minors from accessing inappropriate
content and would like to emphasise that the UK is widely regarded as having one of the most advanced online
safety frameworks.
Family Friendly Filters
—
The main consumer facing ISPs are moving to system where new and existing customers are
presented with an unavoidable choice of whether to apply filters or not. These filters cover the
whole home, ie apply to all the devices used on a connection, put users in control and allow
parents to choose from a list of content that should be filtered including adult content, extremism
and self-harm. This has involved significant investment, both financially and time.
—
Some smaller consumer-facing providers are considering solutions that offer family friendly
filters but can be deployed on smaller scale and at lower costs. ISPA is currently discussing
this issue with its members.
Child Sexual Abuse Content
15
—
ISPA and many ISPs have helped to setup the IWF and have consistently supported the
organisation which is considered to be world class in preventing people from access child abuse
content and facilitating the removal of that content at source.
—
Many ISPs have committed to increase funding of the IWF to enable a proactive remit to
identify and remove child abuse content online. Further funding for education and awareness
campaigns from industry has been forthcoming.
Regulation 18 defines providers as “caches” if the information that is processed by them is “the subject of automatic, intermediate
and temporary storage where that storage is for the sole purpose of making more efficient onward transmission of the information
to other recipients of the service upon their request”, ie, the provider merely processes information and only stores it temporarily.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 81
Allegedly/Potentially Illegal Content
—
Industry has worked with the Home Office and law enforcement regarding the CounterTerrorism Internet Referral Unit (CTIRU) which members of the public can report potentially
unlawful terrorist material from the internet. If hosted in the UK the content is removed and
this framework is underpinned by the Terrorism Act 2006. So far approximately 6,500 pieces
of online content have been removed through CTIRU action.16 However, what constitutes
terrorist material is not always clear cut.
—
Providers of social media platforms and websites that contain user generated content will
remove illegal content whenever they are made aware of it and can apply their terms and
conditions to other types of content that may not be illegal. They also often provide their
consumer with report facilities to flag up any inappropriate behaviour.
—
Providers have been working alongside Government and Parliament to reform the defamation
law to ensure that online freedom of speech is adequately balanced with the right of those who
feel they have been defamed online.
Industry Cannot Solve These Issues on its Own
However, we are concerned that the current policy debate is sometime too strongly focused on finding a
technological fix to a problem that often has societal roots and is sometimes present in both the offline and
online world.
For example, in the relation to the accessibility of adult content, we accept that ISPs should play a role in
empowering their customers to better determine what content should be available in their household. However,
even the most comprehensive filtering solution cannot guarantee that adult content will be unavailable. Over
and underblocking of content is inevitable and it is important that filtering tools are viewed as part of a wider
package alongside education and parental mediation. There needs to be more emphasis on enabling parents
and teachers to teach children how to behave responsibly online, one possible action could be the updating of
sex education in the curriculum so that it keeps pace with technological and societal developments.
In relation to abusive or threatening comments online, we would like to emphasise that ISPs should not be
used as proxy for enforcing the law and perceived societal standards. Social media networks can and often
take action against users that behave inappropriately. However it has to be taken into account that the Crown
Prosecution Service’s Guidelines on prosecuting cases involving communications sent via social media state
that “[j]ust because the content expressed in the communication is in bad taste, controversial or unpopular, and
may cause offence to individuals or a specific community, this is not in itself sufficient reason to engage the
criminal law.” This should not be regarded as a get out clause for providers but it is important to point out that
providers cannot be expected to go beyond what is required by the law. In this context, it worth highlighting
that Parliament has recently amended the Defamation Act which encourages hosting providers to assist the
resolution of disputes between users that cannot be resolved by the hosting provider themselves
Conclusion
We have shown that industry has made available a number of tools, services and advice to help protect
minors from accessing adult content. There is cooperation between industry and law enforcement to tackle
extremist material when legal thresholds are crossed. Websites have in place mechanisms to prevent abusive
behaviour and the law had been used to prosecute individuals in some instances.
The Internet has had a significant impact on modern societies. It has changed how we do business,
communicate, educate and consume content. These changes came about because internet companies developed
innovative products and consumers have found even more innovative and sometime unexpected ways of using
these products. As such the Internet is an extension of the established offline world and it would be wrong to
simply ask ISPs to fix any issues that may arise.
Technological fixes can play a role and support customers but we will only be able to comprehensively
tackle the problems that the Committee outlined in its terms of reference by involving industry, Government,
parents and users and by looking at both the symptoms and causes. The Internet industry has reviewed and
improved its offering to customers in recent times. It is willing to actively engage with the online safety agenda
but we hope that this can be done in a more positive environment based on collaboration.
October 2013
16
HL Deb, 23 September 2013,c421W.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 82 Culture, Media and Sport Committee: Evidence
Written evidence submitted by TalkTalk
Summary
1.1 As the first internet company to introduce a whole home parental control solution, TalkTalk welcomes
the Committee’s Inquiry into Online Safety. We firmly believe that everyone should be able to enjoy the
benefits of the internet, but that as an ISP we have a duty to provide our customers with simple and effective
tools to ensure that they are able to search, download and utilise the internet in as safe a way as possible,
especially when it comes to parents wanting to help protect their children online.
1.2 TalkTalk is currently the only ISP in the UK to offer a whole home parental controls system that protects
every device connected to the home broadband. Built into the broadband network itself, HomeSafe gives
families the ability to control the types of websites they want their household to access across all devices using
the connection and restrict access to content like pornography, gambling, and suicide and self-harm. So far
over 1.2 million homes are using HomeSafe, and of these 400,000 have chosen to activate parental controls.
In March 2012, we launched so called active or unavoidable choice for all new customers signing up to
broadband and in June of this year, we extended this to existing customers and began the process of asking all
our existing customers to make a choice about using parental controls.
1.3 1 in 3 customers—roughly equivalent to the number of UK homes with children in—are choosing to
use filtering when presented with a choice. Whilst much of the debate on internet safety has focused on children
accessing pornography, suicide and self-harm remains the biggest category that customers are choosing to
block, followed by pornography which is then followed narrowly by weapons and violence. This reflects the
range of concerns that parents have when it comes to their child’s safety online and indeed many of their
concerns aren’t about accessing inappropriate content, but are behavioral such as cyber bullying, sexting and
grooming.
1.4 When it comes to filtering, our research shows our customers like being asked to make a choice—80%
think it’s a good thing and 60% wouldn’t have activated controls if they hadn’t been prompted. We support
this approach and believe that it is critical that parents make an engaged decision, appropriate to their individual
family, about the use of filtering tools. We would never filter access to content that is legal (even if it is agerestricted) unless our customers asked us to. Parents should make the decision as part of their approach to
helping keep their children safer online.
1.5 As a member of the IWF, TalkTalk does take the IWF list and voluntarily block this at network level so
none of our customers can access websites that they have identified to contain illegal child abuse images. In
the summer, we—and other ISPS—also implemented a new form of wording on the block pages to act as a
stronger deterrent. We are also big supporters of the IWF’s move to proactive detection.
1.6 There is no silver bullet when it comes to internet safety, and much relies on education and awareness.
TalkTalk conducts extensive marketing and communications to customers about HomeSafe and internet safety
and we have also partnered with the likes of the The Parent Zone, Girl Guiding UK and the UK Safer Internet
Centre on various educational initiatives. Along with other ISPs, we have committed to launch of a national,
sustainable education campaign targeted at parents in the New Year.
1.7 We have long called for internet safety to be included in the Primary curriculum and we are really
pleased that will be the case going forward. We think of internet safety as the road safety of our children’s
generation, in that it required a range of different inventions including the Green Cross Code, air bags and
mandatory seat belt wearing, and required us all to play our part. Internet safety requires the same approach
where use of technical tools needs to be supported by awareness raising and education for parents and young
people alike.
2. How Best to Protect Minors from Accessing Adult Content
2.1. TalkTalk shares the commonly held belief in the importance of protecting children online. The internet
is an integral part of people’s lives now and has been transformational in terms of the benefits that it brings.
However, as with life offline, there are unfortunately risks associated especially for children and young people
and we believe we have a role to play in trying to protect them from harm.
2.2. TalkTalk thinks of internet safety as the road safety of our children’s generation; it requires a range of
different inventions, and the use of technical tools needs to be supported by awareness raising and education
for parents and young people alike. We think technical tools have a really important role to play, but alone
they will not solve the issue. We believe the most important thing a parent can do is take an active role and
talk to their children about what they do online, and combine this with using controls as appropriate.
HomeSafe
2.3 Back in May 2011, TalkTalk introduced HomeSafe, the UK’s first and still only whole home parental
controls system that allows parents to protect every device connected to the home broadband and control the
types of websites their family is able to visit. From our own customer research we found that parents want an
easy to use system that allows them to activate their chosen filters across all devices with one click without
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 83
having to do this separately on each device in the home. We also found that many parents found existing
device level controls confusing and we developed HomeSafe to fill that gap.
2.4 HomeSafe has three simple features; Kids Safe, Virus Alerts and Homework Time.
—
Kids Safe—parental controls that allow the account holder to block content they don’t want to
be accessed on their connection. There are nine different categories, and customers can also
choose to block other specific websites.
—
Virus Alerts—an alert system that blocks access to web pages infected with malware and
phishing sites.
—
Homework Time—this allows parents to block social networking and online gaming sites—
common sources of distraction for children from homework—during a specified time of day.
2.5 The most commonly blocked categories are, in order, Suicide & Self-Harm; Pornography; Violence &
Weapons; Dating; Drugs, Tobacco and Alcohol; Gambling; File-sharing; Gaming Websites; and Social
Networking.
Asking our customers to make a choice
2.6 In March 2012, TalkTalk became the first internet service provider to introduce a genuine unavoidable
choice for new customers when they signed up to TalkTalk broadband, as per the recommendation of the Bailey
Review. Customers are asked to make a “yes” or “no” decision as to whether they want to filter access to
content that might be inappropriate for under 18s on their broadband connection or not. TalkTalk then applies
this to their internet connection as soon as its live and no further action is required by the customer. If the
customer wants this on, then Suicide and Self Harm; Pornography; Weapons and Violence; Dating; Drugs,
Alcohol and Tobacco; and Gambling will be blocked. The settings can be easily changed at anytime through
the customer’s online My Account facility.
2.7 At the beginning of June of this year we extended this unavoidable choice to existing customers and
began the process of asking all of our existing customers who haven’t yet engaged to make a choice about
using Kids Safe filtering. We are doing this via the use of a pop-up screen when a customer who isn’t using
parental controls logs into their My Account. As with new customers, they have to make a decision either way
before they progress any further and access their My Account. We have already asked a million of our existing
customers to decide about using Kids Safe.
2.8 We currently have 1.2 million customers who have activated HomeSafe and over 400,000 customers
have activated Kids Safe filtering. 1 in 3 customers are choosing to turn on Kids Safe when presented with an
unavoidable choice. This is about equivalent to the number of households with children in and so we believe
represents the level of uptake that one should expect, given that not all parents will want to use filtering but
other customers, like grandparents or those who simply want to filter content out due to personal beliefs, might
want to.
2.9 By the end of 2013 we will move to pre-ticking the “yes” option, which we believe is a strengthening
of our current position. It is incredibly important however to note that this does not constituent default blocking
and all customers will still be asked to make a decision about filtering; we would not filter without them
requesting it. We think it is critical that parents make an engaged, informed choice about whether or not they
wish to use filtering as part of their approach to helping keep their children safer online.
2.10 Our customers agree with this and our research shows they like being asked to make a choice. A survey
of our customers found:
—
Around 80% felt that being prompted to set up parental controls was a good thing.
—
60% of those who set up controls said they wouldn’t have done it if they hadn’t been prompted.
—
Of those who didn’t activate parental controls 60% said they didn’t turn them on because they
didn’t have children and 30% said they didn’t want/need them.
Closed feedback loop
2.11 There are a number of safeguards in place to help ensure that children aren’t changing the settings
without their parents’ knowledge. Firstly, the initial set up is done during the sign up process for fixed line
broadband where a customer will likely be credit checked, have to provide bank details and be arranging for a
new broadband service to be installed/an existing service to be taken out. They are also entering into a 12 or
18 month contract and our terms and conditions state our customers must be aged 18 or over. Secondly, the
log in details for a customer’s My Account—where they make any subsequent changes to their HomeSafe
settings—are set up during the joining process by the account holder. At any time, from anywhere, the account
holder can log in to My Account to check the status of their home’s filtering preferences. Every time a change
is made to the settings an email is sent to the email address the account holder signed up with confirming the
changes. Finally, every customer can also opt in for free additional notifications via SMS to their mobile.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 84 Culture, Media and Sport Committee: Evidence
Over-blocking
2.12 We know that there are concerns about over-blocking of websites in relation to whole home filtering
systems. Firstly we think it is really important to remember that filtering systems have been in place for years—
they were just applied at the device level rather than at the network level. When it then comes to issues around
inaccurate blocking, we think there are two key things; firstly obviously that the filtering technology itself is
being constantly improved and refined, and secondly that there are clear reporting mechanisms which are
actioned.
2.13 In developing HomeSafe we deliberately tried to take a more holistic approach to website categorisation
which looks at the website in its entirety—so for example an article on the BBC website which mentions
pornography wouldn’t be blocked. Similarly, a website like Frank which offers advice for those suffering from
drug addiction wouldn’t be blocked, whereas a website advising on how to make drugs would be. We think
this is a more pragmatic approach. Clearly no technology is 100% perfect and so we continue to work to
improve the filtering. We have clear reporting mechanisms, either directly via a large “report” button on the
block page itself or through a feedback form in the customer’s My Account. Website owners can also contact
us via email. This enables us to address any issues of misclassification and also employ any learning to avoid
other instances. The volume of this is respectively low and we aim to action reports within five days.
2.14 We have also formed partnerships with a number of companies to either “white” or “black” list. Papyrus,
the suicide prevention charity, provides us with a list of websites that promote suicide and we include those in
the suicide and self-harm category. Similarly, we have an agreement with the British Association for Shooting
and Conservation (BASC) to ensure that websites with content about lawful shooting interests aren’t blocked.
Marketing, advice and guidance
2.15 HomeSafe is actively supported by comprehensive marketing and communications that aim to educate
our customers about internet safety and the things they can do to help protect themselves and their children
online. We have a dedicated internet safety hub (talktalk.co.uk/security) that provides a wealth of guidance.
This is currently being updated, including aligning with the UKCCIS guide on child internet safety—which
we helped to develop. We also send regular communications about HomeSafe and internet safety to our
customers. Other marketing has included extensive television and outdoor advertising; we believe TalkTalk is
the only company to have advertised about child internet safety on television. We have also run dedicated
security workshops for our customers in London and the key insights and advice from the day are then shared
back to the rest of our customer base. For Safer Internet Day this year we ran an internet safety road show
which visited 5 locations around the country and offered advice and guidance on internet safety to the public.
2.16 In addition to the above, we run or are involved in a number of additional initiatives to educate and
raise awareness about internet safety. We have been working with The Parent Zone to develop The HomeSafe
Family Challenge’, a three-term programme of activity for schools to help them engage parents on child
internet safety. Each term covers a specific theme—Digital Footprint, Digital Know How and Digital Living—
and within each theme there are a range of topics covered for parents and teachers. The HomeSafe Family
Challenge is designed to get parents to engage with the issue and take action. Each theme will have a challenge
associated with it to engage parents and encourage action, for example getting a certain number of parents to
take the “HomeSafe family challenge quiz” with a prize for the school if they do. We have been trialing this
with a number of schools and are now rolling this out to more schools across the UK.
2.17 We have also partnered with Girl Guiding UK to help them redevelop their Computer badge for
Brownies and as part of this will produce supporting educational materials on internet safety for both the
Brownies and their volunteers as internet safety forms one of the elements of getting the Computer badge.
This will be supported by a significant communications campaign about internet safety to the Brownies and
their volunteers.
2.18 We have also worked with The Parent Zone to develop a very simple guide to parental controls which
has been sent to schools across the UK and is available for free download from their website, and also with
the UK Safer Internet Centre to create and promote a video on setting up HomeSafe.
2.19 Finally, along with the other ISPs, we have committed to launch of a national, sustainable education
campaign targeted at parents in the New Year.
3. Filtering out extremist material, including images of child abuse and material intended to promote
terrorism or other acts of violence;
3.1 Like other industry members, TalkTalk takes a zero tolerance approach to illegal child abuse images
online. Currently our efforts are focused on helping prevention. We are a longstanding member of the Internet
Watch Foundation and pay the maximum membership fee. This summer we, along with the other ISPs,
committed to a further £1 million of funding for the IWF to continue its work in identifying illegal child abuse
images on the internet. TalkTalk also chaired the working group that supported the membership review of the
IWF which has led to their move to proactively seeking out illegal child abuse images online in addition to
following public reports.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 85
3.2 We automatically download the block list of known websites containing illegal images from the IWF on
a daily basis, and apply it to our filtering system, which means that every customer is blocked from accessing
those websites across the whole network.
3.3 Customers who try and access these websites are presented with a splash page informing them the
content is blocked because it has been found to be illegal by the IWF. In the summer, a group comprising the
IWF, CEOP and Lucy Faithful Foundation, a charity who work with offenders, agreed a new stronger set of
wording that TalkTalk implemented in August, along with the other three major ISPs. The wording is:
Access has been denied by your internet access provider because this page may contain indecent
images of children as identified by the Internet Watch Foundation.
Deliberate attempts to access this or related material may result in you committing a criminal
offence.
The consequences of deliberately accessing such material are likely to be serious. People arrested
risk losing their family and friends, access to children (including their own) and their jobs.
Stop it Now! can provide confidential and anonymous help to those with concerning or illegal
internet use. They have helped thousands of people in this situation.
0808 1000 900 | [email protected] | www.stopitnow.org.uk
If you think this page has been blocked in error please contact [email protected] or visit:
http://www.iwf.org.uk/accountability/complaints/content-assessment-appeal-process
3.4 We also work closely with CEOP, and our Chief Executive sits alongside CEOP on the UKCCIS Board.
We fully cooperate with their requests for information under RIPA and are currently in discussions about other
ways in which we could support their work further. We signpost customers to their helpline via the “Report
Abuse” button that sits on our customer safety information centre.
3.5 We belive the issue of extremist material and material intended to promote terroism or acts of violence
is predominatly a matter for law enforcment agencies and should be addressed via a joined up, multi-pronged
approach to tackling extremism and terroism. As an ISP we adhere to requests to remove content that has been
found to be illegal under current UK legislation. This may include material found to be extremist or material
intended to promote terroism or acts of violence. We are also in the process of finalising an agreement with
the Home Office to include their filtering list of known terroist websites that breach Terrorism Act 2006 but
that the policing Counter Terrorism Internet Referral Unit have been unable to get taken down, for example
because they are hosted abroad. We intend to add this into our “weapons and violence” category within
HomeSafe, so that our customers can choose to filter access if they want.
4. Preventing Abusive or Threatening Comments on Social Media
4.1 Our own research supports that cyberbullying is one of the main concerns for parents and children alike
when it comes to the internet and it is clear that more needs to be done both to prevent cyberbullying and
support the victims. Predominantly however we believe that the issue of cyberbullying is a behavioural issue
and therefore one that needs to be tackled through education, and then those websites owners and as oposed
to action we can take technically as an ISP.
4.2 In terms of what we are able to do, we investigate any abusive or threatening comments, posted on sites
by our customers when we are provided with the log information that supports the complaint. In these cases
where we can, we will identify the account and contact the account owner to make them aware of the complaint.
We inform the customer that this action is undertaken because of a breach of our Terms and Conditions of
supply. Depending on the severity of the abuse, some cases may result in court action. In these instances, we
would not contact the customer to make them aware, but we will disclose relevant data to the third party
solicitor, on receipt of a fully sealed court order from a UK court.
4.3 We do provide customers with the option of blocking social media via HomeSafe however we do not
think that this solves or helps the issue of cyberbullying. Our safety centre also includes a range of information
about cyberbullying for parents and we signpost to organisations like BeatBullying for further support.
About TalkTalk Group
TalkTalk is the UK’s leading value for money TV, broadband and phone provider with 4 million customers
across the UK. TalkTalk operates the UK’s largest Next Generation Network that covers 95% of UK homes.
TalkTalk is one of seven partners behind YouView, the internet-connected TV service, along with the BBC,
ITV, BT, Channel 4, Arqiva and Five. YouView launched to UK homes in 2012 and in August 2013 TalkTalk
announced it had signed up over 500,000 customers to the service. TalkTalk is also the only provider to provide
a whole home parental controls service, HomeSafe, to its customers free of charge. HomeSafe protects every
device using the broadband from accessing inappropriate content and since launch has been activated by 1.2
million customers. TalkTalk Business has been a leading supplier of broadband and voice for almost 20 years
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 86 Culture, Media and Sport Committee: Evidence
to over 180,000 businesses across the UK. TalkTalk is a founding partner of charity Go ON UK, which aims
to make the UK the most digitally included nation in the world.
October 2013
Written evidence submitted by the Mobile Broadband Group
Introduction
1. The Mobile Broadband Group (“MBG”), whose members are the UK businesses of EE, Telefonica UK
(O2), Three and Vodafone, welcomes the Culture, Media and Sport Select Committee’s Inquiry into online
safety.
2. We also welcome the opportunity to submit information on the significant steps that the mobile operators
have taken over the years, both individually and collectively, to mitigate risks that customers, particularly
children, may face when accessing content and services on-line.
3. The mobile operators’ work in this area is underpinned by a Code of Practice that was first published
nearly ten years ago in January 2004—“The UK code of practice for the self-regulation of new forms of
content on mobile”. The second edition of the code was published in 2009 and the third (and current) edition
in July 2013, reflecting the mobile operators’ determination to keep the code up to date and relevant to evolving
market conditions.17
4. The Code was first published in anticipation of the mass take-up of new mobile devices with enhanced
features, including colour screens, picture messaging, video cameras and Internet browsers. Today the majority
of children have a mobile and an increasing proportion of them go online using a mobile phone or smart
phone.18
5. The intention behind the Code is that parents and carers should have access to the information with which
to show their children how to use mobile devices responsibly and the tools with which to influence the type of
content accessible to children.
6. The Code was the first of its kind and was used as the boiler plate for similar codes introduced my mobile
operators throughout the EU. The UK has thus made an enormous contribution to child safety in Europe and
beyond in this field. It is a universal topic, where international co-operation can and must complement work
done in the UK.
Provisions of the Code of Practice
7. The Code covers a broad range of topics: commercial and Internet content, illegal content, malicious
communications, spam communications and customer education.
Commercial and Internet content
8. The mobile operators’ respective responsibilities for commercial content—where they have contractual
agreements in place with content providers—as against general content on the Internet are different.
9. Commercial content is classified in accordance with a framework provided by an independent body and
any content with an 18 rating is placed behind access controls. Such content is not made available to customers
until they have been through a robust age verification process (acceptable methods of which are set out in
the Code).
10. In 2013, the mobile operators appointed the British Board of Film Classification (BBFC) as the provider
of the independent framework (replacing the Independent Mobile Classification Body).. The agreement took
effect on 2nd September 2013.
11. The mobile operators’ partnership with the BBFC, who have increasing expertise in the on-line arena,
enables consistent, evidence based and transparent decisions about the categorisation of commercial content
and the calibration of Internet filters.
12. The Code requires that the mobile operators offer a filter to customers. The BBFC’s role, on the basis
of all relevant evidence, is to provide advice to mobile operators on where to set their Internet filters. The
framework focuses on content that may be harmful for children (ie those under the age of 18) to view—such
as pornography, violence, hate, promotion of suicide, illegal drugs and is set here: http://www.bbfc.co.uk/whatclassification/mobile-content/framework.
13. The mobile operators, though, are not setting out to police the Internet or online providers. It is
Government’s role to set and enforce societal standards. Providers of age restricted products such as gambling,
weapons, publications and alcohol should still be putting in place their own processes to ensure that minors
are not accessing, or taking delivery of, their restricted goods and services.
17
18
http://www.mobilebroadbandgroup.com/documents/UKCodeofpractice_mobile_010713.pdf
Ofcom, Children’s media use 2012
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 87
14. In addition to providing the independent standards framework, the BBFC has within its remit the ability
to review complaints about whether or not particular websites should or should not be placed behind the filter
and provide guidance to the mobile operators on the filtering of the sites that are the subject of the complaint.
Where there have been instances of inadvertent or unfounded over-blocking, there is a quick and transparent
process for having web-sites removed from filtering.
15. The BBFC framework is binary—18 or unrestricted. This is because 18 is the only age at which it is
currently practical to implement convenient, ubiquitous and robust on-line age verification. Stricter filters are
available in the market for parents that may want a narrower range of content for younger users but these fall
outside the Code.
16. The “18” filter is set by default for prepaid phones on all networks. For contract phones, where the
customer has to prove that he or she is at least 18, the filter is set to on by default for three of the four
networks. The operator Three will also put the filter on by default in 2014.
Measuring effectiveness
17. Obtaining robust, quantitative research as to whether the approach has been effective at protecting
children from harmful content is difficult, not least because of the practical problems associated with
researching this topic with children.
18. Ofcom and the London School of Economics led EU Kids On-line project provide two potential sources
of indicative information.
Seeing things that make me feel sad, frightened or embarrassed (%)
TV
Internet
Mobile
8–11
22
16
10
12–15
14
10
5
Fig 113
Fig 114
Fig 119
Data from Ofcom “Children and Parents: Media Use and Attitudes Report”, October 201219
Table 6: Child has seen sexual images online or offline in past 12 months, by age and gender (%)
9–12
Boys
Girls
Boys
TV/film/video
On any web site
In magazine or book
On a mobile
10
6
6
0
7
3
3
0
22
22
18
12
13–16
Girls
23
13
13
8
Data from EU Kids online project—UK report20
19. The MBG describes this data as only indicative of the effectiveness of filtering. It is not clear that content
that “makes me feel sad, frightened or embarrassed” equates to harmful or inappropriate content. Children
could be seeing age inappropriate (ie by reference to accepted standards of taste and decency) content that
does not make them feel such emotions, and vice versa.
20. Children could also be seeing content on their mobiles that is not captured by the filtering systems—
such as text and picture messages that are sent directly between individual customers or content that is accessed
over networks outside the control of the mobile operators, such as domestic and public wi-fi. However, the
providers of public and domestic wi-fi have made considerable progress recently on the provision of filters to
cover all devices in public spaces and in the home and this should go some way to mitigating the risk of
children accessing age inappropriate content on mobile devices through any network.
21. It is encouraging to note from the data that there is a degree of consistency between the two and that for
both Internet and particularly mobile, the incidence of age inappropriate content being seen is far lower than
for the regulated TV market. Anecdotally, from the hundreds of millions of contacts, both in store and on helplines, between customers and the mobile operators’ across a very broad range of topics, only a small handful
are raised in a situation where a customer has expressed concern about exposure to age inappropriate content.
22. The overriding impression is that children are being appropriately and effectively protected by the
measures being taken.
Illegal Content
23. Mobile operators work with the Internet Watch Foundation (IWF) and law enforcement agencies to deal
with the reporting of illegal child abuse imagery. If illegal content were to be identified on servers hosted by
19
20
http://stakeholders.ofcom.org.uk/binaries/research/media-literacy/oct2012/main.pdf
http://eprints.lse.ac.uk/33730/7/Risks%20and%20safety%20for%20children%20on%20the%20internet%20%20full%20report%20%28LSERO%29.pdf
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 88 Culture, Media and Sport Committee: Evidence
a mobile operator, including web or messaging content, it would be removed quickly, in accordance the relevant
notice and take-down provisions. It is very rare to receive such a notice.
24. Mobile operators receive the IWF’s list of URLs containing potentially illegal content and, in order to
protect customers from being exposed to such content (and thus committing an offence), block the Internet
browser from accessing any URL on the list. Any customer, inadvertently or otherwise, attempting to access
such a site would be directed to a landing page explaining the reason that the page had been inaccessible (using
one of the standard texts endorsed by the IWF).
Malicious Communications
25. Under the Code, the mobile operators commit to procedures for dealing with malicious communications
that occur on their services (ie not third party platforms and services) and are reported to them. Mobile
operators ensure that customers have ready access to mechanisms for reporting to them concerns (such as to a
specialist nuisance calls desk). It is an offence to misuse electronic communications services and, where
requested by law enforcement agencies, mobile operators can and do assist with the compilation of evidence
against perpetrators.
Unsolicited Communications (Spam Text Messages)
26. For a number of years, mobile operators have provided customers with a short code 7726 (the numbers
for S-P-A-M on an alphanumeric keypad) to which unwanted texts can be forwarded. This provides vital early
intelligence to MNOs, who can investigate the content and source of the messages.
27. In the current year, under the auspices of the global mobile trade body, the GSM Association, the UK
mobile operators and the Information Commissioner have signed an MOU with a view to sharing intelligence
about 7726 reports. This enhanced intelligence sharing is enabled through a trial of the new GSMA Spam
Reporting Service.
28. In short, the platform provides a range of tools to analyse data being sent into the 7726 shortcodes of
the UK mobile networks. Approximately 10–15k reports are received each week, resulting in approximately
500 SIM cards being disconnected weekly by the mobile networks.
29. A more detailed memorandum on this topic has been submitted to the Committee in relation to its
inquiry into Nuisance calls and texts.
Information and Advice
30. Mobile operators provide extensive advice to customers—including children, parents and carers—across
a broad range of topics relating to the use of mobile devices and services, such as: privacy, security, Apps,
billing, cost control, and all the matters covered by the content code. Mobile operators, through selected
partnerships and other arrangements also support relevant media literacy activities by others which are designed
to improve the knowledge of consumers in this area.
31. The mobile operators have demonstrated their commitment to customer safety by investing significantly
in filtering and age verification processes over the last ten years—both in terms of capital expenditure and
operating costs. We believe that our approach has been effective. But technical solutions are not foolproof and
it remains essential that all channels of communication are used to raise awareness among parents and children
of how best to stay safe online and how to behave appropriately.
Device Level Filtering
32. Beyond matters that are within the Code (and the remit of the MBG), the mobile operators would like
to see manufacturers continue to develop and improve the filtering and control mechanisms that are available
on portable mobile devices such as smart phones.
33. At present, some of the filtering tools available can be quite blunt and may only offer an “on” or “off”
ability to browse the Internet. With the increasing sophistication of operating systems for mobile devices, it
should become possible for manufacturers to install software that will aid filtering and support more
comprehensive solutions.
34. Software that enables the consumer to make more decisions about what the device can or cannot do,
particularly when browsing the Internet, would enable them to have a much clearer sense of the level of
protection they enjoy, and a greater ability to tailor protection to their own specific needs, independent of the
network through which they are connected.
35. The mobile operators have had network filtering in place for nearly ten years. Great pressure has also
recently been put on the domestic ISPs and public wi-fi operators to do the same—and this is happening.
However, all these efforts would be complemented with the availability of better user controls at operating
system and device level. The UK, through UKCCIS and other channels, should continue to examine closely
what the manufacturers and operating system providers are offering in the area of child safety and challenge
them to be as equally committed as the network providers.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 89
Digital Skills for Life
36. With the Government’s focus on digital skills, and the move towards the online delivery of public
services, it is important that children and adults are not afraid to use the Internet, and are equipped with the
skills they need to become confident web users.
37. The Internet makes available a wealth of information, the lack of which can put people at a significant
disadvantage in areas such as finding employment, housing, education, and making cost savings. With an
estimated 7 million people still not using the Internet, with fear being one of the strongest reasons given for
not going online, it is important that Government, industry and the media be supportive of these groups to
encourage them to get online.
38. The mobile networks welcome the Government’s changes to the education curriculum to enable children
to learn, from an early age, digital skills, including coding, from primary school. It is vital, though, that online
safety and wellbeing is taught as well, as part of the curriculum and across subjects. Teachers of all disciplines
must be trained to teach the relevant skills and appropriate on-line behaviour and young people must be
equipped to understand their responsibilities and how to respond if they ever feel uncomfortable whilst online.
39. There are also economic considerations. Research recently commissioned by Telefonica UK (O2) has
estimated that the unused digital skills among the one million unemployed young people would be worth £6.7
billion21 to the UK economy. Those born into the digital generation will have strong digital skills including
web design, coding and social media expertise. While business must do more to realise these benefits, these
skills must be coupled with appropriate skills for staying safe and behaving responsibly while online.
40. The mobile operators therefore remain committed to maintaining the constructive cross-stakeholder work
with Government, regulators, law enforcement agencies, charities and other interested parties to keep the UK
at the forefront of the Internet safety agenda.
September 2013
Written evidence submitted by Facebook
Summary
1. This submission sets out Facebook’s policies and actions around safety to help inform the Committee’s
inquiry into these matters.
2. Facebook’s mission is to make the world more open and connected and to give people the power to share.
Facebook is a global community of more than 1.15 billion people and hundreds of thousands of organizations.
Each person and organization that uses Facebook represents unique opinions, ideals and cultural values. With
this immense diversity, we work to foster and safe and open environment where everyone can freely discuss
issues and express their views, while respecting the rights of others.
3. The policies Facebook has adopted are designed to reflect real world interactions. While ignorance still
exists both on an off of Facebook, we believe that ignorance will not be defeated by covering up its existence,
but rather by confronting it head on.
4. We have learned that requiring people to engage in conversations and share their views using their real
names and identities promotes an environment of accountability, where contributors must take responsibility
for their own thoughts and actions.
5. Facebook’s detailed Statement of Rights and Responsibilities (“SRR”) describes the content and behaviour
that is and is not permitted on our service. With respect to safety, our SRR specifically prohibits the following
types of behaviours:
(a) Bullying, intimidating, or harassing any user.
(b) Posting content that: is hate speech, threatening, or pornographic; incites violence; or contains
nudity or graphic or gratuitous violence.
(c) Using Facebook to do anything unlawful, misleading, malicious, or discriminatory.
6. Further, Facebook encourages people to report content that they believe violates our terms. We have
“Report” buttons on every piece on content on our site. When we receive a report, we have a dedicated team
of professionals that investigate the piece of content in question. If the content in question is found to violate
our terms, we remove it. If it does not violate our terms, then we do not remove it. We also take action, such
as disabling entire accounts (eg of trolls) or unpublishing Pages, if deemed necessary.
7. We want everyone on Facebook to feel well-equipped to keep themselves safe when using the service. In
particular, we have focused on educating our teenage users, their parents and teachers. We understand that
younger users warrant additional protection and we work in partnership with external audiences to educate
them on our safety tools.
21
http://news.o2.co.uk/?press-release=young-peoples-digital-skills-valued-at-6–7billion-by-uk-businesses
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 90 Culture, Media and Sport Committee: Evidence
How we Combat Child Sexual Exploitation
8. Facebook has a zero tolerance policy for child exploitation and abuse and we fight against these activities
aggressively. We employ advanced technology to protect minors on our site, and we deploy innovative,
industry-leading measures to prevent the dissemination of child exploitation material. We have also built
complex technical systems that either block the creation of this content altogether, including in private groups,
or flag it for immediate review by our safety team.
9. For instance, in collaboration with Microsoft and the National Center for Missing and Exploited Children
(“NCMEC”), we utilize a technology called PhotoDNA that allows us to instantaneously identify, remove and
report known images of child exploitation content to NCMEC. PhotoDNA is a game-changing technology that
has helped enormously in preventing the sharing of abusive materials on Facebook and other services. Once a
piece of content is reported to NCMEC, NCMEC then coordinates with law enforcement authorities around
the world to investigate and prosecute people who are creating and sharing these images online.
10. In the rare instance where child exploitation content is reported or identified by our users on Facebook
(Ie, if the content is a new image of abuse which is not already in our database of known images), we similarly,
take it down as soon as possible and report it to NCMEC.
11. All related reports for the UK are referred to CEOP. CEOP is then able to evaluate each case and where
appropriate, engage local police forces for further investigation and subsequent prosecution.
12. Additionally, we provide a dedicated escalation channel for the IWF and other Internet hotlines to inform
us of illegal images being shared on Facebook,. It is a real testament to the effectiveness of our counter
measures, and particularly PhotoDNA’s technology, that very few reports of such content are received from
these hotlines in any 12-month period,
13. We also work hard to identify, investigate and address grooming behaviour—direct communication by
an adult with a minor with the objective of illegal sexual contact. We work closely with law enforcement,
external partners, expert academics and the Facebook community itself, to spot grooming and to combat it.
How we Keep People Safe
Handling reports
14. We have a comprehensive and well-resourced User Operations (“UO”) team that services all of our users
twenty-four hours each day, seven days a week. This team handles every report that is submitted to Facebook
(including by people who do not have a Facebook account). Hundreds of employees work on the User
Operations team, which is located in four offices across the globe, namely Hyderabad (India), Dublin (Ireland),
Austin (US) and Menlo Park (US). They handle reports in over twenty-four languages and cover every time
zone. Structuring the teams in this manner allows us to maintain constant coverage of our support queues for
all our users, no matter where they are in the world.
15. In order to effectively review such reports, UO is separated into four specific teams, which review
different report types—(1) the Safety team, (2) the Hate and Harassment team, (3) the Access team, and (4)
the Abusive Content team. When a person reports a piece of content, depending on the nature of the report, it
will be directed to the appropriate team. For example, if you are reporting content that you believe contains
graphic violence, the Safety Team will receive and assess the report.
16. If one of these teams determines that a reported piece of content violates our policies or our Statement
of Rights and Responsibilities, we will remove it and warn the person who posted it. In addition, we may also
revoke a user’s ability to share particular types of content or use certain features, disable a user’s account, or
if need be, refer issues to law enforcement. We also have special teams dedicated to handle user appeals for
instances where users feel Facebook might have made a mistake in taking a specific action. We recently
published an infographic, which shows the processes involved in Facebook’s user operations for handling
reports.22
17. Further, we provide all users with a tool, “The Support Dashboard,” which is designed to give a user
much better visibility and insight into the reports they make on Facebook. The Support Dashboard enables
people to track their submitted reports and informs them about the actions taken by our review team. We think
this will help people better understand the reporting process and will educate them about how to resolve their
issues in the future. As people see which of their reports result in a removal of content, we believe users will
be better equipped to make actionable reports. We posted about the Support Dashboard when we unveiled it
in April 2012, on our Safety Page.23
Keeping young people safe
18. We provide educational materials through our Family Safety Centre24 which provides information for
anyone interested in keeping children safe online. The Centre includes specific guidance for teenagers and also
provides information about our global Safety Advisory Board.
22
23
24
https://www.facebook.com/notes/facebook-safety/what-happens-after-you-click-report/432670926753695
https://www.facebook.com/notes/facebook-safety/more-transparency-in-reporting/397890383565083
http://www.facebook.com/safety
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 91
19. Our Safety Advisory Board is comprised of five leading Internet safety organizations from North America
and Europe. These organizations serve in a consultative capacity to Facebook on issues related specifically to
online safety. The members of our Safety Advisory Board include the following:
—
Childnet International.
—
National Network to End Domestic Violence.
—
Connect Safely.
—
The Family Online Safety Institute.
—
WiredSafety.
20. We maintain a Facebook Safety Page,25 which has been “liked” by over one million people. All these
fans can therefore see the latest information on safety education directly in their Facebook Newsfeeds. We
regularly post information, tips, and articles about safety on Facebook and highlight debates on the topic of
digital citizenship, as well as links to useful content from third-party experts.
Under 13s
21. It is well understood that, like many other online services, Facebook’s SRR requires users to be 13 to
sign up for a Facebook account. We therefore require all users to provide their date of birth when signing up
for Facebook to ensure that people under 13 are not able to open an account.
22. However, we are well aware of different studies in the UK and elsewhere that demonstrate how many
under-13s are falsifying their ages to open accounts, often with their parents’ knowledge and even their help.
There is no fail-safe way of addressing the issue of under-age users and we continue to explore best approaches
to the issue with policymakers, our Safety Advisory Board, and our safety partners.
UK partnerships and schools
23. We work with a range of expert partners in the UK who go into schools to provide training and guidance
for teachers, pupils and parents. We take a partnership approach because we have found it to be particularly
effective. Organisations like Childnet, the South West Grid for Learning, the Diana Award (Anti-Bullying),
Childline and Parentzone are trusted safety and advice brands, who are able to help schools and parents
navigate the great variety of online services used by children including Facebook—from online games to
mobile devices, from Twitter to Snapchat and Ask.fm.
24. We particularly support the work of Childnet and the South West Grid for Learning, which runs the
InSafe hotline for professionals in education, and work directly in schools. In recent years, we have funded the
production of thousands of Facebook guide booklets, which these organizations distribute widely. In November
2012 we launched an anti-bullying initiative in partnership with Childline, which encouraged bystanders to
take action when they saw others being bullied.
25. We participate and support the annual “Safer Internet Day” run by the UK Safer Internet Centre,
which comprises these two organisations and the IWF.
26. Facebook was the principal sponsor of The Festival of Education at Wellington College in June 2013.
The Festival is the biggest gathering of its kind in the UK with over 2,500 delegates attending. Over 500
different schools, colleges and universities were represented at the Festival. In addition to these delegates, more
than 600 pupils from over 50 schools took part. At the Festival, the Education Foundation partnered with us
to launch an updated Facebook Guide for Educators.26 This guide provides up-to-date advice on privacy and
safety topics as well as a general introduction to Facebook. The guide features two pilot programmes where
Facebook has been used for learning in two different schools, the Wellington College and the Nautical School
in London. A short film summarising the work was premiered at the Festival (see link on the Education
Foundation website). A copy of the Facebook Guide was provided to every attendee at the Festival and is
also available to download for free from the Education Foundation website.
27. We will continue to partner with these and other relevant organisations over the coming months and
years to help equip young people, teachers and parents with the best information and tools they need to promote
online safety and healthy digital citizenship.
September 2013
25
26
http://www.facebook.com/fbsafety
http://www.ednfoundation.org/2013/06/21/facebook-guide-for-educators/
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 92 Culture, Media and Sport Committee: Evidence
Written evidence submitted by Twitter
Introduction
1. Twitter is an open, public platform built around small bursts of information, which we call Tweets. Each
Tweet can be no more than 140 characters long. In addition to text, Tweets can include photographs, videos,
or links. Users can control their experience on the platform by choosing who to follow and what they wish
to see.
The Twitter Rules
2. Twitter provides a global communication service which encompasses a variety of users with different
voices, ideas and perspectives. With 200 million active users across the world and 15 million in the UK alone,
the platform now serves 500 million tweets a day. Like most technology companies we are clear that there is
no single silver bullet for online safety, rather it must be a combined approach from technology companies,
educators, governments and parents to ensure that we equip people with the digital skills they will need to
navigate the web and wider world going forward.
3. As a general policy, we do not mediate content. However, there are some limitations on the type of content
that can be published with Twitter. These limitations comply with legal requirements and make Twitter a better
experience for all. These limitations include prohibitions on the posting of other people’s private or confidential
information, impersonation of others in a manner that does or is intended to mislead, confuse, or deceive others,
the posting of direct, specific threats of violence against others, and trademark and copyright infringement.
4. Our rules and terms of service clearly state that the Twitter service may not be used for any unlawful
purposes or in furtherance of illegal activities. International users agree to comply with all local laws regarding
online conduct and acceptable content.
5. Full details of Twitter’s rules and terms of service be found in our support centre:
https://support.twitter.com/articles/18311-the-twitter-rules
Illegal Content—Child Sexual Exploitation Policy
6. We do not tolerate child sexual exploitation on Twitter. When we are made aware of links to images of
or content promoting child sexual exploitation they will be removed from the site without further notice and
reported to The National Center for Missing & Exploited Children (“NCMEC”); we permanently suspend
accounts promoting or containing updates with links to child sexual exploitation.
7. We have established a line of communication and maintain an ongoing dialogue with the Child
Exploitation and Online Protection Centre, both in relation to its investigative work and also in its important
work around education and awareness raising. Twitter is also a member of the Internet Watch Foundation.
8. We are in the process of implementing Photo DNA into our backend technologies and, like other
technology companies, are engaged with law enforcement and non-governmental organisations on global efforts
to track and eliminate child abuse images online.
9. Twitter is part of the technology task force for Thorn (http://www.wearethorn.org), a cross industry
foundation which aims to disrupt the predatory behavior of those who abuse and traffic children, solicit sex
with children or create and share images of child sexual exploitation. Thorn exists to continue the work started
by the Demi and Ashton (DNA) Foundation in 2009.
10. In the rare instance that a user finds a Twitter account which they believe to be distributing or promoting
child sexual exploitation while using Twitter, they are asked to notify us by sending an email to [email protected].
They are also asked not to Tweet, retweet or repost child sexual exploitation for any reason and rather to report
it to us immediately so we can take steps to remove it.
Online Safety
11. Twitter provides a global communication platform which encompasses a variety of users with different
voices, ideas and perspectives. As stated above, the platform now serves 500 million tweets a day.
12. Our policy is that we do not mediate content or intervene in disputes between users but we do have a
clear set of rules which govern how people can behave on our platform. These rules are designed to balance
offering our 200 million global users, a service that allows open dialogue and discussion all around the world
whilst protecting the rights of others. As such, users may not make direct, specific threats of violence against
others; targeted abuse or harassment is also a violation of the Twitter Rules27 and Terms of Service.28
13. To help users navigate issues they may be confronted with online, we offer a number of articles in our
Safety and Security Centre29 that guide users to make better decisions when communicating with others.
27
28
29
http://support.twitter.com/entries/18311#
https://twitter.com/tos
https://support.twitter.com/groups/57-safety-security#
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 93
14. Over the course of the next months, we will be publishing new content with more information, including
local resources provided by our partners in the safety community across the world. Some of the new content
will be aimed at bystanders—those who may witness abusive behavior, but are not sure what actions to take.
Teaching users to help each other as well as knowing when and how to contact Twitter for help is vital to
Twitter being a safe space for our users.
15. We also work with a number of safety organizations around the world to provide content to our users
that may be useful for navigating problems online, and plan to host more content from them and other safety
experts in future iterations of the site.
16. We often use the analogy of Twitter as a public town square where users are free to speak with and
interact with others; that said, just as in a public town square, there are behaviors that are not allowed.
17. In addition to reporting violations of our Terms of Service and Rules, users can control what content
they see through features like blocking and sensitive media warnings. When our users see or receive an @reply
they don’t like, they can unfollow the author of the Tweet (if they are following the user), which will remove
that user from their home timeline. They can also block the user30 which will mute the communication. When
you block a user you will not be notified when they mention you, retweet you, favourite your content, add you
to a list or subscribe to one of your lists. You also won’t see any of these interactions in your timeline. This
prevents you from receiving unwanted, targeted and continuous @replies on Twitter. Abusive users often lose
interest once they realize that their target will not respond.
18. Another option available to our users is to protect their accounts.31 Users with protected accounts can
approve requests from other users to follow their accounts on a case-by-case basis. Additionally, their tweets
are only viewable and searchable by themselves and their approved followers. As such, they can prevent any
unwanted followers from viewing their content.
19. Sometimes users see content they don’t like in the form of images or video, and for that we have settings
that allow users to label their media for the appropriate viewers, and select whose media will display on their
own Twitter homepage. We ask users to mark their Tweets as sensitive if they contain media that might be
considered sensitive content such as nudity, violence, or medical procedures.
20. For the viewer, the default setting is that if a Tweet is marked as containing media that might be sensitive,
they will be required to click through a warning message before that media is displayed to them.
21. If another user notices that Tweets have not been marked appropriately, that user may flag the image or
video for review. The content will then be reviewed and a determination made as to whether that media requires
an interstitial in order to comply with Twitter’s Media Policies.32
22. We are continuing to expand our user support and safety teams to ensure we are supporting our growing
userbase appropriately. We also continue to invest heavily in our reporting system from a technological
perspective and recently rolled out a significant update which simplified our system, allowing people to report
violations33 to us via an in-Tweet reporting button as well as via the forms in our support centre.34
23. Once a user files a report to Twitter, they are sent a ticket number and the form is routed to a team of
trained reviewers. If accounts are found to be acting in way which breaks our rules, for example posting
specific threats, then we take action, up to and including suspending accounts. At the end of the process the
specialist reviewer follows up with the person who filed the ticket to let them know that their report has
been addressed.
24. Reports that are flagged for threats, harassment or self-harm are reviewed manually. We also use
automated tools to help triage reports filed for other types of complaints, such as spam.
25. It is important to stress that where a user believes the content or behavior he or she is are reporting is
prohibited by law or if they are concerned for their physical safety, we advise them to contact local law
enforcement so they can accurately assess the content or behaviour. If Twitter is then contacted directly by the
Police, we can work with them and provide assistance for their investigation.
26. Twitter works closely with UK law enforcement. We have updated our existing policies around handling
non-emergency requests from UK law enforcement to more closely mirror our process for US-based law
enforcement requests (including providing notice to users). We have always responded to emergency disclosure
requests from law enforcement in situations involving danger of death or serious physical injury to any person.
27. Twitter invokes these emergency disclosure procedures in the United Kingdom if it appears that there is
an exigent emergency involving the danger of death or serious physical injury to an individual. Twitter may
provide UK law enforcement with user information necessary to prevent that harm.
30
31
32
33
34
http://support.twitter.com/articles/117063#
https://support.twitter.com/articles/14016-about-public-and-protected-tweets#
https://support.twitter.com/articles/20169199-twitter-media-policy#
https://support.twitter.com/articles/15789-how-to-report-violations#
https://support.twitter.com/forms
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 94 Culture, Media and Sport Committee: Evidence
28. Twitter has published Guidelines for Law Enforcement (“Guidelines”; https://t.co/leguide) as well as a
webform (https://t.co/leform) through which UK Law Enforcement can file a request or make an inquiry
29. Since opening our UK office, Twitter has worked to build relationships with stakeholders in the online
safety sector. Twitter is a member of UKCCIS. The company has also established a single point of contact for
the UK Safer Internet Centre and has worked with the South West Grid for Learning.
30. We participated in Safer Internet Day for the first time in 2012 implementing our “Tweets for Good”
programme whereby we promoted safety messaging on the platform. We will continue this involvement and
our work with our safety partners and wider voluntary organisations into the future.
September 2013
Written evidence submitted by Ofcom
Section 1
One page summary
1.1 Ofcom welcomes the opportunity to contribute to the Committee’s inquiry on online safety.
1.2 This document sets out Ofcom’s current duties and powers relevant to the Committee’s inquiry. It also
describes what we consider to be some relevant market and consumer developments captured by our research
that may assist the Committee.
1.3 The Committee’s inquiry considers three specific issues relating to online safety across a broad range of
media services. In order to identify how we may be able to assist the Committee in future, we have structured
our response to consider each of the three issues in turn; and for each we have described Ofcom’s relevant
duties and/or role across the following services:
— Broadcast television and radio.
— ATVOD notified on-demand services.
— Online content more broadly.
Main conclusions of this document:
1.4 The internet has brought significant benefits to consumers and citizens and made the media a much more
social, global and participative environment. However, the internet also brings new regulatory challenges. The
connected environment and changes in consumer behaviour can result in audience protection risks and in
particular, new challenges in securing the protection of children from potentially harmful content.
1.5 In a connected society there is particular pressure on consumers’ understanding of the increasingly
complex regulatory environment, of the practices they can expect of service providers and of their awareness
of their own responsibilities.
1.6 There is a task for the Government, regulators and industry to ensure the provision of clear information,
education and a framework of personal responsibility through which individuals and families can exercise an
informed choice.
1.7 Any approach to online safety should include a combination of the traditional content standards
protections and industry-led measures which provide the right framework of information, online safety tools
and content controls for audiences to protect themselves and their families.
1.8 It is important for consumers that online safety tools and content controls are easy to use and made
available in as many internet connected devices as possible. Consumers should also be adequately informed of
where such tools can be used.
1.9 The effectiveness and impact of online safety tools and content controls will, however, depend on actions
being taken by individuals and parents. Therefore, in order to ensure potentially harmful content is subject to
comprehensive oversight, further consideration may be needed as to the roles and responsibilities of all industry
players in the online environment.
Section 2
Regulatory context
2.1 As the United Kingdom’s independent regulator for the communications sector, Ofcom’s principal duty
in carrying out our functions (set out in section 3(1) of the Communications Act 2003) is:
(b) to further the interests of citizens in relation to communications matters; and
(c) to further the interests of consumers in relevant markets, where appropriate by promoting
competition.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 95
Television and Radio
2.2 In carrying out this duty, Ofcom has other statutory duties and powers relevant to the Committee’s areas
of inquiry. In particular, Section 3(2) of the Communications Act 2003 states that Ofcom is required to secure:
— The application, in the case of all television35 and radio services, of standards that provide
adequate protection to members of the public from the inclusion of offensive and harmful
material in such services; and
— The application, in the case of all television and radio services, of standards that provide
adequate protection to members of the public and all other persons from both:
— unfair treatment in programmes included in such services; and
— unwarranted infringements of privacy resulting from activities carried on for the purposes
of such services.
2.3 All of Ofcom’s decisions must comply with the Human Rights Act 1998 which incorporates into UK
law the European Convention on Human Rights. Of particular relevance to Ofcom’s decisions are: the right to
a fair trial (Article 6), the right to privacy (Article 8) and the right to freedom of expression (Article 10). We
are also subject to the Freedom of Information Act 2000.
2.4 Ofcom is also required to set standards in relation to TV and radio advertising.36 Ofcom established a
co-regulatory partnership with an industry body, the Advertising Standards Authority (“ASA”), in 2004. The
arrangements are underpinned by an enabling statutory instrument37 and a Memorandum of Understanding.38
2.5 Although Ofcom has devolved the exercise of this function to the ASA (in this case, through an Order
made under the Deregulation and Contracting Out Act 1994), it remains ultimately responsible for ensuring
that broadcasters observe relevant standards.
2.6 Television services are provided over different platforms: digital terrestrial wireless transmission, via
satellite, cable or over the internet. Under Article 2 of the AVMS Directive, a broadcast service will fall within
the UK’s jurisdiction if it is either based in the UK, either editorially or by other measures. If a service has no
presence in the UK or EU then secondary technical criteria are used to determine jurisdiction, which includes
the location of the satellite uplink.
Video on Demand
2.7 The Communications Act makes provisions for the regulation of on-demand programme services
(ODPS), which are essentially services whose principal purpose is the provision of programmes the form and
content of which are comparable to the form and content of programmes normally included in television
services, ie TV-like video on demand (VOD) services. These services can be made available on any platform
and are subject to a notification scheme if the editorial control of the service is generally based in the UK.
Notified ODPS must comply with minimum content standards under the AVMS Directive, which has been
implemented in the UK by Part 4A of the Communications Act 2003.
2.8 Ofcom has formally designated the Authority for Television On Demand (ATVOD) as the co-regulator
for editorial content,39 and the ASA as the co-regulator for advertising content. Ofcom remains ultimately
responsible for ensuring that providers of on-demand services observe relevant standards.
Online
2.9 Our role in relation to internet services is much more limited. We regulate television channels delivered
over the internet and notified ODPS when they are established in the UK; but we have no statutory powers to
regulate any other online content. Ofcom also has media literacy duties which address media consumption
online. These are detailed below, along with some information on further roles Ofcom conducts in relation to
online services.
2.10 Section 11 of the Communications Act 2003 places a responsibility on Ofcom regarding the promotion
of media literacy. We fulfil this duty through the publication of Media Literacy research into adults’ and
children/parents’ media use and attitudes. We publish two substantial annual reports: “Children and Parents:
Media Use and Attitudes” and “Adults Media Use and Attitudes”. These provide detailed evidence of media
use, attitudes and understanding among adults and among children and young people aged 3–15. The children’s
report also provides evidence of parents’ concerns about their children’s media use and the ways that they seek
to monitor and mediate that use. Ofcom also publishes a range of other consumer research, exploring similar
35
36
37
38
39
Under 232 (2) of the Communications Act 2003, a service falls within a “television licensable content service” if itE+W+S+N.I.
This sectionnoteType=Explanatory Notes has no associated (a) is provided (whether in digital or in analogue form) as a service
that is to be made available for reception by members of the public; and (b) consists of television programmes or electronic
programme guides, or both.
See section 321 Communications Act 2003.
The Contracting Out (Functions Relating to Broadcast Advertising) and Specification of Relevant Functions Order 2004.
http://stakeholders.ofcom.org.uk/binaries/consultations/reg_broad_ad/statement/mou.pdf.
http://stakeholders.ofcom.org.uk/binaries/broadcast/tv-ops/designation180310.pdf.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 96 Culture, Media and Sport Committee: Evidence
themes. We share the findings of our research widely, including with the Government, industry, academia and
the third sector.
2.11 Ofcom also has a seat on the Executive Board of the UK Council for Child Internet Safety (UKCCIS)40
which brings together more than 200 organisations across the government, industry, law enforcement, academia
and charity sectors, who work in partnership to help keep children safe online. Ofcom provides its research
data to UKCCIS to help inform debates about these areas and has been involved in a number of UKCCIS
initiatives, for example: the “Advice for Child Internet Safety 1.0—universal guidelines for providers”41; and
the “Good practice guidance for providers of social networking and other inter-active services”.42
2.12 In addition, the Government has recently proposed, through its published communications strategy
paper “Connectivity, Content and Consumers—Britain’s digital platform for Growth”,43that it expects industry
across the internet value chain “to ensure that all internet-enabled devices, whether a TV, a games console or
a smart phone, are supplied with the tools to keep children safe as a standard feature”. The Government has
asked Ofcom to have a role in overseeing this work and we are currently awaiting further details on this.
2.13 Ofcom was also significantly involved in developing ParentPort, which is a website with the single aim
of protecting children by helping parents make their views heard about inappropriate programmes, advertising,
products and services.44 The website makes it easier for parents to complain about material they have seen or
heard across the media, communications and retail industries.
Section 3
Market context and consumer context
3.1 Ofcom carries out research which informs the delivery of our duties and our programme of work. It also
keeps us, and others, informed about new technology developments and the impact these may have on the
sectors that we regulate. This section provides a brief overview of Ofcom’s observations on the effects of
convergence on the way consumers and industry engage with media, and the challenges this may create
for audiences.
3.2 Ofcom publish three reports that are of particular relevance to the Committee’s inquiry:
—
The UK and Nations’ Communications Market Report,45 last published in August 2013;
—
Ofcom annual report “Children and Parents: Media Use and Attitudes”46; and
—
Ofcom annual report “Adults Media Use and Attitudes”.47
3.3 Below we have set out a summary of these reports and you can find the full findings using the links
provided above. Ofcom would also be pleased to provide a deeper and wider ranging discussion of our evidence
base should the Committee require it.
Technology continues to evolve rapidly, resulting in an increasing range of new connected devices and
changing consumption of AV content services
3.4 UK household internet access rose to 80% in Q1 2013, just one percentage point higher than in Q1
2012. However, mobile internet access rose ten percentage points to 49% of adults, the second fastest growth
on record.
40
41
42
43
44
45
46
47
http://www.education.gov.uk/childrenandyoungpeople/safeguardingchildren/b00222029/child-internet-safety
http://media.education.gov.uk/assets/files/ukccis%20advice%20on%20child%20internet%20safety.pdf
http://media.education.gov.uk/assets/files/industry%20guidance%20%20%20social%20networking.pdf
https://www.gov.uk/government/uploads/system/uploads/attachment_data/file/225783/Connectivity_Content_and_Consumers_
2013.pdf
www.parentport.org.uk. ParentPort has been jointly developed by the Advertising Standards Authority (ASA), the Authority for
Television On-demand (ATVOD), the BBC Trust, the British Board of Film Classification (BBFC), Ofcom, the Press Complaints
Commission (PCC) and the Video Standards Council (VSC)/Pan-European Game Information (PEGI).
http://stakeholders.ofcom.org.uk/market-data-research/market-data/communications-market-reports/cmr13/
http://stakeholders.ofcom.org.uk/market-data-research/media-literacy-pubs/
http://stakeholders.ofcom.org.uk/market-data-research/media-literacy/media-lit-research/adults-2013/
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 97
Figure 3.1
HOUSEHOLD INTERNET ACCESS: 2005–2013
Proportion of adults (%)
100
Internet
80
60
64
60
73
68
65
67
60
75
77
79
80
76
72
75
71
65
74
67
58
52
40
42
38
41
27
30
21
32
20
12
15
17
31
20
72
50
49
Total
broadband
Fixed
broadband
Mobile data
user
39
5
13
0
2005 Q1 2006 Q1 2007 Q1 2008 Q1 2009 Q1 2010 Q1 2011 Q1 2012 Q1
Internet on
mobile
Mobile
broadband
Source: Ofcom technology tracker, Q1 2013. Base: All adults aged 16+ (n=3750).
3.5 The power and functionality of devices connected to the internet are making them increasingly attractive
to consumers, especially the scope that they give consumers to access, create and share content across different
devices. The majority of devices continue to grow in terms of penetration.
3.6 There has been a particularly remarkable growth in the use of tablets. In spite of their relative novelty,
in February 2013 tablets accounted for 8% of web-page views, double the level in 2012. Tablets are viewed as
the main method of connecting to the internet by a third of users and just over half (56%) of tablet owners use
their device for watching audio visual content.48
There are increasing levels of take-up and usage of internet delivered services
3.7 There continues to be a demographic split in use of the internet. Younger people are more likely than
older people to use the internet for activities related to converged media and this becomes more marked the
older the age bracket. Conversely, younger people are more likely to access their media simultaneously meaning
that, while they are consuming a greater volume of media than older people, they are doing so in a shorter
time-period.49
Figure 3.2
WEB-BASED COMMUNICATION ON COMPUTERS, MOBILE PHONES, AND BOTH, AMONG 16–24
YEAR OLDS
% of 16-24 year olds using web-based communication at least daily
with friends and family on computers, mobile phones and on both
100
80
61
60
40
20
49
Computer
51
39
Mobile
33
36
26
30
27
17
16
13
1517
13
9
11 7
9
0
Source: Kantar Media Omnibus
Base: All 16–24s in UK (N=375)
Q.2A How often do you use [insert statement] to communicate with friends and family?
48
49
Ofcom Communications Market Report 2013. 51.
http://stakeholders.ofcom.org.uk/binaries/research/811898/consumers-digital-day.pdf
Both
15
5
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 98 Culture, Media and Sport Committee: Evidence
Computers were defined as desktop PC, laptop, netbook or tablet
3.8 There is also evidence of growth in online video consumption. YouTube in particular has experienced
strong growth in the last year with three quarters (74%) of laptop and desktop internet users visiting YouTube.
Its unique audience grew 4% in the year to May 2013.50
3.9 One of the key observable trends in web content consumption has been the rise of social networking, in
particular among young people. Three-quarters of 15–24 year olds use social networking sites.
Figure 3.3
PROPORTION OF ADULTS WHO ACCESS SOCIAL NETWORKING SITES AT HOME
Adults 16+ (%)
77
80%
69
60%
45
49
49
52
48
41
40%
37
42
24
20%
12
2
0%
UK
15-24 25-34 35-54 55-64 65-74 75+
AB
C1
C2
DE
Male Female
Source: Ofcom Consumer Research Q1 2013
Base: All adults aged 16+ (n = 5812 Q1 2008, 1581 Q3 2008, 6090 Q1 2009, 9013 Q1 2010, 3474
Q1 2011, 3772 Q1 2012, 3750 Q1 2013) QE5. Which, if any, of these do you use the internet for?
NB Question wording for QE5 prior to 2013 asked about household use of the internet at home.
In 2013 QE5 asked about individual use of the internet anywhere.
Consumers continue to have a high expectation for content standards on their television
3.10 In February 2012, Ofcom published a deliberative research report entitled “Protecting audiences in a
converged world”.51 The research looked at public attitudes within the context of convergence, in order to
understand the public’s expectations for protection and how content should be regulated in the future.
3.11 This research indicated that knowledge of prevailing content regulation is currently high for broadcast
services but lower for other services (such as catch-up and VOD). The research also suggests that viewers have
high expectations of content regulation on broadcast television, and associated VOD and catch-up services.
Converged or connected TVs, which incorporate broadcast, VOD and open internet services, are considered to
be closer to a TV-like experience. Audiences therefore expect content over these devices to be regulated more
similarly to broadcast TV, in contrast to internet content accessed through devices such as PCs and laptops.
Parents are adapting the way they protect their children from certain kinds of content in light of
technological change
3.12 Ofcom’s annual reports “Children and Parents: Media Use” and “Adults Media Use and Attitudes”
provide detailed evidence of media use, attitudes and understanding among adults aged 16 and above and
among children and young people aged three—15. Below are some of the key findings from the 2013 Adults52
and Children and Parents research.53
3.13 In 2012 79% of households (aged 16+) had internet access, this is in line with 2011. The use of devices
to go online at home, as seen among children, is diversifying with PC/laptop/netbook use remaining stable at
74%, although there were significant increases seen using other devices to access the internet at home. Mobile
phone54 access increased to 53% (up from 45% in 2011), tablet (16% vs. 6%) and games console (16%
vs. 11%).
3.14 In 2013 91% of children aged five—15 live in a household with access to the internet through a PC,
laptop or netbook and Ofcom’s research shows that children’s patterns of media consumption are changing.
50
51
52
53
54
Ofcom commissioned qualitative research earlier this year to give us a better insight into the complex and diverse phenomenon
of UGC. This provides a snapshot of the different players, types of content, business models and opportunities in this sphere.
http://stakeholders.ofcom.org.uk/binaries/research/research-publications/content.pdf
http://stakeholders.ofcom.org.uk/binaries/research/tv-research/946687/Protecting-audiences.pdf
Adults’ media use and attitudes report 2013 (http://stakeholders.ofcom.org.uk/market-data-research/media-literacy/media-litresearch/adults-2013/)
The Media Literacy research for 2013 is due to be published in October 2013.
Mobile phone access includes smartphones. Personal smartphone use increased in 2012 from 44% to 54%.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 99
Children are also increasingly likely to use a wide range of portable and personal devices, particularly
smartphones and tablets. In 2013 43% of children aged five—15 owned a mobile phone with 62% of 12 -15s
owning a smartphone. Parents’ ability to monitor what their children are doing online is becoming more limited
than it was when internet access took place at a fixed point in the heart of the family home.
3.15 Among adult internet users over half (55%) say they use the same passwords for most websites. A
quarter (25%) of users said they had problems remembering passwords, and a similar proportion (26%) said
they tend to use easy to remember passwords like birthdays or names.
3.16 For adults, take-up of Wi-Fi protection and experience of deleting cookies from the web browser have
increased since 2011. Very few internet users say they do not use online security measures or safety features
because they don’t know how they work, although a significant minority of people are unaware of a range of
security features. Awareness and take-up of the security measures we asked about is typically lower among
older users, DE adults and women.
3.17 43% of parents of children aged five—15 said that they use technical parental controls packages for
desktop PCs, laptops and netbooks.55 Uptake of controls for television is slightly higher at 45%. Use of filters
on mobile phones where the phone can be used to go online is at 42% while uptake for controls on mobile/
fixed games consoles is 16% and 19% respectively.
3.18 Parents generally view controls as an extra tool to supplement rather than a replacement for parental
mediation (rules, monitoring and talking to children). 79% of parents of five—15 year olds have spoken to
their child about staying safe online and just over half of parents have rules in relation to parental supervision.
Overall, 85% of parents employ some kind of online mediation strategy to help their child stay safe online.
3.19 Other findings from the 2013 survey show that in households where five—15s use a PC/laptop or
netbook at home 62% have some kind of technical method of control in place. 44% have safe search settings
on search engine websites; 43% have parental controls56 installed; 19% have the Safety Mode set on YouTube;
11% have software to limit the time spent online; 8% have a PIN/password set on broadcasters’ websites. The
most common reason reported by parents of older children (12–15) for not using access controls was that they
trust their child to be sensible/responsible; and parents of younger children (five—sevens and eight—11s) are
more likely to say it is because their child is supervised, or that the child is too young for a lack of controls to
be a problem.
3.20 The majority of parents (79%) report that they know enough to keep their child safe online, but around
half of parents continue to feel that their child knows more about the internet than they do, including 14% of
parents of children aged three—four. Also, parents of eight—11s are more likely to agree with this statement
now than in 2012 (44% vs. 35%).
Section 4
Protecting minors from accessing adult content
4.1 We have significant experience of regulating content that is unsuitable for children on broadcast television
and radio services. Through our co-regulatory relationship with ATVOD we also have experience of regulating
adult content on notified video on demand services. This section sets out details of some of our experiences of
regulating this type of content on these services, which may be useful for the Committee. While Ofcom has
no statutory role to regulate online services to protect children from accessing adult content, outside of
television channels delivered on the internet and notified ODPS, this section also sets out some relevant research
findings regarding adults’ and children/parents’ media use and attitudes that complements our understanding
of this sector.
Ofcom’s experience of regulating television and radio broadcast services
4.2 Under Section 319(1) of the Communications Act 2003, Ofcom has a statutory duty to set standards for
broadcast content as appear to it best calculated to secure the standards objectives. These standards are set out
in Ofcom’s Broadcasting Code (“the Code”) in the form of rules which broadcasters must comply with. These
standards objectives include: “that persons under the age of eighteen are protected”.
4.3 Ofcom considers the standards it has set for the protection of children, as set out in Section One of the
Code, to be amongst the most important. The Code includes rules which help ensure the protection of minors
from harmful content and from material that is unsuitable for them (covering content such as drugs, smoking,
alcohol, violence and dangerous behaviour, and offensive language).57 It also has rules that specifically relate
to protecting children from sexual material, including prohibiting the broadcast of material equivalent to the
55
56
57
This figure includes all available technical parental controls, for example parental controls provided by the ISP, but also parental
controls built into the computer’s operating system (windows or Mac), and third party parental controls such as net nanny or
family shield.
This includes parental controls provided by ISPs, the computers own operating system, or controls such as McAfee loaded by
other people in the household.
http://stakeholders.ofcom.org.uk/broadcasting/broadcast-codes/broadcast-code/protecting-under-18s/
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 100 Culture, Media and Sport Committee: Evidence
British Board of Film Classification (BBFC) R18 rating and requiring “adult-sex material”58 to be shown only
between 10pm and 5.30am and behind mandatory restricted access.59
4.4 Ofcom has sanctioned a number of adult-chat broadcasters for broadcasting sexually explicit material
without the relevant protections.60 For example, in November 2010 Ofcom imposed a financial penalty on
Bang Channels and Bang Media (London) Ltd totalling £157,250 in respect of multiple breaches of the Code
and licence conditions by the adult chat and daytime programming on these services. Ofcom concluded that
the licensees had been operating a wholly inadequate compliance system which amounted to manifest
recklessness. Ofcom revoked all the licences held by these licensees on the basis that they were no longer fit
and proper to hold those licences.61
Ofcom’s experience of co-regulation for on-demand content
4.5 ATVOD has published Rules and Guidance to ensure compliance of all notified ODPS with certain
minimum standards.62 Rule 11 of the ATVOD Rules reflects section 368E(2) of the Communications Act and
states that, “if an on-demand programme service contains material which might seriously impair the physical,
mental or moral development of persons under the age of eighteen, the material must be made available in a
manner which secures that such persons will not normally see or hear it”.
4.6 ATVOD has adopted a precautionary approach to its interpretation of the wording of the Act and includes
R18 material (or material equivalent to content classified in that category) as “material that might seriously
impair”. Therefore R18 material or R18-equivalent content should only be made available in on-demand
programme services in a manner which secures that persons under the age of eighteen will not normally see
or hear it.
4.7 In the past year Ofcom has imposed financial penalties on three ATVOD notified ODPS for a breach of
Rule 11. These sanctions were imposed on the services “Playboy TV”,63 “Demand Adult”64 and “Strictly
Broadband”65 after these services provided R18 equivalent material without adequate measures in place—a
content access control system—to ensure that those under 18 would not normally see or hear it.
Relevant research findings regarding online services
4.8 Paragraphs 3.12 to 3.20 set out some relevant research findings from Ofcom’s recent Media Literacy
research into adults’ and children/parents’ media use and attitudes, which looks at the behaviour and
experiences of children and parents online.
4.9 Our research indicates that the majority of parents use a combination of mediation strategies to help
keep their children safe online. Over four in 10 parents have parental controls installed, and feel that their
children are safer as a result.
4.10 Most parents report that they know enough to keep their child safe online, but around half of parents
continue to feel that their child knows more about the internet than they do, including 14% of parents of
children aged 3–4. This may reflect the fast pace of technological change and the ease which younger children
in particular can embrace and explore new mobile technology and apps and suggests that parents need
continuing support to build their own skills and confidence.
4.11 Our research also shows that despite the vast majority of young people stating that they are confident
internet users and that they know how to stay safe online, there is an increase in 12–15s undertaking risky
activities online. A substantial minority of 12–15s have a social networking profile which may be visible to
people not known to them and this has increased since 2012 (33% vs 22%). These children are more likely to
be in the 15% who have undertaken some kind of potentially risky online behaviour such as adding people
that they don’t know in person or sending photos or personal details to people only known online.
4.12 Some of these activities may be difficult to manage through technical controls; this reinforces the
importance of developing children’s media literacy skills—to help them analyse and assess content, and manage
the potential risks and unintended consequences of online activity—irrespective of technological changes.
4.13 How to protect children from the risk of exposure to harmful content online is an issue that has been
closely considered by UKCCIS. There are already examples of industry-led approaches to support the
protection of minors from such content. For example: all major ISPs provide customers with parental control
software either for individual devices or for the whole home, via network-layer filtering. All major ISPs
recently committed to offering their customers a network-layer filtering option and search providers offer “safe”
58
59
60
61
62
63
64
65
Material that contains images and/or language of a strong sexual nature which is broadcast for the primary purpose of sexual
arousal or stimulation.
Mandatory restricted access means there is a PIN protected system (or other equivalent protection) which cannot be removed
by the user, that restricts access solely to those authorised to view.
http://stakeholders.ofcom.org.uk/enforcement/content-sanctions-adjudications/?pageNum=1#in-this-section
http://stakeholders.ofcom.org.uk/binaries/enforcement/content-sanctions-adjudications/bangmedia-revocation.pdf
Under the AVMS Directive, which has been implemented in the UK by Part 4A of the Communications Act 2003.
http://stakeholders.ofcom.org.uk/binaries/enforcement/vod-services/Playboy_TV_Sanction.pdf
http://stakeholders.ofcom.org.uk/binaries/enforcement/vod-services/Demand_Adult.pdf
http://stakeholders.ofcom.org.uk/binaries/enforcement/vod-services/Strictly-Broadband.pdf
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 101
versions of their search results, which aim to exclude sexually explicit images and video. Paragraphs 3.17 and
3.19 include data on what technical method of control parents currently use.
4.14 Controlling illegal content and protecting minors from potentially harmful content are fundamental
policy goals which require regulatory intervention, regardless of changes in technology and delivery of content.
4.15 The above findings highlight the importance of ensuring that any protection framework for online
content should include a combination of the traditional standards protections and measures which provide the
right framework for audiences to protect themselves and their families.
4.16 Convergence and the expanding consumption of connected devices will create further pressure on
consumers’ understanding of the regulatory environment, of the practices they can expect of service providers
and of their awareness of their own responsibilities.
4.17 There is therefore a task for the Government, regulators and industry to ensure the provision of clear
information, education and a framework of personal responsibility through which individuals and families can
exercise an informed choice.
4.18 Industry will need to play a vital role in ensuring consumers are provided with clear information,
education and a framework of tools and controls through which they can manage their own and their children’s
access to content. Such controls should be easy to use and find, and potentially available across all connected
devices.
4.19 It is important to recognise, however, that the effectiveness and impact of online safety tools and content
controls will depend on actions being taken by individuals and parents. Therefore, in order to ensure potentially
harmful content is subject to comprehensive oversight, further consideration may be needed as to the roles and
responsibilities of all industry players in the online environment.
Section 5
Preventing extremist and unlawful material
5.1 Ofcom has clear statutory duties to regulate extremist material on UK licensed broadcast services,
including television channels delivered on the internet. We also have statutory powers to regulate such material
for on-demand services notified in the UK. This section sets out some details of our experience of regulating
this type of content, which may be useful for the Committee. Ofcom has no statutory powers to regulate
extremist material on other online services.
Ofcom’s experience of regulating television and radio broadcast services
Material intended to incite hate
5.2 Ofcom has statutory powers to set standards to protect audiences from harm from religious extremism
broadcast on television and radio. Rule 3.1 of the Code states that “material likely to encourage or incite the
commission of crime or lead to disorder must not be included in television or radio services”. When considering
potential breaches of Rule 3.1, a key factor that Ofcom takes into account is whether the content, as it was
presented, contained a call to action which would be likely to encourage or incite the commission of crime
or disorder.
5.3 Ofcom takes the issue of incitement to crime or violence extremely seriously. On the occasions where
this issue has arisen, we have taken robust and appropriate action. For example, we have recently published four
sanction decisions relating to the broadcast of material that was likely to encourage or incite the commission of
crime (including hatred) and/or was abusive towards a religion or belief.66
Unlawful content
5.4 If a broadcast service includes unlawful content, such as child abuse images, this would clearly be a
matter for the criminal justice system. Ofcom would therefore report any such activity and the licensee
responsible to the relevant authorities. Ofcom also has statutory powers to act quickly in response to the
inclusion of any such content.
Ofcom’s experience of co-regulation for on-demand content
Material intended to incite hate
5.5 Rule 10 in the ATVOD Rules reflects section 368E(1) of the Act and states that: “An on-demand
programme service must not contain any material likely to incite hatred based on race, sex, religion or
nationality”.
66
http://stakeholders.ofcom.org.uk/binaries/enforcement/content-sanctions-adjudications/Rehmatul-DM-Digital.pdf
http://stakeholders.ofcom.org.uk/binaries/enforcement/content-sanctions-adjudications/regis1limited.pdf
http://stakeholders.ofcom.org.uk/binaries/enforcement/content-sanctions-adjudications/noor-tv.pdf
http://stakeholders.ofcom.org.uk/binaries/enforcement/content-sanctions-adjudications/takbeer.pdf
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 102 Culture, Media and Sport Committee: Evidence
5.6 To date there have been no cases where ATVOD has breached a notified ODPS under Rule 10.
Unlawful content
5.7 ATVOD’s Rule 11 of the ATVOD Rules reflects section 368E(2) of the Communications Act and states
that, “if an on-demand programme service contains material which might seriously impair the physical, mental
or moral development of persons under the age of eighteen, the material must be made available in a manner
which secures that such persons will not normally see or hear it.”
5.8 ATVOD’s Rule 11 captures content which is illegal, for example criminally obscene or indecent images.
If content of this nature was included in a notified ODPS ATVOD would refer the service to Ofcom
immediately. Ofcom would then immediately seek to suspend the service and notify the relevant authorities.
Ofcom’s role for online content
5.9 Other than the regulatory duties we have set out above, which relate to television channels delivered on
the internet and notified ODPS, Ofcom has no statutory powers to regulate hate material or abusive images
online.
5.10 Illegal material such as child abuse images are within the remit of two bodies in the UK. The Internet
Watch Foundation (IWF) is a self-regulatory regime funded by telecommunications and internet companies,
which works to restrict the availability of child abuse images online; and the Child Exploitation and Online
Protection Centre (CEOP), is the UK Police body dedicated to eradicating the sexual abuse of children. CEOP
tracks and seeks prosecution of offenders, such as those who create, distribute or consume child abuse images.
These bodies work closely together under a formal agreement.
5.11 Extremist material online is a substantial concern and poses significant risk to UK citizens. This will
continue to be a clear regulatory challenge for the future. Industry-led approaches and tools will continue to
play an important role in protecting audiences from such harmful content online.
Section 6
Preventing abusive comments on social media
6.1 Ofcom has experience of regulating abusive content on television and radio services, which may be of
assistance to the Committee, details of which are set out below. While Ofcom has no statutory role to regulate
abusive content, outside of television channels delivered on the internet, this section also sets out some research
findings regarding children’s behaviour and experience of online social media, which may be of both interest
and use to the Committee.
Ofcom’s experience of regulating television and radio broadcast services
6.2 Ofcom’s Broadcasting Code contains rules which aim to protect audiences from abusive content on
broadcast television and radio services. For example:
— Rule 3.1 of the Code states that “material likely to encourage or incite the commission of crime
or lead to disorder must not be included in television or radio services”;
— Rule 4.2 states that “the religious views and briefs of those belonging to a particular religion
or religious denomination must not be subject to abusive treatment”;
— Rule 2.3 states that “in applying generally accepted standards broadcasters must ensure that
material which may cause offence is justified by the context. Such material may include, but is
not limited to...humiliation, distress, violation of human dignity, discriminatory treatment or
language (for example on the grounds of age, disability, gender, race, religion, beliefs and
sexual orientation)”; and
— Rule 7.1 states that “Broadcasters must avoid unjust or unfair treatment of individuals or
organisations in programmes”.
6.3 Ofcom recently sanctioned the licensee Takbeer TV Limited for serious breaches of Rule 4.1
“broadcasters must exercise the proper degree of responsibility with respect to the content of programmes
which are religious programmes” and Rule 4.2 “the religious views and briefs of those belonging to a particular
religion or religious denomination must not be subject to abusive treatment”. In this case the broadcast
contained abusive statements about the Ahmadi Muslim community and beliefs.67
6.4 In addition, Ofcom has recorded a series of breaches of Rule 4.2 of the Broadcasting Code against the
licensee Ummah Channel Limited for broadcasting abusive content on its service, Ummah Channel. This
television channel broadcast programmes which included terms and references when taken together amounted
to “abusive treatment” of religious views and briefs.68
67
68
http://stakeholders.ofcom.org.uk/binaries/enforcement/content-sanctions-adjudications/takbeer.pdf
http://stakeholders.ofcom.org.uk/binaries/enforcement/broadcast-bulletins/obb167/issue167.pdf;
http://stakeholders.ofcom.org.uk/binaries/enforcement/broadcast-bulletins/obb196/obb196.pdf
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 103
Ofcom’s experience of co-regulation for on-demand content
6.5 ATVOD has no rules to regulate abusive content on notified ODPS, ie content that is not considered to
be hate speech or material likely to incite crime, or does not amount to R18 equivalent material.
Relevant research findings of online services
6.6 Our Media Literacy research into adults’ and children/parents’ media use and attitudes indicates that for
the first time there is a decrease in the number of children with social networking profiles and there appears to
be greater diversity in the types of social networking sites being used.
6.7 Other relevant findings include:
— Compared to 2012, children are less likely to know how to block messages from someone they
don’t want to hear from (53% vs. 68%) and to have done this in the past year (32% vs. 42%).
— 19% of 12–15s say they personally have had a negative experience, other than bullying, of
online or mobile phone activity in the past year, where the most likely experience is gossip
being spread (13%).
— Girls aged 12–15 are more likely than boys to have experienced gossip being spread about
them online or through texts (17% vs. 10%) or to have seen/received something troubling online
like a chain letter or something that makes them feels scared (4% vs. 0%).
— Girls aged 12–15 are more likely than boys to say they know of someone who has been bullied
through a mobile phone (33% vs. 20%) and to say they have experienced bullying in this way
(12% vs. 3%).
— Almost one in ten 12–15s and one in twenty-five 8–11s say they have experienced online
bullying in the last year.
— Since 2012, those aged 12–15 with an active social networking profile are more likely to have
a profile on Twitter (37% vs. 25% in 2012).
6.8 The above findings highlight further the importance of providing consumers with clear information,
education and the right framework through which to protect themselves and their families from potentially
harmful content of this kind.
6.9 As already stated, self-regulatory approaches will continue to play an important role in providing this
framework in order to help protect adults and children in an online world. For example, there is increasing
responsibility for industry to provide consumers with effective reporting tools, community standards or terms
of use; and the need for transparent and demonstrably effective procedures for handling complaints and
removing non-compliant content. It should be clear to consumers how to complain or what action to take if
they see or hear content that is harmful or unsuitable for children.
6.10 The research findings also highlight the importance of providing children in particular with the right
tools and media literacy skills to address this issue and understand the potential risks.
October 2013
Written evidence submitted by the Home Office
Introduction
1. This paper sets out the Home Office’s evidence to the Culture, Media and Sport Committee inquiry into
online safety.
2. Ensuring safety online, particularly for children, is a priority for Government. In a speech to the National
Society for the Prevention of Cruelty to Children (NSPCC) the Prime Minister set out a range of measures to
ensure that the cultural challenges of protecting children from accessing potentially harmful or age
inappropriate content, and the challenges of restricting the availability of, and access to, criminal content such
as child abuse material are met..
3. The Home Office’s evidence focuses particularly on the Committee’s concerns about illegal child abuse
images online and illegal extremist materials online. It should be considered alongside evidence submitted by
the Department for Culture, Media and Sport on harmful or age inappropriate content and threatening or
abusive behaviour..
Tackling Illegal Child Abuse Images Online
4. Research by the NSPCC indicates that about 5% of children in the UK suffer contact sexual abuse at
some point during their childhood. Child abuse has many forms, one of which is online and multinational
activity using sophisticated technology and techniques to evade law enforcement attention. The internet
(including the so-called “hidden internet” which facilitates a sense of a “safe” community of offenders) is used
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 104 Culture, Media and Sport Committee: Evidence
for the proliferation of indecent images of children. Work to tackle this between government, law enforcement
and industry is an important part of the Organised Crime Strategy that was published on [7 October].
5. Images of child sexual abuse are illegal. It is illegal to take, make, circulate and distribute child abuse
images (Protection of Children Act 1978)—such offences carry a maximum sentence of ten years’
imprisonment. It is illegal to possess indecent images of children (Criminal Justice Act 1988)—such offences
carry a maximum sentence of five years’ imprisonment. Such offences apply equally online as they would
offline.
6. The Government has worked with industry to develop some of the tightest controls in the world so that
child abuse images are removed from the internet where possible, blocked from being accessed from the UK
where they cannot be removed, and investigated so that those involved in their production, distribution or
possession are brought to justice and so that the victims of these crimes can be protected and safeguarded. The
Child exploitation and Online Protection Command of the National Crime Agency is central to this effort.
7. Tackling such abuse on the internet forms an important aspect of our wider approach to protecting children
and vulnerable people—a key priority of this Government. The Government has established a Sexual Violence
Against Children and Vulnerable People National Group (SVACV). This is a panel of experts and policy
makers brought together by the Home Office to coordinate and implement the learning from recent inquiries
into historic child sexual abuse and current sexual violence prevention issues. Tackling the misuse of the
internet by those seeking to sexually abuse children, or gain personal gratification from it, is an important
aspect of the National Group’s work. On 24 July 2013, the Government published a progress report and action
plan on the work of the SVACV National Group. This contained clear activity already delivered by the
Government to protect children online and an action plan to take forward further activity identified by the
Group. The progress report and action plan can be found at the following link:
https://www.gov.uk/government/publications/sexual-violence-against-children-and-vulnerable-peoplenational-group.
Removing Illegal Child Abuse Images from the Internet and Preventing Access to Them
8. The Internet Watch Foundation (IWF) takes reports of illegal child abuse images from the public. It has
the authority to hold and analyse these images through agreement with the Crown Prosecution Service and the
Association of Chief Police Officers. The IWF assesses images, determines whether they are illegal and, if so,
the severity of the images using Sentencing Council guidelines. Work is ongoing to expand the role of the IWF
so that it can proactively search for child abuse images rather than being reliant on reports from industry and
the public.
9. If the site hosting the image is located in the UK then the details will be passed to law enforcement
agencies and the Internet Service Provider will be asked to take down the website using the “notice and
takedown” process. In 2012, the Internet Watch Foundation found that 56% of UK child sexual abuse images
were removed in 60 minutes or less from the point at which the Internet Service Provider or hosting provider
was notified of the content. IWF Members remove 90% of such content within one hour, and 100% within two
hours. In the past 17 years, over 400,000 webpages have been assessed—100,000 have been removed for
containing criminal content. Less than 1% of child sexual abuse content is now thought to be hosted in the
UK, down from 18% in 1996.
10. Where images are hosted outside the UK, the IWF will pass the details to their equivalent body in the
country identified as hosting it so that they can take action. Often it is taken down completely, and the average
time taken to remove child sexual abuse content hosted outside the UK had reduced to ten days in 2011
compared to over a month in 2008. Until the content is taken down the Internet Watch Foundation adds the
URL to its “URL list” or “blocking list” which IWF Members can use to block access by their customers to
the identified images. Such blocking arrangements currently apply to about 98.6% of domestic broadband lines.
11. The IWF is working with its members to introduce “splash pages”—these are warning messages that
appear if a user attempts to access a webpage that has been removed for hosting illegal child abuse images,
and they deliver a hard-hitting deterrence message to users seeking to access child abuse images.
12. These arrangements have been enormously successful, but the Government is committed to going further
still in order to prevent offenders from being able to access child abuse images through the internet. That is
why the Prime Minister, in his speech of 22 July, announced a further set of initiatives.
13. Internet search engine providers have been asked to go further in restricting access to illegal child sexual
abuse images. They have been asked to do three things:
(a) To work with law enforcement to develop effective deterrence messages to target those users
who try to access child abuse images through their search engines.
(b) To ensure that illegal child sexual abuse images are not returned in search results, and to use
their search algorithms to guide users away from searches that could contain pathways to child
abuse images.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 105
(c) To prevent any search results from being returned when specific search terms are used that
have been identified by CEOP as being unambiguously aimed at accessing illegal child sexual
abuse images.
14. The Government continues to work with search engine providers to put these measures in place. If
progress is not forthcoming then the Government will consider bringing forward legislation to require their
compliance.
15. The objective for all of these actions is to make it more difficult for unsophisticated users to find a route
from open searching to more sophisticated offending environments, make it more difficult for inquisitive nonoffenders to access indecent images of children, and make it less likely that members of the public could
inadvertently come across such images.
Investigating Those Responsible for Illegal Child Abuse Images and Protecting Victims
16. The Child Exploitation and Online Protection Command of the National Crime Agency (CEOP) is the
national lead for tackling sexual exploitation and sexual abuse of children. It is an integral part of the new
National Crime Agency, that was established on 7 October 2013 to lead the UK’s fight to cut serious and
organised crime.
17. CEOP has done excellent work over the last seven years. It is dedicated to eradicating the sexual abuse
of children, and supports both operational work to protect children and educational programmes to help children
protect themselves online. It operates a reporting centre for children who want to report online sexual threats
(in 2012–13 CEOP received an average of 1,600 reports per month of abuse from the public and the internet
industry) a 14% increase on the previous year. This will have included direct reports from children. CEOP
maintains close relationships with a wide range of stakeholders from law enforcement, industry and the
educational sectors nationally and internationally. In 2012–13, CEOP safeguarded and protected 790 children,
an increase of 85% on the previous year, and its work led to the arrest of 192 suspects. CEOP has now
protected more than 2,250 children in its seven-year history.
18. Government funding for CEOP was higher in 2012/13 (£6.381m) than it was in 2009/10 (£6.353m). The
CEOP budget has effectively been protected in cash terms since 2011/12 and there are now more people
working in CEOP than at any time in its history.
19. Becoming a part of the NCA will bring advantages to CEOP. Its existing operating principles will be
preserved as agreed when the Plan for the NCA was published, including its innovative partnerships with
private and voluntary sectors. It will also gain access to more capacity to deal with complex cases of child
sexual exploitation and abuse. It will gain greater resilience for front-line operational services and benefit from
support from other NCA specialist functions such as the National Cyber Crime Unit. Its influence, as part of
the national leadership the NCA will have across law enforcement, will be enhanced. The Crime and Courts
Act 2013 places a statutory duty on the NCA to safeguard and promote the welfare of children in England and
Wales, which means that all officers, not just those directly involved in child protection, will receive mandatory
training on safeguarding children. This means that every one of over 4000 officers will have a legal duty to
safeguard and promote child welfare so will be able to protect even more children from harm.
20. CEOP take the creation, distribution and accessing of illegal child sexual abuse images online extremely
seriously. Not only do such images capture a child being sexually abused, but the viewing of such images can
lead to an escalation in offending as well as continuing the victimisation of the child being abused in the
image. A CEOP report in 2012 (“A Picture of Abuse”) noted that research had identified a link between the
possession of illegal images of children and the contact sexual abuse of children.
21. The Government will support CEOP and the police to tackle those involved in the creation, dissemination
and accessing of child abuse images online. This includes ensuring that law enforcement agencies have the
powers that they need to investigate offenders and bring them to justice, and to develop new capabilities that
will enhance their effectiveness. The Prime Minister, in his speech of 22 July, announced plans for the creation
of a new child abuse image database that would hold a national record of the images seized by the police. This
would enable the police to identify known images more quickly on suspects’ computers, will improve their
ability to identify and safeguard victims from the images, and has the potential to enable industry to use the
unique numerical “hash values” derived from the images to search for them on their networks and remove the
before offenders can access them.
22. CEOP is also at the forefront of efforts to tackle more sophisticated offending where encryption and
anonymisation is used by offenders to share illegal child abuse images through photo-sharing sites or networks
away from the mainstream internet. This includes work in collaboration with the internet industry and the
Internet Watch Foundation to tackle the peer-to-peer sharing of child abuse images.
Working to Tackle Child Abuse Images Globally:
23. In addition to the work of the IWF considered above, CEOP work extensively with their partners overseas
to tackle the global threat from illegal child abuse images. CEOP participates in, leads and assists international
investigations into illegal child sexual abuse images and online child sexual exploitation. CEOP is also a
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 106 Culture, Media and Sport Committee: Evidence
founding member of the Virtual Global Taskforce which brings together law enforcement, public and third
sector partners from across the world to tackle child abuse materials.
24. The UK strongly supports the work of the Global Alliance between the EU and US, with CEOP leading
the work for the UK. The Deputy Director of CEOP is leading as the Driver on the EU EMPACT (European
Multidisciplinary Platform against Criminal Threats) Cyber Child Sexual Exploitation initiative. The first
meeting took place in October 2013 at Europol and agreed the Strategic Goals and Action Plan for 2014.
25. The Prime Minister has asked Joanna Shields, the Business Ambassador for Digital Industries, to lead
industry engagement for a new UK—US taskforce to bring together the technical ability of the industry in both
countries to tackle the use of the internet for child sexual abuse images. The Government expects to be able
to make further announcements about the work of the taskforce in due course.
Prevention—Helping Children Protect Themselves Against Becoming Victims
26. Online safety for children is of paramount importance to Government, and there are major crossGovernment efforts taking place to ensure that children learn about staying safe online both in schools and
through other initiatives.
27. CEOP’s work to educate children and young people focuses particularly on the implications of their
online behaviour, the “digital footprint” they leave, and the range of contact, content and conduct risks they
face online. CEOP has developed a specific educational programme called ThinkuKnow to tackle this issue.
The programme includes educational films and cartoons, teachers’ packs and other online resources for use
with children aged four—18. In 2012–13, over 2.6 million children saw the ThinkuKnow resources. In the
same year, over 800 professionals in education, child protection and law enforcement have been trained by
CEOP to educate children about online safety and how to respond to incidents. The Government, through
CEOP, has now trained over 25,000 professionals who deliver these key messages to children in classrooms
and other youth settings.
Tackling Material Intended to Promote Terrorism or Other Acts of Violence Online
28. Extremists use the internet to spread extremist messages to committed terrorists and those vulnerable to
radicalisation. This material covers a spectrum from illegal AI Qaeda propaganda which incites violence like
Inspire magazine, through to dedicated extremist forums and hate preacher videos on social media.
29. We know from recent attacks in Boston and Woolwich that online terrorist and extremist material remains
one of a range of variables that can contribute to an individual becoming radicalised, and/or tipping into the
acceptance of undertaking violent acts.69 Our primary focus under the Prevent Strategy (2011) has therefore
been to remove online material that breaches terrorism legislation. Proposals for going further to restrict access
to this material will be discussed at the Extremism taskforce in October 2013.
Current Approach—Limiting Access to Illegal/Terrorist Material
30. A specialist police unit, the Counter Terrorism Referral Unit (CTIRU) currently take down material that
breaches the Terrorism Act and is hosted in the UK, or—where we have strong relationships with industry.
CTIRU proactively seeks to identify illegal material, and members of the public concerned about terrorist or
extremist content online are able to refer material for investigation and potential removal by the CTIRU. In
this respect, the CTIRU acts like the Internet Watch Foundation (IWF) in receiving reports and ensuring content
is removed where possible.
31. Since it was established in 2010, CTIRU have taken down over 6000 pieces of material (more than
5,000 of which in the last 12 months). Even so, we would like to take down more terrorist content. However,
a significant amount is hosted overseas and domestic law, even when it has extra-territorial jurisdiction, is of
limited use. It would be difficult to prosecute a person or service provider operating outside the UK—they
would have to be extradited to the UK to face prosecution and they could well argue that they had not engaged
in criminal conduct in the country in which they were operating.
32. Where we cannot take down material, it is filtered from parts of the public estate (we have prioritised
schools and some libraries). The filtering project has developed and our content list (of over 1000 URLs) is
assessed by both CTIRU and the Crown Prosecution Service (CPS) to contain only URLs that are illegal. The
main drawback with filtering is that UK users are still able to choose to view this illegal terrorist content (via
desktop settings). We are considering how we can further restrict access to illegal terrorist material (potentially
at the network level), further aligning with the IWF’s approach.
33. There is no consistent global definition for what constitutes extremist material and the majority of
harmful, offensive and inappropriate content online does not meet legal thresholds for action by law
enforcement. For extremist material, this includes al-Awlaki’s series of online sermons and practical instructions
that can be of use to terrorists.
69
Joint Terrorism Analysis Centre (JTAC) assessment 2013
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 107
Improve the Industry Response to Terrorist Content
34. We aim to encourage the public and civil society groups to report legal but harmful content and seek to
get it removed under industry terms and conditions. We are therefore working with industry to ensure their
acceptable use policies are clear about what is tolerated on their platform and provide easy mechanisms to flag
concerns about extreme content.
35. We are keen to see a consistently effective response to terrorist content across industry. We have seen
some improvement following bilateral negotiations on individual acceptable use policies, but inconsistencies
remain. Some companies run schemes which prioritise law enforcement referrals, however others make no
explicit mention of terrorist or extremist content in their policies nor do they provide a single point of contact
for law enforcement. Whilst engaging with industry to ensure that their own acceptable use policies are being
applied rigorously, we are also considering the Home Affairs Select Committee recommendation of establishing
a code of conduct for internet companies, distinct from their own terms and conditions, to improve the response
to terrorist material (eg including “terrorism” as a category under unacceptable use).
36. We also tackle this problem from the user-end and run digital awareness projects that work with
practitioners, parents and young people to ensure they have the skills to recognise and challenge extreme
content. For example, through Prevent funding the Youth Justice Board have trained a hundred youth offending
team practitioners to ensure they have skills to challenge and report extremism online and we have worked
with Association of Chief Police Officers (ACPO) who have disseminated an internet safety DVD to local
prevent projects, raising awareness of the issue. We will work with DCMS to ensure we are linked into
initiatives such as Safer Internet Centre and Get Safe Online, which provide internet safety information and
advice alongside a wealth of internet safety resources for schools and information for parents and children.
Making Extremist Content Less Attractive and Promoting Positive Messages
37. We also seek to divert audiences and those vulnerable to radicalisation from extremist messages. The
Research, Information and Communications Unit (RICU), part of the Office for Security and Counter Terrorism
in the Home Office, aims to challenge and confront terrorist ideology and provide pathways for people to
explore alternatives. Activity includes working with digital communications experts to help civil society groups
make better use of the internet to counter extremist ideology. We intend to scale up this work.
38. RICU actively shares best practice on communications through the Global Counter Terrorism Forum,
the EU Radicalisation Awareness Network, and in bilateral work with the US and other countries.
October 2013
Written evidence submitted by the Department for Culture, Media and Sport
1. Introduction
1. The Government welcomes this inquiry into online safety. The internet has transformed our lives; changing
the way we interact with one another, promoting the development of new and innovative technologies and
driving the UK’s economic ambitions for the future. The UK has built strong digital foundations, and is reaping
the benefits—it is estimated that in 2010 the digital economy accounted for over 8% of GDP, a greater
proportion than in any other G20 country.70
2. Government’s aim is for the UK to be a nation of confident and safe consumers, able to make the most
of the opportunities the digital world has to offer. However, in the face of such rapid technological and cultural
change—as well as maximising the benefits the internet brings—we must be aware of, and responsive to, the
risks it presents. In July this year, Government published a policy paper, “Connectivity, content and consumers:
Britain’s digital platform for growth”71 following its review of the media and telecommunications sectors. This
policy paper considers the issue of online safety but also sites it in the context of the wider media environment.
3. Ensuring safety online, particularly for children, is a priority for Government. In a speech72 on 22 July
to the National Society for the Prevention of Cruelty to Children (NSPCC) the Prime Minister set out a range
of measures to ensure that the cultural challenges of protecting children from accessing potentially harmful or
age inappropriate content, and the challenges of restricting the availability of, and access to, criminal content
such as child abuse material are met. This evidence considers harmful or age inappropriate content (section 2)
and threatening or abusive behaviour (section 3). The Home Office will be submitting written evidence in
relation to illegal content separately.
70
71
72
Boston Consulting Group
https://www.gov.uk/government/publications/connectivity-content-and-consumers-britains-digital-platform-for-growth
https://www.gov.uk/government/speeches/the-internet-and-pornography-prime-minister-calls-for-action
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 108 Culture, Media and Sport Committee: Evidence
2. Harmful Content
4. Currently, 91% of children live in households with internet access and a greater proportion of children
aged 12–15 own smartphones than adults.73 While consenting adults should be free to watch the legal content
they choose, and while children and young people are important consumers of digital content, it is the case
children’s use of the internet can be a cause of concern for parents. Government understands that, first and
foremost, responsibility for keeping children safe online falls to parents and guardians; however, Government
is acting to ensure that parents have the tools and information they need to be able to do so.
UK Council on Child Internet Safety (UKCCIS)
5. Government believes that a multi-stakeholder approach to the issue of limiting the access of children to
harmful content is the most effective way to deliver results. The Government has been working through the
UK Council for Child Internet Safety (UKCCIS), which brings together more than 200 organisations across
the information and communication industries, law enforcement, regulators, academia, and charities—to pursue
a voluntary approach to child internet safety and has called on industry to make the right tools available to
allow parents to protect children online.
Measures for limiting access to harmful content
6. In his speech on 22 July, the Prime Minister outlined the range of measures Government is working with
industry to deliver which will help parents and guardians limit children’s access to harmful or age
inappropriate content:
73
74
—
Domestic internet filtering: Providing parents and guardians with easy-to-use and effective tools to
help limit the content that children can access in the home is a key aspect of the overall package of
measures that should be available. Government has been working with industry to secure the
commitment that, by the end of this year, the four largest internet service providers (ISPs)—BT,
Virgin Media, Sky and TalkTalk making up almost 90% of the fixed broadband market74—will
provide, free of charge, family-friendly network-level filtering (covering all of the devices in the
home which connect to that internet service) for new customers by the end of this year. This will
mean that, when someone sets up a new broadband account, the settings to install family-friendly
filters will be automatically selected; if customers just click next or enter, then the filters would
automatically be on. ISPs have also committed to contacting existing customers through 2014 and
providing them with an unavoidable choice on whether or not to set up filters. Again, this will be
free of charge. While the technology being deployed by each of the ISPs is different, based on the
differing infrastructure they have in place, the family-friendly filtering solutions offered will include
a number of customisable options, allowing parents to select or de-select categories of content,
“black-list” or “white-list” (filter or ensure access, respectively) certain URLs, dependent upon their
own specific circumstances.
—
Verification of those setting the filters: ISPs have also committed to developing processes to ensure
that the person setting the filters are aged 18 or over, so avoiding the situation that it is a child
setting the filtering levels. It is anticipated that this will be delivered through a “closed-loop” email
system where the main account holder will be contacted to verify the setting of filters, and informed
if the settings on the filters are changed subsequently.
—
Smaller ISPs: While the largest ISPs cover almost 90% of the broadband market, Government is
exploring options for the smaller ISPs—through the Internet Service Providers’ Association (ISPA)—
to deliver tools and solutions to their domestic consumers which can put users in control of the
content they, and their family, receive.
—
Filtered public Wi-Fi: The six largest wireless internet providers (Arqiva, BT, Nomad, Sky, Virgin
and O2) covering around 90% of the market committed to, and have implemented, publicly available
family-friendly Wi-Fi in places where children regularly visit, ensuring that children are unable to
access pornographic content.
—
Mobile phone filtering: The Mobile Broadband Group published a UK “Code of Practice for new
forms of Content on Mobile” in 2005 which requires mobile network operators (MNOs) to provide
a filter to their internet access services. Whilst the vast majority of providers are filtering internet
services in this way, coupled with robust age verification processes, Three—as the last major operator
to fully implement these measures—has committed to doing so.
—
Public awareness campaign: The largest ISPs are working to develop an awareness campaign for
parents for next year. This campaign will provide support and advice to parents about bringing their
parenting skills into the online world and will complement plans by Government to highlight internet
safety messages in interactions it has with citizens.
Source: Ofcom Children and Parents: Media Use and Attitudes Report Oct 2012— http://stakeholders.ofcom.org.uk/binaries/
research/media-literacy/oct2012/main.pdf
Source: Ofcom Communications Market Report 2013 http://stakeholders.ofcom.org.uk/binaries/research/cmr/cmr13/2013_UK_
CMR.pdf
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 109
—
Ofcom reporting: Ofcom has been asked to report, by the end of this year, on a number of areas in
relation to child internet safety—predominantly around parental awareness of, and confidence in
using, the internet safety tools available. A further report, in a year’s time, will help assess whether
levels of parental awareness and confidence have increased as a result of these measures.
7. While these will provide a range of tools to help support parents and guardians to make sure children do
not access harmful or age inappropriate content Government is clear that technology is not the only way in
which this issue should be addressed. Ensuring that children are resilient, competent and confident when online
is important and education is a key aspect of this.
Education
8. As part of the government’s reforms to the national curriculum, it will be strengthening the requirements
to teach e-safety as part of changes to the new computing programmes of study. The existing national
curriculum only requires e-safety to be taught to 11 to 16 year olds in key stages 3 and 4. From September
2014 e-safety will be taught to pupils at all key stages, from primary pupils aged five through to secondary
pupils aged 16.
9. There is also the award-winning programme for children and parents Think U Know75 which is produced
by the Child Exploitation and Online Protection Centre (CEOP—part of the Serious Organised Crime Agency,
a Government Agency). Through trained volunteers, this programme reaches hundreds of thousands of children
each year, educating them on topics such as grooming and sexting. CEOP falls under the command of the
National Crime Agency on 7 October.
10. The UK’s Safer Internet Centre76 (supported with EU funds) has a hotline for internet safety information
and advice and provides a wealth of internet safety resources for schools and information for parents and
children. The Centre leads on activity for Safer Internet Day in February each year which attracts significant
media coverage (this year on BBC Radio 1 and 5 Live in addition to coverage on Facebook and Google).
11. Get Safe Online77 is a joint initiative between the Government, law enforcement, leading businesses
and the public sector. Their aim is to provide computer users and small businesses with free, independent, userfriendly advice that will allow them to use the internet confidently, safely and securely. Working with Get Safe
Online, the National Fraud Authority’s e-Confidence campaign begins in October.
12. Childnet is a children’s charity working in the area of policy and has developed a range of awardwinning websites and resources to help provide children, parents and teachers with the information and skills
they need to use the internet safely and responsibly. Examples of Childnet’s work include the Know IT All78
suite of resources, the guidance for schools on cyberbullying, the Let’s Fight it together film, and many more.
13. The South West Grid for Learning79 is a not for profit, charitable trust company, that specialises in
supporting organisations with online safety. It was established in 2000 to provide schools and many other
educational establishments throughout the South West of England with safe broadband internet connectivity;
broadband-enabled learning resources but especially providing professionals, parents and children with advice,
resources and support in using the internet safely across the wider UK.
14. ParentPort80 was launched in October 2011, following the Bailey Review, to make it easier for parents
to complain about material they see or hear across the media, communications and retail industries. ParentPort
is run by the UK’s media regulators and provides parents with useful information about media standards
and internet safety as well as giving parents an easy way to make a complaint and share their views with
the regulators.
3. Abusive or Threatening Behaviour
15. The issue of abusive or threatening behaviour—whether targeted at adults or children—over the internet
is an area of concern for Government. However, the legislative position in relation to this is clear; where
behaviour is illegal off-line it is also illegal on-line and there are a number of pieces of legislation which can
be, and are, used to prosecute.
Legislation and Director of Public Prosecutions (DPP) Guidance
16. Communications sent via social media are capable of amounting to criminal offences in relation to a
range of legislation. Including:
— Communications which may constitute credible threats of violence to the person or damage to
property.
75
76
77
78
79
80
http://www.thinkuknow.co.uk/
http://www.saferinternet.org.uk/
https://www.getsafeonline.org/
http://www.childnet.com/resources/kia/
http://www.swgfl.org.uk/Staying-Safe
http://www.parentport.org.uk/
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 110 Culture, Media and Sport Committee: Evidence
—
Communications which specifically target an individual or individuals and which may constitute
harassment or stalking within the meaning of the Protection from Harassment Act 1997.
—
Communications which may be considered grossly offensive, indecent, obscene or false.
17. Communications which are grossly offence, obscene, indecent or false fall to be considered either under
section 1 of the Malicious Communications Act 1988 or under section 127 of the Communications Act 2003.
18. Section 1 of the Malicious Communications Act 1988 makes it an offence to send to another person:
(a) an electronic communication which conveys:
(i) a message which is indecent or grossly offensive;
(ii) a threat; or
(iii) information which is false, and is known or believed to be false by the sender; or
(b) any electronic communication which is, in whole or part, of an indecent or grossly offensive nature,
provided that one of the purposes of sending the communication is to cause distress or anxiety to
the recipient. The offence is one of sending, delivering or transmitting, so there is no legal
requirement for the communication to reach the intended recipient.
19. Section 127 of the Communications Act 2003 makes it an offence to send or cause to be sent through a
“public electronic communications network” a message or other matter that is “grossly offensive” or of an
“indecent, obscene or menacing character”. The same section also provides that it is an offence to send or cause
to be sent a message that the person knows to be false “for the purpose of causing annoyance, inconvenience or
needless anxiety to another”. The defendant must be shown to have intended or be aware that the message was
grossly offensive, indecent or menacing, which can be inferred from the terms of the message or from the
defendant’s knowledge of the likely recipient. The offence is committed by sending the message. There is no
requirement that any person sees the message or be offended by it.
20. Other relevant legislation includes:
—
Offences Against the Person Act 1861;
—
Computer Misuse Act 1990;
—
Protection from Harassment Act 1997;
—
The Criminal Justice and Public Order Act 1994; and
—
Section 15 Sexual Offences Act 2003 (for grooming)
21. In June 2013, following consultation, the Director for Public Prosecutions published guidelines81 for
prosecutors when considering cases involving communications sent via social media. These guidelines set out
that prosecutors (subject to general guidelines around having sufficient evidence and whether prosecution is in
the public interest) should, when considering whether communications sent via social media are capable of
amounting to criminal offences, make an initial assessment of the content of the communications and the
conduct to distinguish between:
—
There is a credible threat of violence.
—
Where communications specifically target an individual or individuals.
—
Where communications breach a court order.
—
Where communications may be considered grossly offensive, indecent, obscene or false.
22. These guidelines seek to draw the difficult balance between protecting freedom of speech on the one
hand with acting robustly against communications which cross the threshold into illegality.
23. Data from the Crime Survey for England and Wales (below) shows that whilst almost a third of adults
had some form of negative experience online, the majority were related to viruses. The proportions of people
experiencing “upsetting/illegal images” and “threatening/abusive behaviour” are much lower, and some of
those experiences are also likely not to have met the criminal threshold.
Experience of negative incidents online among adults aged 16 and over
2010–11 2011–12
Computer virus
Loss of money
Unauthorised access to/use of personal data
Upsetting/illegal images
Abusive/threatening behaviour
Any of the above incidents
Other (ie experienced no incident)
81
26.0%
2.6%
4.5%
3.1%
1.6%
30.5%
69.5%
24.6%
2.6%
5.2%
3.5%
1.4%
29.4%
70.6%
http://www.cps.gov.uk/news/latest_news/dpp_publishes_final_guidelines_for_prosecutions_involving_social_media_
communications/
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 111
Acceptable use policies and reporting mechanisms
24. In the vast majority of cases communications sent via social media will not cross the threshold into
criminal behaviour but may still cause offence to other users. Where this is the case, Government expects
social media companies, and others, to have robust processes in place promptly when abuse is reported;
including acting expeditiously to assess the report, removing content which does not comply with the acceptable
use policies or terms and conditions in place and, where appropriate, suspending or terminating the accounts
of those breaching the rules in place.
25. A UKCCIS working group has been looking at the issue of transparency of reporting processes following
reports of abuse by social network users and difficulties with complaints and reporting. Parents and children
have indicated that they often cannot easily make a report, that their report may not be handled quickly and
they are given little feedback about how their reports are dealt with. Ministers will consider how this work can
best be taken forward at the next UKCCIS board meeting on 6 November.
October 2013
Supplementary written evidence submitted by the NSPCC
Introduction
1. The NSPCC welcomed the opportunity to respond to the Culture, Media and Sport (CMS) Select
Committee inquiry into online safety in September, and we were delighted to be invited to come before the
committee in October to give oral evidence.
2. Since then, the NSPCC has released further research on the issue of online safety and was the only
children’s charity to be involved in the Prime Minister’s internet safety summit in November. As such, we
have requested to submit supplementary evidence to the inquiry, to be read alongside our original response.
We hope both the statistics and perspectives included here are helpful for the Committee when making its
recommendations.
Harmful Contact Online
New Key Stats
— 10,600 young people contacted ChildLine in 2012–13 for support and advice on how to deal with
internet safety related issues—a 65% increase on the year before82
— 4,507 young people contacted ChildLine in 2012–13 for support and advice on how to deal with
being bullied online—an 87% increase on the year before83
— More than One in Four 11–16 year olds with a social networking profile have experienced something
upsetting on a social networking site.84
3. Research the NSPCC is about to launch research which shows that 28% of young people who have a
social networking profile have experienced something that has upset them in the last year. These experiences
include cyberstalking, being subjected to aggressive or offensive language, being sent sexually explicit pictures
and being asked to provide personal or private information. However, the greatest proportion of this group
(37%) told us that they had experienced “trolling”. Another alarming statistic showed that over half of these
young people said they did not know the person who was responsible for the behaviour that upset them.
4. Alongside this evidence that online bullying is clearly a problem for young people in the UK, the latest
ChildLine statistics showed an 87% increase in 2012–13 in the number of young people contacting us to ask
for support and advice about being bullied via social networking sites, chat rooms, online gaming sites, or via
their mobile phones. We believe the increasing ownership of smartphones and tablets by young people is likely
to be contributing to this trend.
5. In our original submission to the Committee the NSPCC described the need for online safety, including
information about respectful behaviours and managing upsetting experiences, to be a core part of the national
curriculum, taught through compulsory Sex and Relationships Education (SRE). We also called for social
networking sites and other technological leaders to provide straightforward and meaningful ways to report
abuse online. We continue to call for this action and in addition the NSPCC would be delighted if the
Committee could support the following asks:
— The UK Council for Child Internet Safety (UKCCIS) Executive Board should make tackling the risks
children experience on social networking sites a priority
— Ofcom—should play a greater role in providing easily accessible information to parents about social
networking sites
82
83
84
Can I tell you something?, ChildLine Reviewt 2012/13 (Jan 2014) http://www.nspcc.org.uk/news-and-views/media-centre/pressreleases/2014/childline-report/childline-report_can-i-tell-you-something_wdf100354.pdf
Ibid
New NSPCC research due for release end Feb 2014
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Ev 112 Culture, Media and Sport Committee: Evidence
—
Social networking sites and others—should offer a range of options for how to report illegal or
offensive bullying or content and default privacy settings should be set to the highest levels possible
for users under 18. The NSPCC also believes sites should provide incentives for new users to take
safety and etiquette tutorials.
6. Additional to the issue of online bullying, the ChildLine annual review, Can I tell you something? also
showed that 1061 young people specifically talked about online sexual abuse, with girls being the most affected.
ChildLine also saw a 28% increase in contacts from young people who talked about being asked to post
sexually provocative images online, or share them via a mobile phone.
7. The NSPCC remains adamant that children should be given strategies to resist unwanted sexual
approaches, encouraged to report upsetting experiences online and be taught about online safety as part of
compulsory SRE lessons at school. We are pleased that the Government has committed to emailing out to
every school SRE advice developed by Brook, the Sex Education Forum and others. This is a helpful interim
measure to address the current shortfall in up-to-date statutory SRE guidance. However, we believe the
Government should commit to updating the official SRE guidance from 2000 to include issues children face
in today’s world, like staying safe online.
8. We do, of course, believe parents have a central role in talking to their children about how to stay safe
online. We appreciate these conversations might feel daunting, but we believe parents needn’t be experts on
the internet, but instead treat online safety as they would any other form of parenting where rules and
boundaries need to be set.
9. We also recognise that other agencies that work with children and prioritise children’s safety should play
a role alongside schools and parents in providing online safety advice and education. The NSPCC in October
2013 launched the ChildLine ZipIt app to discourage children from sending sexually explicit messages. This
has already been downloaded by 40,000 young people so far. We have also developed in collaboration with
CEOP the Keeping Children Safe Online e-learning course aimed at any organisation or professional working
with children.
Age-inappropriate Content
New Key Stats
—
63% of children aged 11–12, who use a social networking site, have a profile on Facebook, and
28% use Twitter despite the minimum age for registration on both sites being set at 13.85
—
28% of over 601 school pupils surveyed by the NSPCC think pornography definitely “influences
how young people have to behave in a relationship.”86
10. Based on our research, we estimate that half of all the 11 and 12 year olds in the UK use services that
have a minimum age of 13. The NSPCC is concerned that this leaves them vulnerable to engaging with content
and behaviours that are not appropriate for their age. In addition, ChildLine counselled 475 young people who
had been exposed to sexually explicit images or internet pornography and almost one third of children surveyed
by the NSPCC about watching pornography said it affected the way they acted when they were in a relationship.
11. We support plans by the major Internet Service Providers (ISP) to launch a national campaign to help
parents use the filters and tools that are available to them to manage the online content to which their children
have access, and more broadly help parents to understand more about internet safety. To support these efforts,
the NSPCC also calls for:
—
UKCCIS—to undertake a robust and independent evaluation of the ISPs campaign to measure its
effectiveness and push for improvements where needed
—
The Government—should evaluate the effectiveness of active choice plus and be prepared to explore
legislative solutions to children accessing adult content and services.
Child Abuse Images
12. In our original evidence to the Committee the NSPCC explained that a Freedom of Information request
we made in 2010 revealed that nearly 26 million child sexual abuse images had been confiscated in the
preceding two years from just five of the 43 police forces in England and Wales which were able to check
their records. At that time most forces said it would take too long to interrogate files to see how many pictures
of children being sexually abused they had accumulated during investigations. However, estimates of the
number of online child abuse images confiscated by police in England and Wales each year in the UK range
from 100–360 million.87
85
86
87
NSPCC, Younger Children and Social Networking Sites: A Blind Spot? (Nov 2013) http://www.nspcc.org.uk/Inform/
resourcesforprofessionals/onlinesafety/younger-children-report_wdf99929.pdf
NSPCC (Sept 2013) http://www.nspcc.org.uk/news-and-views/our-news/child-protection-news/telegraph-better-sex-education/
online-porn-dictates-young-people-relationships_wda98264.html
NSPCC (Oct 2012) http://www.nspcc.org.uk/news-and-views/media-centre/press-releases/2012/12–10–15-urgent-action-childabuse-images/urgent-action-child-abuse-images_wdn92344.html
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 12:12] Job: 035254 Unit: PG04
Culture, Media and Sport Committee: Evidence Ev 113
13. The NSPCC welcomed the opportunity to contribute to the Prime Minister’s internet safety summit in
November, and we were pleased by the commitment shown by the leading search engines to ensure that tens
of thousands of the most disturbing search terms will return only “clean” results and warnings about child
abuse imagery. We now urge the Government and industry to work together to monitor and evaluate the
effectiveness of these technological solutions and make improvements where needed.
14. Alongside this proactive approach from industry, we want to see a tireless commitment to policing efforts
to tackle child abuse images from the CEOP command within the new National Crime Agency. Images of
child abuse was mentioned briefly in the NCA’s 2013–14 Annual Plan as an operational priority:
“To tackle the enablers of crime that have both utility and impact across several threat areas. For
example cyber-enabled criminality where access to the hidden internet can facilitate criminal acts
such as illicit drugs supply, images of child abuse and the trade in stolen credit card data.”88
15. However, we hope that the Agency’s programme of work on child abuse images will feature much more
extensively in its next annual plan, and more broadly on policing the NSPCC wants to see:
— A greater number of arrests and prosecutions for offences relating to child abuse images.
— A commitment to identifying and destabilising the criminals at the centre of child abuse image
networks.
— A greater focus on tackling child abuse images at a local level from Police and Crime
Commissioners, and in Police and Crime Plans.
16. The NSPCC also eagerly awaits the work of the UK/US taskforce on child abuse images which we hope
will encourage action from the internet and technology industries to apply their technical expertise to develop
innovative solutions to this global challenge.
January 2014
Printed in the United Kingdom by The Stationery Office Limited
03/2014
035254
19585
88
National Crime Agency Annual Plan 2013/14 (Oct 2013), p.7 http://www.nationalcrimeagency.gov.uk/publications/33-ncaannual-plan-2013–14/file
House of Commons
Culture, Media and Sport
Committee
Online safety
Sixth Report of Session 2013–14
Volume II
Additional written evidence
Ordered by the House of Commons
to be published [date]
Published on [date]
by authority of the House of Commons
London: The Stationery Office Limited
The Culture, Media and Sport Committee
The Culture, Media and Sport Committee is appointed by the House of Commons to
examine the expenditure, administration and policy of the Department for Culture, Media
and Sport and its associated public bodies.
Current membership
Mr John Whittingdale MP (Conservative, Maldon) (Chair)
Mr Ben Bradshaw MP (Labour, Exeter)
Angie Bray MP (Conservative, Ealing Central and Acton)
Conor Burns MP (Conservative, Bournemouth West)
Tracey Crouch MP (Conservative, Chatham and Aylesford)
Philip Davies MP (Conservative, Shipley)
Paul Farrelly MP (Labour, Newcastle-under-Lyme)
Mr John Leech MP (Liberal Democrat, Manchester, Withington)
Steve Rotheram MP (Labour, Liverpool, Walton)
Jim Sheridan MP (Labour, Paisley and Renfrewshire North)
Mr Gerry Sutcliffe MP (Labour, Bradford South)
The following members were also a member of the committee during the parliament:
David Cairns MP (Labour, Inverclyde)
Dr Thérèse Coffey MP (Conservative, Suffolk Coastal)
Damian Collins MP (Conservative, Folkestone and Hythe)
Alan Keen MP (Labour Co-operative, Feltham and Heston)
Louise Mensch MP (Conservative, Corby)
Mr Adrian Sanders MP (Liberal Democrat, Torbay)
Mr Tom Watson MP (Labour, West Bromwich East)
Powers
The committee is one of the departmental select committees, the powers of which are
set out in House of Commons Standing Orders, principally in SO No 152. These are
available on the internet via www.parliament.uk.
Publication
The Reports and evidence of the Committee are published by The Stationery Office by
Order of the House. All publications of the Committee (including press notices) are on
the internet at www.parliament.uk/cmscom. A list of Reports of the Committee in the
present Parliament is at the back of this volume.
The Reports of the Committee, the formal minutes relating to that report, oral
evidence taken and some of the written evidence are available in a printed volume.
Additional written evidence is published on the internet only.
Committee staff
The current staff of the Committee are Elizabeth Flood (Clerk), Grahame Danby (Second
Clerk), Kevin Candy (Inquiry Manager), Emily Gregory (Senior Committee Assistant), Keely
Bishop (Committee Assistant) and Jessica Bridges-Palmer (Media Officer).
Contacts
All correspondence should be addressed to the Clerk of the Culture, Media and Sport
Committee, House of Commons, 7 Millbank, London SW1P 3JA. The telephone number
for general enquiries is 020 7219 6188; the Committee’s email address is
[email protected]
List of additional written evidence
(published in Volume II on the Committee’s website www.parliament.uk/cmscom)
Page
1
Dr Peter Dawe OBE
Ev w1
2
Dr Claire Hardaker
3
Remote Gambling Association
4
Ann Farmer
5
Kirsty Hopley
Ev w10
6
Dr Peter Nelson
Ev w11
7
British Board of Film Classification (BBFC)
Ev w13
8
Arqiva
Ev w18
9
James Griffiths
Ev w22
10
Christopher G. H. Thomas
Ev w22
11
John Edgar Cameron Lowe
Ev w23
12
Norfolk Library and Information Service
Ev w23
13
Phil Alexander
Ev w25
14
Ben Hardwick
Ev w25
15
John R Edwards
Ev w26
16
Chris Evershed
Ev w27
17
EE Limited
Ev w28
18
BCS The Chartered Institute for IT
Ev w32
19
Dudley Friend Clayson
Ev w36
20
PAPY RUS Prevention of Young Suicide
Ev w37
21
EU Kids Online
Ev w38
22
Rolf Smith
Ev w43
23
Mediawatch-UK
Ev w45
24
South West Grid for Learning
Ev w47
25
British Sky Broadcasting Limited (‘Sky’)
Ev w51
26
Big Brother Watch
Ev w53
27
BT
Ev w56
28
Russell Hopwood
Ev w59
29
Stuart Taylor
Ev w60
30
Richard Browning
Ev w61
31
Timothy Michael Johnston
Ev w62
32
Lloyd Johnston
Ev w62
33
John Carruth
Ev w63
34
Rob Barry
Ev w64
35
Barry Saltmarsh
Ev w65
36
Safermedia
Ev w66
37
John Reiner
Ev w69
38
Family Online Safety Institute
Ev w69
39
Terry James
Ev w72
40
Film Distributors’ Association
Ev w73
Ev w2 ; Ev w7
Ev w7
Ev w10
41
Information Commissioner’s Office
Ev w74
42
Virgin Media
Ev w77
43
The Police Federation of England and Wales
Ev w84
44
End Violence Against Women Coalition
Ev w85
45
Childnet International
Ev w88
46
Ukie
Ev w92
47
The Children’s Society
Ev w96
48
Eric and Esme Bricknell
Ev w97
49
Stonewall
Ev w98
50
FPA and Brook
Ev w100
51
Michael J Smaldon
Ev w109
52
Prof Andy Phippen
Ev w110
53
Ahmadiyya Muslim Community UK
Ev w113
54
Sex Education Forum
Ev w117
55
Net Children Go Mobile
Ev w119
56
British Naturism
Ev w120
57
Open Rights Group
Ev w122
58
Russell Pillar
Ev w126
59
CARE (Christian Action Research and Education)
Ev w127
60
Malcolm Holmes
Ev w130
61
Microsoft UK
Ev w131
62
The Authority for Tel evision on Demand
Ev w132
63
Intellect
Ev w139
64
Telefonica UK Ltd
Ev w142
65
National Centre for Cyberstalking Research
Ev w143
66
Ethos Capital Ltd
Ev w145
cobber Pack: U PL: CWE1 [SO]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w1
Written evidence
Written evidence submitted by Dr Peter Dawe OBE
1. The Flavoured Internet Foundation
Internet damage limitation. Dr. P Dawe OBE, Founder Internet Watch Foundation, 2013.
2. Introduction
Once again there is a public furore about Internet content. However, this time the debate is messy and illdisciplined, as many voices are speaking about many issues: Child protection, legality of content, censorship,
financial predation, bullying to name just a few.
Policy makers are in an impossible position! Caught between protection and censorship with the added
problems due to the global nature of the Internet.
3. This paper outlines a means of mitigating the problems based on the genesis of The Internet Watch
Foundation by the person who lead the founding of IWF.
4. A Solution
It is our view that the problem is not a legislative problem, but a market problem. Previously, there has been
a diversity of media channels with each channel being moderated by an “editor”.
The Internet needs to offer the public a choice of “flavours” of Internet, similar to how people choose their
newspaper, TV channel or indeed their preferred supermarket.
5. Why is there Market Failure in Selective Internet?
— Currently providing a moderated view (Fenced Garden) of the Internet isn’t economically viable.
— The Internet is so large as to be difficult, nay impossible, to comprehensively categorise all content.
— The choice of “flavour” is also difficult as each person has their own set of prejudices, in many
dimensions. Eg Extent of naked flesh, Violence, Religion.
— Service providers are, understandably, reluctant to be seen as the arbiters of taste and legality.
— The threat of a defamation action when a content provider is censored on the basis of an algorithm,
particularly if the censorship is due to “business morality”.
— Computer applications to implement filtering have been proprietary and not comprehensive in their
use of available tools.
— Multiple legislative jurisdictions have to be considered.
6. Several attempts have been made to provide the public with filtered content, however typically these are
from a singular view-point (eg Western liberal Christian tradition), which means the cost of creating the filter
list has to be spread over a small user base.
7. Creating a Viable Marketplace
— There is a need for an open architecture for classification.
— The problem is too big for a individual players.
— Web authors do not want to negotiate with individual classification services and editors.
— New applications can be implemented using existing infra-structure.
— Editors want to be able to choose their own filtering thresholds and classes.
8. Tactical Implementation Steps
— Political encouragement to entrepreneurs to implement diverse Internet “flavoured” services.
— Extension of the legal protection, enjoyed by newspaper editors against civil claims from content
owners, to filtering services.
— A legal requirement on Internet service providers to be network neutral in regard to providers of
Flavoured Internet.
— The implementation of a series of demonstration technologies and services.
9. Leadership not Committee
The genesis of Internet Watch Foundation should be seen as a model for deployment. Internet Watch was
founded, NOT BY CONSENSUS, but designed and implemented through the maverick leadership of Mr Peter
Dawe. The players (Service providers, Government, Police, Politicians and the Press) are individually impotent.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w2 Culture, Media and Sport Committee: Evidence
Each has conflicting objectives, all consider the other players as responsible, irresponsible, ignorant, idealistic,
imposing impossible costs, ducking their responsibility etc.
10. However, with a more sophisticated set of players, it is less likely that a maverick leader will not be
able to overcome the likely barriers alone. Political endorsement of this initiative, for a finite period (two
years), would encourage many of the other players to also support the initiative. Thus allowing the model
design to be finalised, proof of concept completed and for the model to be promoted to all players world-wide.
11. Note: Other initiatives, such as consensus building committees, can still be undertaken in parallel with
this proposal.
12. Funding
For an effective team to complete the design, prove concept and importantly to sell the model internationally,
we estimate that a budget over the two years of £500,000 would suffice. A requirement of the design is that it
is ultimately self-funding in operation. (Internet Watch cost under £50,000 to set up, and was completed in less
than 12 months, a voluntary levy on the industry has funded it since.)
13. The Dawe Charitable Trust believe it can catalyse the actions needed for this global initiative, recruiting
support from the industry, government, religious, freedom-of-speech and other groups.
14. Dawe Charitable Trust already has an experienced core team committed to working on the initiative.
15. The Dawe Charitable Trust is seed-funding this initiative (£50,000), and the team will seek to design
and implement a model regardless. However adoption of the above recommendation will both speed the
initiative and improve the likelihood of global success.
September 2013
Written evidence submitted by Dr Claire Hardaker
Summary
In this document I discuss the following four points:
— What facilitates abusive or threatening online comments (ie what makes it “easy”?)
— What motivates abusive or threatening online comments (ie what makes people do it?)
— Measures unlikely to reduce abusive or threatening online comments (ie what won’t stop it?)
— Measures likely to reduce abusive or threatening online comments (ie what could stop it?)
Preliminary Comments
(1) Whilst my research has covered strategies of content filtering, it relates better to the discussion of
“preventing abusive or threatening comments”, and less directly to “protecting minors from adult content” or
“filtering extremist material such as child abuse imagery”. I therefore do not comment on these two areas.
(2) A crucial point note, however, is that filtering content intended to promote harmful behaviour (eg
terrorism, violence) is unlikely to deal with the online grooming responsible for inciting these behaviours.
Grooming is typically an interactive process involving a manipulator persuading a vulnerable target to
undertake action that is in the manipulator’s interest, and typically harmful to the target (eg extremist violence,
suicide, self-harm, age-inappropriate sexual behaviour, etc.).
(3) Such grooming is likely to occur in interactive environments such as chatrooms, messageboards, and via
private direct-message software (eg Windows Live Messenger) where content filtering will struggle to apply,
and where processes of moderation are likely to be far more effective. I would therefore suggest that content
filtering as a strategy for dealing with material intended to promote terrorism or other acts of violence is
unlikely to tackle this point with much success.
What facilitates abusive or threatening comments online?
(4) There are arguably three major factors (though easily more besides):
Anonymity
(5) The internet offers a perceived anonymity that has no real parallel offline, and this appearance of
invisibility encourages the user to feel that they can do unpleasant things with a highly reduced risk of suffering
any consequences. In turn, the sense of being invulnerable encourages disinhibition: the user thinks that she is
safe from repercussions, so behaves in a mildly objectionable way. Each time there is no negative consequence,
her behaviour may gradually escalate.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w3
(6) By itself, however, anonymity does not explain why the internet seems to bring the worst out in some,
since many of us are online and anonymous all the time, yet would never think to behave in this way. A second
issue to consider, then, is detachment.
Detachment
(7) Linked to disinhibition is the way that the internet allows us to shut down our empathy, and in particular,
our compassion. In fact, we can choose not to empathise at will, especially when dealing with someone we
dislike. The internet, however, drastically increases this ability, and allows us to emotionally distance
ourselves—not just from people we don’t like but also from those we don’t even know—in several ways:
(a) Dehumanising: Because we lose many indications of emotional response (eye contact, tone of
voice, facial expressions) it is easier to “forget” that we are communicating with another human.
Instead, we have only the words and/or images sent back and forth.
(b) Devaluing: The above means that we can downplay any emotional reaction. If they claim to be
offended or hurt, because we don’t see that reaction, it is easier to believe that they are lying
or exaggerating. It is also easier to quickly forget the incident, whilst it may linger in the other’s
mind for days.
(c) Victim-blaming: It can be easier for the bully/troll to shift the responsibility for what they are
doing from themselves, and blame their victim for what is happening, because, eg, they’re
famous, or they made a tribute page public, etc. This stance is typically also indicated by
“solutions” such as telling the victim that if they don’t like it, they should “get off the internet”.
(d) Self-vindicating: It can also easier to diminish the severity of our behaviour to ourselves.
Because we each inhabit our own heads, it seems obvious to us just how seriously we meant
something, so when trolls/cyberbullies are confronted, we tend to hear excuses like, “Stuff said
on the internet isn’t real. I was just joking. I obviously didn’t really mean that [rape-threat/
defamatory statement/etc].” Our innate bias means that we are more likely to adopt such a selfsupporting viewpoint than to judge our own behaviour as reprehensible.
Entertainment
(8) The third aspect that seems to encourage some to be deliberately mean online long predates the internet.
It is simply the human tendency to enjoy and glorify aggression. Evidence for this can be found worldwide,
stretching back over millennia. For brevity, however, we need only look at ancient military artefacts (eg the
Bayeux tapestry, Homer’s Iliad, Beowulf), the sports descended directly from warfare (javelin, archery, fencing,
boxing, wrestling, martial arts), and archaeology built especially for the enjoyment of bloodshed (eg the
Coliseum, the Hippodrome).
(9) Today, one glance at mainstream entertainment demonstrates that this interest is not remotely abated. For
instance, The Hunger Games is a book series about individuals forced to fight to the death for a reality
television show. Shawshank Redemption tells of a man wrongly convicted for the murder of his wife and
incarcerated in a brutal, corrupt prison where inmates are beaten to death, shot, and raped. The X Factor
television series places competitors into highly stressful, even cruel environments with antagonistic judges.
And many of the biggest games titles, such as the Call of Duty, Halo, and Grand Theft Auto series, involve
players simulating extreme violence and murder.
(10) The above should demonstrate that humans are entertained by violence, and that this is by no means a
new phenomenon brought about by digital media. For some, however, simply consuming aggression-based
entertainment from fiction (eg books, films) or simulation (eg games, sports) is not enough. For them, the
internet presents a virtual Coliseum where they can amuse themselves by attacking others without risk of
“injury”, since they can hide behind a cloak of anonymity, and silence their own conscience by ignoring their
target’s feelings.
What motivates abusive or threatening comments online?
(11) To reiterate a point from above, whilst many of us are frequently online anonymously and enjoy
consuming violence (it is a rare person who hasn’t watched an action movie!), it is still a relatively small
number who feel motivated to take advantage of the conditions that the internet presents, and to actually engage
in abusive or threatening online behaviour. In the same way, many of us drive, but however annoyed we may
become behind the wheel, mercifully few of us are ever turn our car into a lethal weapon.
(12) In other words, each individual needs a trigger, or motive to push them into using the internet as a
weapon. Unsurprisingly, the evidence suggests that there is no single, over-riding motive for being abusive or
threatening online, but that whilst there certainly are trends, there are also many unique cases. Some of those
broad trends include:
(a) Boredom, a need for entertainment: This is perhaps the most trivial reason (and therefore,
perhaps also the most frustrating and difficult to resolve)—simply too much free time after
school, at work, or between jobs. These are the individuals who are killing time, and even
seeking impress others of a similar mindset. Groups of trolls coalesce on sites like 4chan, post
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w4 Culture, Media and Sport Committee: Evidence
links to targets that might prove “fun”, and compete against each other to see who can be the
funniest, cleverest, or most extreme. They may also organise troll raids via more secure systems
(eg IRC, Torchat), and train newer recruits in how to be less identifiable online, eg by sharing
information about identity-protecting software such as TOR, I2P, and Psiphon.
(b) A need for attention: these individuals seem to display a craving for any kind of attention,
whether positive or negative. Such individuals may also post annoyingly implausible stories,
grand claims, age-inappropriate content, etc. This may be symptomatic of emotional, social,
and/or psychological problems.
(c) Revenge, disenfranchisement: these appear to be individuals who feel cheated, short-changed,
or otherwise wronged on a personal level, and appear to want to “fix” this by making the object
of their malice feel as miserable as themselves. This may be symptomatic of a struggling
economy, eg the student who can’t find work and has therefore taken to targeting his local MP.
(d) Politics, activism: whilst the motive in its own right may be good (eg fighting for racial equality)
the method may not meet that same moral standard. This said, it’s easier to comprehend why
individuals might choose this noisy, attention-grabbing route if we consider that letters to
politicians, newspapers, and economists are all too easily ignored. This may be symptomatic,
therefore, of a political system that appears (to the trolls/bullies) to ignore their voices.
(13) In summary, motivations for posting abusive/threatening content can spring from a number, and a
mixture of sources, including emotional, psychological, social, political, economical, etc. This means that there
is likely to be no single approach that will have any meaningful impact on this issue. Instead, it is far more
likely to require addressing on multiple fronts. I discuss this area next.
Measures unlikely to reduce online abuse or threats
(14) There are two classic “solutions” that have been proposed to fix online abusive comments. I deal with
them briefly since they merit very little serious consideration.
Remove Anonymity
(15) In a nutshell, this is borderline impossible, if only because it is unenforceable, and unworkable. Even
if all countries agree to legally mandating online identity disclosure (unlikely) the costs of setting up,
administrating, and then enforcing it would be staggering. Further, we need only consider the risks inherent in
having a child’s name, age, location, etc, available online to realise that online identity disclosure would
actually create more dangers than anonymity currently averts.
Stop Using the Internet (or the Site in Question)
(16) This is not only bizarrely unrealistic, since the internet is now ubiquitous, it is also ineffective (kids
don’t need a single device to be cyberbullied) and puts responsibility on the victim, rather than the attacker
(see victim-blaming, para. (7(c)) above).
Measures likely to reduce online abuse or threats
(17) There is no single answer to this complex problem, but several smaller improvements can, collectively,
advance online safety. I therefore recommend five ways forward.
Education, Resources, and Training for Children and Young Adults
(18) In a perfect world, we would always choose prevention over cure, but this requires well-informed
vigilance both from potential targets and from those responsible for protection. At present, Childnet does
outstanding work, teaching school-children to be safe and kind online, and I welcome the fact that online safety
has been introduced across the 2014 National Curriculum.
(19) For best effect, however, these lessons need to be appropriately supported (see paras. (21)–(23) below),
sufficiently detailed/thorough, and adopted across all schools (private, public, and free). These lessons also
should not be restricted purely to Computing, but should be adopted across a number of classes, including
PSHEE and Citizenship.
(20) However, in order for this to be successful, teachers themselves need to be well-trained in the safety
measures, in tell-tale signs, and in appropriate strategies of intervention. Teachers also need the support of
parents and caregivers, and this leads into a second area of improvement.
Education, Resources, and Training for Parents, Caregivers, and Teachers
(21) The assistance that Childnet currently offers teachers is also invaluable, but it could be substantially
underpinned with compulsory online safety training in all PGCEs, since NSPCC research shows that far too
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w5
many teachers currently don’t feel confident when it comes to advising pupils on safe social network
practices.1
(22) Meanwhile, parents would benefit from proactive, face-to-face training in:
(a) signs indicative of cyberbullying (whether as victim or bully);
(b) ways to keep children safe online2 (technologically, behaviourally); and
(c) the many organisations that offer help and advice (eg BeatBullying,3 Childline,4 Childnet,5
Get Safe Online,6 IWF,7 NSPCC,8 SIC,9 NHS,10 CEOP11).
(23) Whilst these organisations all do invaluable work, they cannot be a standalone solution. A third area of
improvement, then, is through sites themselves.
Site Tools, Responsibilities, and Processes
(24) We generally expect social networks to offer the tools and support to keep us safe. Indeed, the
expectation seems fairly reasonable. If Mr Smith wants to profit from people buying his goods or services, his
products and premises must meet the relevant legal requirements to protect customers from loss and harm.
However, should Mr Smith want to profit from people using his social network, in the UK at least there appear
to be no similar legal requirements.
(25) It seems reasonable to argue that the opportunity to profit from a social network should also entail a
duty of care towards that site’s users. However, whilst most social networks require registrants to enter into
agreements forbidding abusive behaviour, beyond that, sites differ widely in the availability, sophistication, and
transparency of their safety features. At the most basic level, social networks should:
(a) allow users to control their own environment (eg privacy settings);
(b) flag up content for the site to manage (eg report abuse buttons);
(c) proactively use their own management tools (eg IP blocks); and
(d) publish clear guidelines on how they will deal with complaints, handle abuse reports, and
cooperate with criminal investigations.
(26) At present, whilst some sites offer all of these and more, others provide none, and some even appear to
stonewall those trying to tackle online abuse.12 This takes us into a fourth area of improvement—enforcement.
Education, Resources, and Training for Police
(27) Where cases become too serious for a user or site to deal with, we would naturally turn to the police.
However, whilst SOCA13 does tackle cybercrime, its focus tends to divide between child sexual exploitation
and online fraud. SOCA and the Police Central e-Crime Unit14 are forming a new National Cybercrime
Unit,15 but it remains unclear whether online abuse will form part of this organisation’s remit.
(28) This means that for now at least, online abusive messages and threats are more likely to be dealt with
by ordinary police officers, and that makes it difficult for the average non-high-profile individual to find
appropriately trained police who will take their complaint seriously.16 In short, it is vital that we have the
necessary resources and training for police, including:
(a) Perception change: online offences are “real crimes” that should be taken seriously; users can’t
just resolve the problem by simply “switching off” (see para. (16) above).
(b) Basic training: unsurprisingly, non-specialist officers appear to be largely unaware of the sheer
scope of online “misbehaviours” that exist, how online attackers work, their strategies of
concealment, standard denial excuses (eg “yes it came from my computer but I didn’t send it”),
methods of gathering evidence, the investigation that needs carrying out, etc. This is an issue
that can only be resolved with extensive training. Without it, officers are likely to shy away
from or find excuses for avoiding such areas, in turn institutionally embedding the concept that
they only deal with “real” (ie offline) crime.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
www.nspcc.org.uk/Inform/resourcesforprofessionals/onlinesafety/statistics-online-safety_wda93975.html
http://consumers.ofcom.org.uk/2009/12/child-internet-safety-strategy-launched/
www.beatbullying.org/
www.childline.org.uk/Explore/Bullying/Pages/online-bullying.aspx
www.childnet.com/
www.getsafeonline.org/
www.iwf.org.uk/
www.nspcc.org.uk/help-and-advice/for-parents/online-safety/online-safety_wdh99554.html
www.saferinternet.org.uk/advice-and-resources
www.nhs.uk/Livewell/Bullying/Pages/Cyberbullying.aspx
www.thinkuknow.co.uk/
www.theguardian.com/society/2013/aug/09/cyberbullying-mother-fight-askfm
www.nationalcrimeagency.gov.uk/
http://content.met.police.uk/Site/pceu
www.gov.uk/government/policies/keeping-the-uk-safe-in-cyberspace/supporting-pages/setting-up-a-national-cyber-crime-unit
www.theguardian.com/society/2013/aug/09/cyberbullying-mother-fight-askfm
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w6 Culture, Media and Sport Committee: Evidence
(c) More resources: even were the above not an issue, officers also appear to be largely underresourced in terms of time. The solution to this seems to be an extension to the remit of CEOP,
SOCA, or the newly-proposed NCU (see para. (27) above).
Legislation and Guidance
(29) Assuming that a case is investigated by the police and the CPS decides to pursue criminal charges,
many of the Acts that might by employed (eg Malicious Communications Act 1988, §5 of the Public Order
Act 1986, Protection from Harassment Act 1997, Communications Act 2003) were created before the advent
of major social networks, and so are imperfectly equipped to deal with the new behaviours that these sites
have created.
(30) For instance, the most recent, relevant legislation, the Communications Act, came into force in July
2003. However, an Act of this magnitude—more than 250,000 words spanning over four hundred sections—
takes years to write and enact. The section relevant to online behaviour (§127) is therefore far more heavily
informed by the internet of the 1990s than that of the new millennium. Meanwhile, massive, ubiquitous social
networks like Facebook and Twitter were founded 2004 and 2006 respectively. As sites like these have evolved,
so too have the behaviours associated with them, yet the legislation remains largely unchanged.
(31) Whilst the CPS published guidelines17 on prosecuting cases involving communications sent via social
media in June this year, like much current legislation designed for supposedly-similar offline behaviours, these
guidelines do not explicitly address the critical differences that the online environment offers:
(a) the speed with which content can be reproduced;
(b) the breadth and sensitivity (eg family, school peers) of audience that can be reached;
(c) the inability of targets to entirely eradicate malicious content in some cases; and
(d) the expense and difficulty (as outlined above) of prosecuting even very serious online offences.
(32) It is therefore worth considering whether these Acts need updating, or whether we actually need a new
Online Communications Act that specifically covers the increasing range of abusive behaviours on the internet.
(33) This difficulty is further exacerbated by a culture in which offline crime still seems to be seen as more
serious than any supposed online counterpart. For instance, the CPS provides useful guidance on behaviours
such as stalking,18 harassment,19 and school bullying,20 but if one reads these guidelines, it seems that the
offline version is considered the norm, and any online variant, where it is even acknowledged to exist, is simply
considered a sub-type.
(34) This overlooks the fact that supposedly equivalent behaviours like cyberstalking, cyberharassment, and
cyberbullying can have their own unique attributes, methods, and consequences which require dealing with in
fundamentally different ways. Indeed, what little the CPS has to say about online antisocial behaviours tends
to be vague.21
(35) In short, it appears that we need CPS guidance for the police, court system, legal practitioners, etc.
that would provide definitions for, and explanations of the various antisocial online behaviours (eg trolling,
cyberbullying, cyberstalking, etc), and information with regards to issues of jurisdiction and proof.
(36) Finally, across the European Union, the Safer Internet Programme22 is promoting self-regulation,23
and already, several high-profile corporations have signed up to a set of Safer Social Networking Principles.24
This is a great start, but these principles are purely voluntary, and whilst it would be inappropriate to enforce
such principles across all EU-based social networks, it is worth returning to the example of Ms Smith, and to
ask whether those profiting from social networks should have similar duties of care towards their site users.
Conclusion
(37) The full potential of the internet is yet to be realised. If we are to fully enjoy the benefits, and reduce
online threatening and abusive behaviour, my belief is that we must approach this in a comprehensively and
holistically. I hope the above may offer some ways in which to achieve this.
September 2013
17
18
19
20
21
22
23
24
www.cps.gov.uk/consultations/social_media_guidelines.pdf
www.cps.gov.uk/legal/s_to_u/stalking_and_harassment/#a02b
www.cps.gov.uk/legal/s_to_u/stalking_and_harassment/#a02a
www.cps.gov.uk/legal/v_to_z/youth_offenders/#a25
www.cps.gov.uk/southwest/cps_southwest_news/news_articles/in_the_dock__prosecuting_cyber_bullying__electronic_stalking_
and_trolling/
http://ec.europa.eu/digital-agenda/self-regulation-better-internet-kids
http://ec.europa.eu/digital-agenda/self-regulation-better-internet-kids
http://ec.europa.eu/digital-agenda/sites/digital-agenda/files/sn_principles.pdf
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w7
Further written evidence submitted by Dr Claire Hardaker
This is a brief note to add to a previous document submitted to the Culture, Media and Sport Committees
inquiry into Online Safety. In brief, the CASS Centre at Lancaster University was successful in their bid for
ESRC funding on a project that aims to investigate the rape threats sent by Twitter trolls to a number of high
profile women (ES/L008874/1). The project will investigate what the language used by those who send rape/
death threats on Twitter reveals about…
1. Their concerns, interests, and ideologies; what concept do they seem to have of themselves and
their role in society?
2. Their motivations and goals; what seems to trigger them? What do they seem to be seeking?
3. The links between them and other individuals, topics, and behaviours; do they only produce
misogynistic threats or do they engage in other hate-speech? Do they act alone or within
networks?
We’re hoping that this project will offer timely insight into an area where policy, practice, legislation, and
enforcement is currently under intense scrutiny and requires this type of research to help shape future
developments. As such, we would be pleased to pass on the results of this project as they come in, and it goes
without saying that we would be delighted to also make available all of the final, official findings, reports,
articles and so forth at the end of the project.
October 2013
Written evidence submitted by the Remote Gambling Association
Introduction
1. The Remote Gambling Association (“the RGA”) has noted that the Culture, Media and Sport Committee
has decided to investigate a number of aspects of online safety that are currently raising concerns, in particular
how best to protect minors from accessing adult content. We would, therefore, like to set out the experience
of the online gambling sector and the success we have achieved in addressing a number of key risks.
2. By way of introduction, the Remote Gambling Association (RGA) is the largest on-line gambling trade
association in the world, representing the world’s largest licensed and stock market-listed remote gambling
operators and software providers. The organisation provides the remote gambling industry with a single voice
on all issues of importance to regulators, legislators, and key decision-makers. Our members are listed on our
website which is www.rga.eu.com.
3. The members of the RGA are committed to the protection of young and vulnerable people. We promote
responsible gambling and have worked with regulators and legislators to ensure that the rules affecting
gambling are both workable and proportionate to the harm that could be caused.
Why Regulated Remote Gambling is Different to other “Adult” Products
4. The Gambling Act 2005 allowed for a wider range of advertising of gambling products in Great Britain.
To be able to advertise a gambling operator has to hold an operating licence issued by the Gambling
Commission, or an equivalent licence issued by an EU gambling regulator or an overseas regulator which
issues licences with equivalent standards to the UK regulator. These licences require that before bets can be
settled the customer is over 18 and has had his or her identity verified.
5. As far as we are aware, no other adult service providers are required by law to ensure that their customers
are over the age of 18. This puts the regulated online gambling industry in a different position to other ecommerce sectors. because there are mandatory safeguards in place, but especially where children are
concerned we believe that the principles at least should be applied equally.
The Current Requirements for Licensed Online Gambling Operators
6. The requirements of the licences held by our members mean that all operators require customers to open
accounts. As part of this process their identity and age will be verified.
7. UK and similar legislation around Europe that applies to online gambling requires very high standards of
age verification and it is commonly an offence to knowingly accept bets from minors. We support this position
and have proactively worked with third party suppliers to improve the range and reliability of data that is used
to identify whether someone is under 18 or not.
8. In practice this means that during the account opening process potential customers are asked to provide a
range of information and, amongst other things, this is checked by third parties (such as Experian) to confirm
the customer’s identity and payment details. Details will be checked against any databases (public or private)
to which they have access and they will see whether the applicant has a credit history, which of course will
not be the case for children.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w8 Culture, Media and Sport Committee: Evidence
9. Under the Gambling Commission’s Licence Conditions and Codes of Practice, customers will not be able
to withdraw any winnings until the age verification process has been satisfactorily completed. In any event, if
the customer has failed to satisfactorily complete the age verification checks within 72 hours of registering a
relevant bank card and placing their first bet, the operator will freeze the account, prevent the customer from
gambling any further until they have satisfactorily completed these checks and, if upon completion of age
verification the customer is ultimately shown to be underage, the operator will return all stakes to the young
person.
10. In practice the necessary checks for the vast majority of account applicants can be undertaken almost
immediately, but the 72 hour window mentioned above allows companies if required to seek documentary
evidence of identity and age if the electronic verification does not provide the right level of assurance.
11. There are occasionally cases where people allege that a child has opened an account in their name or
has illegally used the existing account of an adult. In reality this happens very rarely and always comes to light
when funds are withdrawn from the adult’s bank account (and even then it presupposes that the minor has
knowledge of the requisite passwords etc in order to gain access). If someone believes that their credit or debit
card has been used fraudulently they should contact the operator and their bank and then the cardholder will
be able to challenge the charges made, in exactly the same way they would if the card had been used to
purchase other goods or services. Although we would suggest strongly that there should be a greater element
of parental control in these situations, if it can be proven that the bets were placed by someone who is under
18 then the bets will be cancelled and stakes returned (see para 9 above).
12. With regard to advertising, only suitably licensed operators are able to market gambling products in the
UK. This provides another safeguard and all advertising, irrespective of where the operator is based, must
comply with the relevant requirements from the Advertising Standards Authority and OFCOM which have
clear and enforceable rules to prevent the targeting of advertisements at minors. In addition, the British
gambling industry’s own advertising code of practice has put in place further protections, including the 9.00pm
television watershed, for all gambling except for betting around live events and bingo. It all bans any form of
gambling sponsorship from appearing on children’s products.
Online Gambling as an Example of Regulatory Best Practice of Online Child Protection
13. In July 2011 the European Commission sought views on these issues as part of its consultation on a
Green Paper on Online Gambling in the Internal Market.
14. Some of the questions asked how on-line age controls are imposed. A response was submitted by the
Children’s Charities’ Coalition on Internet Safety, which includes Action for Children, Barnardo’s, BAAF, The
Children’s Society, NCB and NSPCC. Their response stated:
“Since the online age verification laws came into force in the UK in September 2007, the children’s
organisations have not been made aware of a single instance where a child has beaten the system
and got online to gamble. There have been instances where a child has ‘borrowed’ a parent’s credit
or debit card and has been able to pass themselves off as the parent, but that raises different issues.
There is nothing that laws can do about that, neither is there an easily foreseeable technology fix
that can solve it. However, we are not aware of any instances where a child was able to lie about
their age and get away with it in an online environment, as they used to do quite easily before the
law was changed. By comparison it may not be so difficult to ‘PhotoShop’ a fake ID to use in a real
world setting.
The law should never require an online gambling [operator] to get everything 100% right 100% of
the time. That is impossible. But society is entitled to expect gambling web sites to be able to
demonstrate that they are making reasonable efforts. Asking someone to tick a box to confirm their
age, without more, is not reasonable. It is negligent.”
15. On the 23 November 2010 Claire Perry MP speaking in an adjournment debate on the topic of Internet
Pornography in the House of Commons looked to the controls on the online gambling industry as an example
of how age verification works. She said:
“The previous Government sensibly introduced workable age-verification restrictions on online
gambling sites in 2005, an excellent model that works well and searches financial and electoral
databases to verify that users are old enough to enter the site and engage in the gambling within. It
is a workable model, the previous Government introduced it, and they are to be commended for
doing so.”
16. In September 2008 the European Commission held a forum on Cross Media Rating and Classification
and Age Verification Solutions. The subsequent report stated that:
“the British model, restricting access to online gambling sites [is] seen as the most effective in
preventing access to sites using these technologies.”
17. There are other examples that could be cited, but we hope these will suffice to show that the success of
these measures is not something that the online gambling industry is merely claiming.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w9
18. Following on from this, we believe that the legal requirements to ensure that all people gambling are
over the age of 18 are the most appropriate controls for the online gambling industry. We believe that the
requirements to age verify online gamblers works. There is very little evidence that young people are attracted
to, or participate in remote gambling.
Are Young People Attracted to Online Gambling?
19. There should never be any complacency about the issue of under-18s and gambling, however, in assessing
what the most proportionate controls should be it would be wrong to assume that there is a huge interest in
this activity.
20. The only significant reference we have been able to find on whether young people are particularly
attracted to on-line gambling is in Professor Gill Valentine’s report “Literature review of children and young
people’s gambling” (September 2008). In that she stated “To-date there is not enough evidence to assess
whether the advertising/promotion of gambling—including free demonstration games available on websites—
has a direct effect on gambling participation.” While she recommended that further research is undertaken no
such research has been completed as at August 2013.
21. The main research undertaken into the prevalence of gambling among young people has been undertaken
on behalf of the National Lottery Commission. As the legal age for participation in the National Lottery is 16
the research they have undertaken is on people aged 15 and under. The most recent research conducted by
Ipsos MORI and the Social Research Institute published in July 2012 showed that 18% of 11–15 year olds say
they have gambled in the past week. Rates of participation in online gambling are lower than rates of
involvement in offline gambling. Thirteen per cent% of children say they have played free online gambling
games; this is most often through Facebook. Seven per cent of children aged 11–15 say they have some
experience of playing online gambling games for money, although around half of this can be explained by
children playing alongside parents, with their parents’ permission.
Further Protections that could be Provided to Children and their Parents
22. Advertising: a review could usefully be undertaken to consider which of the advertising rules that apply
to gambling might be extended to all adult products and whether certain rules should be more product-specific.
In saying that we are mindful of the high levels of compliance where such rules are in place. From a gambling
perspective there has been close engagement by the industry with the ASA in the development of codes. In
2009 the ASA reviewed compliance by the gambling industry with the CAP and BCAP codes and found
“fewer than 1% of the gambling product advertisements [….] seemed to breach the Code, an exceptionally
high compliance rate”.
23. Raising awareness amongst parents: in order to participate in gambling on a regulated gambling website,
customers have to open an account and have their age and identity verified. Although it is not a major problem
in terms of the number of cases, a weakness in the system is where a child uses their parent’s identity and
legitimate credit or debit card to open or use an account.
24. It is self-evident that if this can happen with gambling then it can happen with other adult products.
Parents need to understand that they should keep their credit or debit card, password and pin numbers details
safe at all times and to supervise the use of credit cards for online shopping. For example if a parent allows a
child to use a credit card to purchase a non-age restricted product such as a book they should ensure that the
child returns the credit card to them and ensure that the card number is not retained by the child.
25. We also note and support proposals that have previously been made to promote awareness about the
parental controls that are available for internet-enabled devices.
26. Highlighting tools for parents to use: The RGA has produced technical guidelines to raise standards
among industry operators and regulators. As part of its work on responsible gambling it has included the
following in that document:
“The operator’s responsible gambling page should provide a link to a recognised filtering programme
to assist customers/parents in preventing underage individuals from using gambling sites.”
27. This provision has been adopted into licences issued by UK and European regulators and is now the
industry standard. It is a practical tool for restricting child access to unsuitable sites, but is not in any sense
burdensome for operators and, again, could be used for all sites offering adult products.
Conclusions
28. As an industry which is subject to statutory regulation involving an adult product it is right that
expectations and compliance should be high. However, this has led to the industry proactively seeking out
and devising practical measures to achieve the regulatory objective of preventing under-18s from accessing
gambling products.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w10 Culture, Media and Sport Committee: Evidence
29. Experts in child protection have endorsed those efforts and, most importantly, their effectiveness.
Although it may not be appropriate to set the bar at such a high level for non-regulated industries, we do feel
that there might be lessons to learn from what we have done.
30. As ever, if it would be at all helpful we would be very willing to expand on any of the points made in
this letter,
September 2013
Written evidence submitted by Ann Farmer
I would like the Committee to take into account the dangers of online pornography to children and young
people, especially in negatively affecting them mentally and emotionally.
My research into eugenics and population has found a significant overlap between these movements and the
campaign for “value-free” sex education; both movements have historical links with the pornography industry.
The aim of the pornography industry is to maximise profits by turning people into sex addicts, searching for
sexual material on a regular basis, often of an increasingly extreme nature. Eventually they may seek more
extreme “thrills” by acting out what they have seen online: two recent cases of rape and murder of young girls,
April Jones and Tia Sharpe, have featured the use of pornography by men unknown to police.
There is also evidence that young boys are themselves becoming child molestors, acting out the pornographic
scenarios they have seen online.
Pornography dehumanises women and children, depicting them as sex toys to be used at will, often in a
brutal manner; even those women who participate willingly do not remain psychologically unharmed. It also
dehumanises the viewers of pornography, reducing them to their sexual appetites, and encouraging them to
believe that there is a “right to sex”.
Even without these extreme negative outcomes, we must face the prospect of a society where the younger
generations have learned to view other people through the distorted prism of a perverted sexual desire; the
history of pornography shows that lust swiftly turns to cruelty, thus we must tackle this problem of easily
accessible online pornography if we do not wish society to be composed increasingly of uncaring sociopaths
who care chiefly about their own sexual appetites.
September 2013
Written evidence submitted by Kirsty Hopley
Online pornography is free, easily accessible and the content is extreme. Anyone can access any type of
pornography including, violence, rape, acted child pornography, s and m, extreme degradation of women and
more. There is no log in page and no age check. This is unacceptable. This is not tolerated in pubs or strip
clubs so why is it freely available online? A simple answer is to make online pornography sites operate account
only sites, the use must provide their details and have a user name. This will reduce the amount of times
people stumble upon online porn without affecting the freedom of choice for users of online porn.
Secondly, internet filters are great for parents and I am in favour of the automatic filter. It makes sense that
those who wish to view pornography should be the ones making the choice rather that everyone getting it and
then having to choose not to have access. However, there are serious concerns.
I have Talk Talk Homesafe settings which are good in theory. They allow the user to turn on and off filters.
In practice this rarely works. I need to turn filters off to read material for my work but usually when I want to
turn it back on, it freezes or takes hours. Secondly, some sites are blocked which do not contain pornography
such as rape support sites. Similarly, even with my homesafe settings on I could still type in sex or porn and
get some extreme sexual images through google images. It is crucial that this is monitored and dealt with
swiftly. Clearly the technology is already there, it just needs monitoring.
Next issue is advertising. Many sites have pop ups which can lead to sexual material. I have seen instances
of this on school approved websites and on kindle children’s books (Alice in Wonderland was one). Advertising
that is aimed at adults should never appear on sites that are aimed at children. This should be a criminal offence
as it is grooming.
Online safety is something parents need to understand to take seriously. With channels such as Cebebies
encouraging very young children to go online and the rise of little children having tablets, education of parents
is key. My suggestion is that this is discussed by Health Visitors in the early years checks. It should be a face
to face discussion of the dangers and a leaflet given of how to keep up. Every device that is sold must have
an easy to understand guide for parents to set blocks and this should be discussed at purchase. This needs to
be regulated so that it happens every time. It is not like the days when parents had to make sure toddlers didn’t
choke on a piece of lego, parents do not understand the technology that their kids do. The digital age has
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w11
approached so quickly. Following this, schools should offer compulsory help for parents to understand keeping
children safe online. The world has moved on and parenting needs to catch up.
This is an issue which the government needs to support parents to protect children. The starting point is to
regulate the porn industry (like we do with tobacco, alcohol and gambling industries). It is unbelievable that
the porn industry is given such a free reign to provide such an unrealistic and harmful portrayal of what is
actually a wonderful act that humans should enjoy.
September 2013
Written evidence submitted by Dr Peter Nelson
PORNOGRAPHY: BRITAIN’S BIG PROBLEM
A young man walks into a newsagent and buys a copy of the Sun or similar newspaper. On page 3, he sees a
picture of a young woman wearing little or no clothing. What goes through his mind?
Does he think, “I must get a good job, earn a living, save some money, get married, have children, and help
bring them up”?
Or does he think, “I must go out at the weekend, have some drinks, find a woman, and have sex with her”?
If the latter is the case, then society has a big problem.
Consequences
In the first place, young men who think like this do not make good fathers. They make women pregnant,
but take little or no interest in bringing up the children. This leaves women bringing up children on their own.
While some do this remarkably well, the evidence is that children from single-parent families do less well, and
are less well-behaved, at school.1 Even in homes where a father is present, there can be domestic violence
which has an adverse effect on children. Many of the problems in education today can be traced back to poor
parenting. Teaching youngsters who lack discipline and encouragement to study is very difficult.
Secondly, the open display of sex in the media brings sex to the attention of children from an early age. This
leads them to engage in sexual activities before they are old enough to appreciate the consequences. Teenage
pregnancy rates in Britain are very high. This is despite attempts to bring them down by promoting the use of
contraceptives in schools.2 These fail because they raise children’s interest in sex still further. Even good sex
education is limited in what it can do, given the big influence the media have on young people.
A recent NSPCC study reveals that nine out of ten young people between 13 and 17 years old in England,
Scotland, and Wales have experienced some form of intimate relationship.3
Thirdly, pornography encourages men to be promiscuous. Promiscuity leads to transmission of sexually
transmitted-diseases (SDIs). Health agencies promote the use of condoms to try to stop this, but condoms are
not always used, and frequently fail. SDI levels in Britain are very high and rising. The number of new
diagnoses in the UK is now running at 1.3% of the population per year.4
Fourthly, pornography inflames men’s sexual passions. Men who cannot control these may attack women.
The incidence of rape and indecent assault on women in Britain is very high. On present figures, one in ten
women will be raped or assaulted during their lifetime.5 The NSPCC study referred to earlier found that as
many as one in three girls in Britain have experienced some form of sexual violence from a partner.3
Adult pornography may also be implicated in the development of paedophilia.6
Fifthly, sexual images of women contribute to the pressure on men to assert their masculinity. They see
women vaunting their sexuality and want to show off theirs. This leads them to engage in loutish behaviour,
drink heavily, drive recklessly, and so on. Anti-social behaviour is a big problem in Britain today, and much
of it is macho-driven.
Sixthly, pornography contributes to mental illness in society. Many mental problems have a sexual
component. Emotions are interconnected, and difficulties in one area can lead to problems in another. Some
men have so much difficulty coping with pornography that they have set up a help group, “Men and Porn”.
Many suffer on their own, and are reluctant to admit that they have sexual problems. The incidence of mental
illness in Britain today is very high and rising.
While most pornography is aimed at men, it encourages women too to think of sex as something to be
pursued as and end in itself, and not just as part of a stable relationship within which children can be brought
up. Pornography aimed at women will do this even more.
Proof
The contribution of pornography to these various problems is evidenced by the extent to which they have
grown over the last sixty years. Sixty years ago, media standards were high, and there was almost no
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w12 Culture, Media and Sport Committee: Evidence
pornography on sale to the public. Since then, standards have been relaxed, and the volume and explicitness
of pornography have increased. At the same time, the incidence of all the above problems has grown.
This is shown by the following comparisons, taken from my 2005 briefing paper.4 There are large rises in
all the indicators from the 1950s to the 2000s.
Teenage pregnancies
The following chart compares the number of girls in England and Wales becoming pregnant below the age
of 14 per thousand members of their age group.
1955 |
2002 |
| 0.1
| 1.1
Sexually-transmitted infections
The chart below compares the number of new cases of syphilis, gonorrhoea, herpes, and genital warts in
England and Wales per thousand members of the population. The comparison is limited to these STIs because
there was no HIV/AIDS in the 1950s, and no records were kept of chlamydia or other infections.
1955 |
2004 |
|
0.7
| 2.1
Sexual crime
The following chart compares the number of rapes and sexual assaults on women and girls per thousand
members of the female population. This is based on the number of offences reported to the police. Reporting
of sexual crimes against women has become easier since the 1950s, but against this, many women take
protective measures that they did not take then (eg avoiding going out alone at night).
1955 |
2004/5 |
|
0.3
| 1.4
Family breakdown
Sexual infidelity is one cause of family breakdown. (Other factors are money, alcohol, and drugs.)
The chart below gives the number of lone parents with dependent children in Britain as a percentage of the
total number of families with dependent children.
1955 |
2004 |
|
5
| 24
Mental illness
Sexual problems are one cause of mental illness. A measure of the level of mental illness in society is the
number of anti-depressant drugs people use. The following chart compares the number of prescription items
for these drugs in the community in England per person. Anti-depressants were introduced in 1957.
1960 |
2003 |
|
0.05
| 0.55
Macho violence
Much of the violence in society is driven by machismo. Alcohol is often implicated, but heavy drinking by
men is itself, in many cases, macho-driven. Violence in the media and drugs are other factors.
The chart below compares the number of violent offences against the person recorded by the police in
England and Wales per thousand members of the population. The comparison is only very approximate because
the police have changed the way they count violent crimes. I have doubled the 1955 figure to allow for this.
1955 |
2004/5 |
0.4
| 19.5
This evidence leaves little doubt that pornography is very harmful. This finding contradicts the conclusions
of Home Office reports on the effects of pornography, published in 1979 and 1990.7 I have discussed the
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w13
weaknesses in these reports elsewhere.8 The biggest one is that they do not answer the question, “If
pornography is not a major cause of the sexual problems now afflicting society, what is?”
A more recent study of hard pornography by the Ministry of Justice concludes that this does have a harmful
effect on adults.9 Other recent studies, including one by Dr Linda Papadopoulos for the Home Office, similarly
conclude that exposure to pornography has a harmful effect on children.10 Dr Papadopoulos’ report is alarming.
Would that the proponents of permissiveness in the 1960s had heeded the warnings of those who predicted
what the consequences of this would be.
Remedy
The remedy for these problems is for the Government to enact legislation to restore standards in the media
to where they were sixty years ago. This will be difficult. Users of pornography will protest, and liberals will
complain about loss of freedom. The above problems are, however, huge and need to be addressed.
My suggestion is that the Government sets up a Royal Commission to re-examine the effects of pornography
and to make recommendations. Acting on a Royal Commission’s report would be easier for Government than
acting on its own.
I have answered objections to raising standards and discussed possible legislation elsewhere.11 The answer
to the objection of loss of freedom is that pornography takes away even more freedom—from women who
would like to be able to go out alone at night, from young mothers who would like to do what their other
friends do, from men who find pornography addictive, from old people shut in for fear of macho-driven yobs,
and so on.
Pornography makes for a less civilized society. It is an evil, and needs to be treated as such.
References
1. Lords and Commons Family and Child Protection Group report 1998.
2. Cf. M Henderson et al, “Impact of a theoretically based sex education programme (SHARE) delivered by
teachers on NHS registered conceptions and terminations: final results of cluster randomised trial,” BMJ, 2007,
334, 133; Meg Wiggs et al, “Health outcomes of youth development programme in England: prospective
matched comparison study,” BMJ, 2009, 339, b2534.
3. Christine Barter et al, Partner Exploitation and Violence in Teenage Intimate Relationships, NSPCC, 2009.
4. P G Nelson, Sex in the Media: Is it Harmful? mediawatch-uk, 2005.
5. Andy Myhill and Jonathan Allen, Rape and Sexual Assault of Women: the Extent and Nature of the Problem,
Home Office Research Study 237, 2002.
6. P G Nelson, Nudity and Sexual Activity in the Media, Whittles, 1998, Sect. 1.5.2.
7. Report of the Committee on Obscenity and Film Censorship, chairman Bernard Williams, 1979; Dennis
Howitt and Guy Cumberbatch, Pornography: Impacts and Influences, 1990.
8. Nudity and Sexual Activity in the Media, Chap. 2.
9. Catherine Itzin, Ann Takets, and Liz Kelly, The Evidence of Harm to Adults Relating to Exposure to Extreme
Pornographic Material, Ministry of Justice Research Series 11/07, 2007.
10. Michael Flood, “The harms of pornography exposure among children and young people,” Child Abuse
Review, 2009, 18, 384; Linda Papadopoulos, Independent Review into the Sexualisation of Young People, Home
Office, 2010.
11. Nudity and Sexual Activity in the Media, Chaps. 3–4.
September 2013
Written evidence submitted by the British Board of Film Classification
Executive Summary
— The British Board of Film Classification (BBFC) is the UK’s regulator of film and video. The
BBFC’s primary aim is to protect children and other vulnerable groups from harm through
classification decisions which are legally enforceable and to empower consumers through content
information and education.
— Although the ecology of the internet is different from that of the more regulated offline environment,
the risk of harm, particularly to minors, from both legal and illegal material is at least as great online.
Government, industry and regulators should work together to reduce the harm risk wherever possible.
There is a case for more joined up co and self-regulation online drawing on existing, effective, and
often industry-led models, encouraged and supported by Government.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w14 Culture, Media and Sport Committee: Evidence
—
—
—
—
The starting point for co- and self-regulatory models to protect children online should be clarity of
approach and consumer and industry support. The BBFC sets out in its submission the factors for
the success of online co- and self-regulatory systems: child protection at its core; effective labelling
of content; broad coverage; low cost; efficient, flexible and innovative.
The BBFC is ready to continue to make available its expertise and experience, working with
Government, industry and other regulators to improve child protection online. Examples of the
BBFC’s involvement in this area include:
— voluntary self-regulation by the BBFC of Video-on-Demand (VOD) content applying granular
age ratings and content advice;
— the BBFC-run regulation of mobile content, including websites, using a binary classification
system based around the BBFC 18/R18 standard which enables the filtering of content and
protection of minors; and
— development by the BBFC and NICAM of a User Generated Content (UGC) ratings tool. This
self-regulatory questionnaire is tailored to reflect national differences and is designed to be
completed by those uploading the content or the audience, allowing the use of online filters to
prevent children accessing inappropriate content and the facility to report abuse.
Online pornography is an important example of where co or self-regulation has not worked well.
The BBFC removes any material from pornographic works it classifies which is potentially harmful
or otherwise illegal under the Video Recordings Act (VRA), including so-called “rape porn”. But
online, where the VRA and its harm test do not apply, harmful material including “rape porn” is
available. Greater regulation of such adult content, and content which the BBFC will not classify at
any category, would be welcome for the better protection of minors.
There are some obvious areas where Government intervention may be appropriate to better protect
children online. Lack of industry self-regulation of online pornography is leading to children being
able to access inappropriate and potentially harmful content. Equally, industry could do more, for
example through effective moderation or through the age rating of online music videos, to ensure
that their service is safe for children and the vulnerable online.
1. Introduction
1.1 The British Board of Film Classification (BBFC) is grateful for the opportunity to respond to the Culture,
Media and Sport Committee Inquiry into Online Safety.
1.2 The BBFC is an independent, non-governmental body which classifies films and videos distributed by
way of packaged media and, increasingly, on the internet. The BBFC’s primary aim is to protect children and
other vulnerable groups from harm through classification decisions which are legally enforceable and to
empower consumers, particularly parents and children, through content information and education. The BBFC
will not classify material which is illegal in the UK, including content which promotes terrorism or which
features indecent images of children. The BBFC became the independent regulator of content delivered by
mobile networks in the UK from 2 September 2013. The Director of the BBFC, David Cooke, sits on the
Executive Board of UK Council for Child Internet Safety (UKCCIS) in recognition of the BBFC’s role in the
protection of children online.
1.3 The BBFC classifies films and videos/DVDs according to its Classification Guidelines, with ratings
ranging from U for Universal to R18 which are restricted videos only allowed to be sold in licensed sex shops.
The BBFC’s Guidelines are the result of extensive public consultation with over 10,000 people across the UK
being consulted during the most recent Guidelines consultation in 2013. Research demonstrates that the public
agrees with the BBFC’s classification decisions most of the time. 92% of recent film viewers agreed with
classification of films they had seen recently. (Goldstone Perl, Bernice Hardie 2013); and recognises and
understands the BBFC’s symbols 89% of the time (rising to 97% for parents) (TNS, 2007).
1.4 The BBFC also publishes (on its website and Apps) more detailed information, BBFCinsight, aimed
primarily at parents, which outlines the key content issues in a film and thereby enables viewers to make fully
educated viewing decisions for themselves and their families. This information is available both when the
content is distributed by way of packaged media and online. Independent research (Slesenger, 2009) suggests
that 86% of parents of primary schoolchildren find this information useful.
1.5 The BBFC runs an extensive education programme both online and through a series of seminars at the
BBFC and throughout the country, with visits to schools and colleges, speaking to over 12,000 students in
2012 on media literacy and the classification process. The BBFC further contributes to a number of media
literacy initiatives run by industry and child protection groups, including Childnet UK, advising children and
adults on how to keep themselves and their family safe online.
2. The BBFC and Home Entertainment Industry Self-regulation Online
2.1 For historical reasons, the current statutory regime for film and video/DVDs does not cover online
content. The Video Recordings Act was originally passed in 1984 before digital downloads of videos/DVDs.
However, media convergence is increasingly blurring the distinction between physical and online material.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w15
2.2 Independent research commissioned by the BBFC shows that while the public considers that the internet
brings greater choice, freedom and flexibility, the majority of viewers still consider it important to be able to
check the suitability of audio-visual content they download. 85% of the public consider it important to have
consistent BBFC classifications available for Video-On-Demand (VOD) content; rising to 90% of parents of
children under 16 (Bernice Hardie, 2011).
2.3 In recognition of public demands for regulatory protection online, the BBFC has worked in partnership
with the home entertainment industry on a number of voluntary, self regulatory services which bring trusted
BBFC classification standards and well known age ratings and symbols online. BBFC digital services offers
labelling and content advice services designed for content providers and platform owners (digital video
services). These services cover all forms of video-on-demand including subscription models (SVOD) and
transactional (TVOD).
2.4 The BBFC has now rated more than 200,000 videos for online distribution by such companies as Warner
Bros, Walt Disney, 20th Century Fox, Paramount, Universal, Sony; with BBFC ratings used by platforms such
as iTunes, Netflix, Blinkbox/Tesco, Sainsbury’s, Dixons, BT Vision and Talk Talk. The new industry initiative
known as UltraViolet—which promotes the digital ownership of content—is also working with the BBFC.
Displaying BBFC labelling enables consumers to make informed choices when purchasing retail and rental
digital video, thereby empowering parents to protect their children.
2.5 BBFC digital services include a ratings service aimed specifically at content that is only available online
and that would otherwise not be provided with a rating. This service—known as Watch and Rate—is becoming
more popular and well-used. Compared with the same period in 2012, there has been a 141% increase in
industry take up as of 10 September 2013.
3. Online Pornography
3.1 It is generally accepted that exposing children to pornography is potentially harmful to them and is likely
at the very least to impair their development. This is why many countries, including the UK, have enacted
measures to prevent children from accessing pornography.
3.2 In the UK, the BBFC classifies all pornographic content released both on video and theatrically. The
BBFC removes any material from pornographic works which is potentially harmful or otherwise illegal. As
well as policing the border between legal and illegal pornography, the BBFC polices the border between the
strongest, hardcore pornography, and the less strong, softcore pornography. The BBFC classifies hardcore
pornography as R18, which means that it may only be supplied through licensed sex shops, as an extra
precaution against underage viewing. However, the risk of children accessing even the strongest legal
pornography is far greater online. In addition, there are fewer effective controls on the distribution online of
pornography which the BBFC would not classify at any category.
3.3 Online pornography is an important example of where co or self-regulation has not worked well. While
the majority of the home entertainment industry has chosen to opt for voluntary self regulation, no members
of the adult industry currently use the BBFC’s digital services. The market imperatives which in the UK have
driven the mainstream home entertainment industry to adopt best practice self regulation, work in the opposite
direction in relation to adult content. While the mainstream video industry recognises that parents want the
reassurance of trusted labels on online content, much of the adult industry wants to advertise its product
as “uncensored”.
3.4 The BBFC removes any material from pornographic works it classifies which is potentially harmful or
otherwise illegal under the VRA, including so-called “rape porn”. The BBFC intervenes to remove from a
pornographic work any portrayal of sexual activity which involves lack of consent, whether real or simulated.
The BBFC will also remove material, including dialogue, which is likely to encourage an interest in sexually
abusive activity, including rape. In short, “rape porn” does not exist in works classified by the BBFC. But
online, where the VRA and its harm test do not apply, “rape porn” is available. The BBFC believes that
consideration needs to be given into how protections that are applied to physical R18 material can also be
applied online if much of the adult industry continues to refuse to participate in self-regulatory measures. The
BBFC therefore welcomes the Prime Minister’s declaration on 22 July 2013 that Government will ensure:
“that videos streamed on line in the UK are subject to the same rules as those sold in shops. Put
simply—what you can’t get in a shop, you will no longer be able to get online.”
3.5 The Government has recognised the dangers of extreme pornography and in 2008 made possession of
extreme pornography an offence under the Criminal Justice and Immigration Act. A BBFC classification is a
defence against a prosecution under this Act therefore purchasing a legally classified work is a protection
against inadvertently possessing extreme pornographic material. The BBFC regularly assists Local Government
trading standards officers in ensuring that pornographic material has been classified by the BBFC and restricted
for sale to licensed sex shops. However, these methods of enforcement are not available online.
3.6 While physical video works rated R18 by the BBFC may only be supplied to adults through licensed
sex shops, restrictions to prevent children accessing such material are not maintained online, except for UK
services regulated by the Authority for Television On Demand (ATVOD). These services form a very small
proportion of the market, making it easier for those under 18 to access potentially harmful pornography. In
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w16 Culture, Media and Sport Committee: Evidence
addition, ATVOD can only intervene and impose rules to require that the material must be made available in
a manner which secures that children will not normally see or hear it, if a programme distributed online by a
UK provider contains material which might seriously impair the physical, mental or moral development of
persons under the age of 18.
3.7 The BBFC believes that there is a case to strengthen the definition of content online likely to seriously
impair the development of minors so that it is clear that hardcore pornography constitutes such content. One
option might be an amendment to the Audiovisual Media Services Directive so that explicit depictions of real
sexual activity, or activity involving realistic infliction of pain or injury, or real or simulated lack of consent to
sexual activity, must be provided in such a manner which ensures that under 18s cannot normally see or hear
it. Such an approach would be consistent with that in the Government’s paper: Connectivity, Content and
Consumers: Britain’s digital platform for growth.
4. The BBFC and Mobile Networks’ Self-regulation of Online Content
4.1 From 2 September 2013, the BBFC became the independent regulator of content delivered via the UK’s
four main mobile networks (EE, O2, Three and Vodafone). Under this new self regulatory system, the BBFC
has created a Classification Framework, based on the trusted, transparent and consistent standards set out in its
Classification Guidelines, which specifies what content should be placed behind access controls/filters. The
Classification Framework is a living document which will be updated regularly to reflect evolving public
attitudes and societal concerns.
4.2 The standards in the BBFC’s Classification Guidelines are used to ensure content which would be age
rated 18 or R18 by the BBFC will be placed behind access controls and internet filters by mobile operators to
restrict access to that content by those under 18. This content includes pornography and other adult sexual
content, pro-smoking and pro-Ana (anorexia nervosa) websites and content which promotes or glorifies
discrimination, drug misuse or real life violence. The BBFC and the mobile operators believe that this
partnership will be trusted by the public, more transparent than the previous mobile regulatory regime and
better able to adapt to changes in societal attitudes and concerns in relation to child protection.
4.3 The BBFC believes that child protection would be improved if those operating public WiFi services in
places regularly visited by minors applied the same, consistent standards as those used by the mobile operators.
The BBFC would be able to offer assistance in setting and applying these standards, as it has done in
partnership with the mobile operators for content accessed via mobile networks.
5. User Generated Content Online
5.1 The BBFC was a third party member of the EU’s CEO Coalition to make the Internet a better place for
children. In response to a call for action from the European Commission, members of the Coalition in 2012–13
undertook to take positive action to make the Internet a better place for children in the European Union. The
Coalition aimed to make the internet safer for children and the vulnerable through action in five core areas:
simple and robust reporting tools, age-appropriate privacy settings, wider use of content classification, wider
availability and use of parental controls and effective takedown of child abuse material.
5.2 In response to this initiative, the BBFC and its counterpart, the Dutch regulator, the Netherlands Institute
for Classification of Audio Visual Media (NICAM) together developed a tool for ordinary people to age rate
User Generated Content (UGC) across different countries and platforms and according to recognised and trusted
national standards. UGC is growing in significance and needs to be addressed to allow parents the tools to
prevent children and the vulnerable from accessing inappropriate content.
5.3 The UGC tool, using a single, simple, free to complete questionnaire, instantaneously produces an age
rating which can be shown on the video hosting platform and linked to internet filters. The ratings differ from
country to country to reflect different national concerns over content. For example, strong language is an issue
for consumers in the UK but not in the Netherlands.
5.4 The tool is simple to use and understand. It contains six questions about the content of the UGC, on
behaviour (including racism and other discriminatory behaviour, self harm and suicide), drugs, horror, language,
sex/nudity and violence. Completing the questionnaire takes less than two minutes. It also includes a facility
for viewers to report content which in their view is abusive or even illegal, such as material intended to
promote terrorism or child sex abuse images.
5.5 The questionnaire is flexible. It may be completed by those uploading content or it may be completed
by those viewing the content. The ratings can then be linked to online filters. This new initiative will shortly
be trialled by Mediaset in Italy and the BBFC and NICAM are looking for trial partners elsewhere across the
EU. This initiative could make the online world safer for children and has been welcomed by the EU
Commission’s CEO Safer Internet Coalition and the UK Government.
5.6 The BBFC would be happy to demonstrate to the Culture, Media and Sport Committee how this tool
could help to make the internet safer for children.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w17
6. Moderation
6.1 One core area that the BBFC believes could significantly improve child protection online is effective
moderation of sites that appeal to children and teenagers. The case of instances of inappropriate behaviour on
the site Habbo Hotel and Ask.fm demonstrates the importance of robust moderation and effective follow up
action if children are to be properly protected.
7. Exemptions from the Video Recordings Act and Online Music Videos
7.1 When the Video Recordings Act (VRA) was passed over 25 years ago certain video works were made
exempt from classification by the BBFC because they were considered unlikely to be harmful. The threshold
at which music, sport, documentaries and religious works lose their exemption from the VRA remains
extremely high despite the content of these exempt works having changed beyond all recognition since 1984.
This high threshold has meant that inappropriate and potentially harmful content in such works is exempt from
statutory classification, allowing it to be legally supplied to children.
7.2 On 24 May 2013, following a consultation, the UK Government announced that it plans in 2014 to lower
the exemptions threshold for all currently exempt material in order to prevent children accessing potentially
harmful material. Once implemented, the decision will improve the protection children enjoy from potentially
harmful media content by ensuring that video content such as drug misuse, strong violence, racist language,
and certain sexual content falling short of actual sex will no longer be legally able to be freely supplied to
children. Instead, the BBFC will classify such content to keep it away from vulnerable and impressionable
children.
7.3 However, there is particular concern about the content of currently exempt music videos and online music
videos, highlighted among others by Reg Bailey in his review into the commercialisation and sexualisation of
Children which resulted in the report Letting Children be Children. The Government’s proposed amendment
to the VRA will only bring physical music videos within the BBFC’s regulatory ambit. In the Department for
Education’s Progress Report on the Bailey Review in May 2013, the Government stated:
“On music videos, we want industry to develop solutions no later than the end of the year to ensure
online videos—particularly those that are likely to be sought out by children and young people—
carry advice about their age-suitability and content.”
7.4 The BBFC hopes to work with the recorded music industry towards the goal of achieving well understood
and trusted age ratings and content advice for online music videos as we have done successfully with the home
entertainment industry in relation to other online videos. One company—Universal Music—has begun to
submit long form music video content to the BBFC for online age ratings. A number of platforms which use
BBFC ratings through BBFC Digital Services have indicated that they would carry music videos if they were
age rated by the BBFC. This would then mean that the ability to download these music videos could be
controlled by age related filters which would be an important mechanism for child protection online.
8. Conclusion: The Appropriate Regulatory Environment to Protect Children Online
8.1 In a converging media world, it is the content, and not the means of distribution, that is important when
considering protection of children. The ecology of the internet is of course very different from that of the
offline environment where harm risks are well understood and regulated for. But online, the risk of harm,
particularly to minors, from both legal and illegal material is at least as great and in some cases (for example
access to legal pornography or pro-self harming and suicide content) arguably greater because of a much
greater ease of access.
8.2 In several areas, industry and regulators have worked effectively together to reduce the online harm risk.
The initiatives outlined in this response for example, which build on the BBFC’s partnerships with, among
others, mobile operators, digital video services and content providers, and our cooperative relationships with
fellow regulators both in the UK and overseas, demonstrate this. The public should continue to benefit from
these initiatives. Drawing on its experience and expertise, the BBFC stands ready to work further with
Government, industry and other regulators to continue collectively to improve child protection online.
8.3 The key criteria for any regulation should be clarity of approach and consumer and industry support.
The BBFC believes that the public policy priority should be to create an appropriate framework enabling both
existing and future effective self, co and statutory regulation to flourish, empowering consumers to inform and
protect themselves and their families.
8.4 In the BBFC’s experience the key factors for the success of an online regulatory system are:
— child protection at its core;
— effective labelling of content so that the standards are trusted and understood and the symbols
used are recognisable;
— broad coverage so that the system creates a known standard; and
— low cost; efficient, flexible and innovative service so that it can keep pace with technological
change and not be burdensome on industry.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w18 Culture, Media and Sport Committee: Evidence
8.6 The BBFC’s own self-regulatory solutions fulfil these criteria and have widespread industry and public
support. We would encourage Government to continue to encourage and support these, and other, solutions.
However, as is clear from our evidence, there are some obvious areas where more robust Government
intervention is required to ensure proper protection of children online, for example online pornography, where
currently lack of industry self-regulation is leading to children being able to access inappropriate and potentially
harmful content. Equally, industry could do more, for example through effective moderation, to ensure that
their service is safe for children and the vulnerable online.
September 2013
Written evidence submitted by Arqiva
Summary of Key Points and Recommendations
— Arqiva welcomes the opportunity to respond to the Culture, Media & Sport Select Committee’s
timely new Inquiry into Online Safety.
— Arqiva believes that children and young people should be able to enjoy the benefits of digital
technology and the internet, whilst having the right to be safe online. Indeed, the commitment to
ensure child safety online is a matter of grave importance—and should be a shared responsibility
between parents, businesses and the wider communications industry including ISPs, web publishers,
WiFi providers and government. Therefore, Arqiva is pleased to be a member of the Internet Watch
Foundation (IWF) and to have fully engaged with the UK Council for Child Internet Safety
(UKCCIS), as well as government; key stakeholder organisations, businesses, retailers and
manufacturers on this important issue.
— Children gain access to the Internet from many devices: PCs, smartphones, tablets and games
consoles; both their own and those owned by their friends and parents. Many of these devices gain
access to the Internet via multiple networks: fixed broadband at home, mobile broadband and Public
WiFi outside of the house. In simple terms, there are two main ways of filtering the Internet content
presented to children:
I. The network can filter content, for all users, or for children only, where the network can
firmly identify the user and the age of that user.
II. The device can be set to filter content, with device passwords to prevent children from
changing these filter settings.
— Arqiva believes that both measures are required to minimise the adult content available to children.
The first would benefit from a consistent approach to content filtering for all Internet access networks
that provide service in public places. The second requires education for parents and children, as well
as a policy that all devices (and internet access software on those devices) should be supplied with
adult content filtering turned on by default.
—
Arqiva recommends that government place greater emphasis on internet safety themes across the
educational curriculum.
—
We believe that policy makers should do more to research and improve online protections for more
vulnerable children; define inappropriate content and improve the means for identifying it online. In
addition, more research should be conducted to understand the many different ways that children
and young people are using internet enabled devices, especially in the home and within social
environments.
—
We also believe that media and communications companies, including the public service
broadcasters, play a key role in educating the public.
About Arqiva
Arqiva is a media infrastructure and technology company operating at the heart of the broadcast and mobile
communications industry and at the forefront of network solutions and services in an increasingly digital world.
Arqiva provides much of the infrastructure behind television, radio and wireless communications in the UK
and has a growing presence in Ireland, mainland Europe and the USA.
Arqiva was responsible for UK Digital “Switch-Over”—engineering from analogue television to Freeview—
a huge logistical exercise which touched every Parliamentary constituency, requiring an investment by Arqiva
of some £630 million and was successfully delivered to time and budget.
Arqiva is also a founder member and shareholder of Freeview (Arqiva broadcasts all six Freeview
multiplexes and is the licensed operator of two of them) and was a key launch technology partner for Freesat.
Arqiva is also the licensed operator of the Digital One national commercial DAB digital radio multiplex.
Arqiva operates five international satellite teleports, over 70 other staffed locations, and thousands of shared
radio sites throughout the UK and Ireland including masts, towers and rooftops from under 30 to over 300
metres tall.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w19
In addition for broadcasters, media companies and corporate enterprises Arqiva provides end-to-end
capability ranging from:
— satellite newsgathering (30 international broadcast trucks);
— 10 TV studios (co-located with post-production suites);
— spectrum for Programme-Making & Special Events (PMSE)25;
— playout (capacity to play out over 70 channels including HD); to
— satellite distribution (over 1200 services delivered); and
— Connect TV—who launched the first live streaming channel on Freeview.
Arqiva’s WiFi network includes almost every UK airport—and reaches cross the hospitality and leisure
sector, providing WiFi to 85,000 rooms in leading hotel chains, and many restaurants, retail chains and shopping
centres and local high streets.
Elsewhere in the communications sector, the company supports cellular, wireless broadband, video, voice
and data solutions for the mobile phone, public safety, public sector, public space and transport markets.
Arqiva’s major customers include the BBC, ITV, Channel 4, Five, BSkyB, Classic FM, the four UK mobile
operators, Metropolitan Police and the RNLI.
The Questions
Key, Overarching Issues:
— How best to protect minors from accessing adult content;
— Filtering out extremist material, including images of child abuse and material intended to
promote terrorism or other acts of violence;
— Preventing abusive or threatening comments on social media.
1. Arqiva welcomes the opportunity to respond to the CMS Select Committee’s timely new inquiry into
Online Safety.
2. Over the past decade the UK’s media, broadcasting, telecommunications and technology industries have
undergone unprecedented, profound and exciting change. As a result of digitisation, the traditional boundaries
between what were once separate universes: Content, Communications and Computing are eroding. Such
“convergence” has seen the blurring of boundaries between the media, telecoms and information technology
sectors.
3. Technology is becoming an integral part of both modern life and children’s lives. According to Ofcom:
90% of children live in a household with access to the internet through either a PC, laptop or netbook;26
whilst 65% of children use the internet “almost every day”.27 Young people are increasingly taking part in a
wide range of activities online, enabling them to discover and access a wide range of content, connect with
their friends and family, as well as offering the potential to create and distribute their own content. Young
people are understandably excited, stimulated and motivated by these developments. Indeed, recent research
suggests that primary age children are “highly engaged” with digital technology.28 It is essential for policy
makers and educators to continue to research and understand the many different ways that children and young
people are using internet enabled devices, especially in the home and within social environments.
4. However, as well as bringing exciting new benefits, new forms of communication and emerging platforms
bring new opportunities for misuse. Whereas, traditionally, the public service broadcasters were trusted by
parents to provide quality viewing and listening content; suitable for children and family/inter-generational
viewing and delivered in a linear fashion; and at a suitable time (ie: pre or post-9pm watershed); we have
witnessed how in recent years emerging commercial “new media” platforms and social networking, in a largely
unregulated space, have exposed young people to a range of new dangers—from the exposure to “adult
content”, to the phenomena of “cyberbullying”, “grooming”, “sexting”, “self-harm”, gaming, gambling, as well
as illegal downloading and file sharing, or online fraud.
5. Child safety (online or offline) is a hotly debated issue. As Dr Tanya Byron has noted “A focus on the
most terrible but least frequent risks can skew debate in a direction that sends out negative and fear-based
25
26
27
28
Such as the wireless cameras operated by the BBC and Sky News, and the radio microphones used in virtually all television
production and many West End shows.
Ofcom’s Communications Market Report 2013
This is split across; 38% of 5–7 year olds, 62% of 8–11 year olds, 87% of 12–15 year olds—Ofcom’s Children and Parents:
Media use and attitudes report (Oct’12)
UK Safety Internet Centre—Safer Internet Day—Have Your Say Report February 2013: 86% of 7–11s use some form of online
communication tool, such as social networks and virtual worlds (56%), chat functions in online gaming (38%) or using webcams
(28%). The internet is crucial for schoolwork, and 94% of 7–11s say they have little trouble finding information for their school
work on the internet. Secondary age children are prolific online communicators: 96% of young people age 11–19 use some form
of online communication tool, including services such as social networks (74%), emails (72%), instant messaging (68%),
webcams (52%), chat functions in online gaming (45%), chat rooms (17%) and blogs (14%). Young people are also contributing
to the production of online content: a quarter (24%) of 11–19s have created a website, 21% have created a game, 14% have
created an app and 12% have created a blog
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w20 Culture, Media and Sport Committee: Evidence
messages to children, young people and families.”29 In this context, it is interesting to note that Ofcom studies
suggest that parents are more concerned about the media content delivered via the television (23% very or
fairly concerned) than via a child’s access to the internet (17%).30
6. In his recent speech on Internet Safety the Prime Minister distinguished between two challenges: “the
criminal and the cultural”, namely, the proliferation and accessibility of child abuse images on the internet, and
the fact that many children are viewing online pornography and other damaging material at a very early age.31
We agree with the Prime Minister that “…these challenges are very distinct and very different. In one we’re
talking about illegal material, the other is legal material that is being viewed by those who are underage.” It is
critical that these debates are left separate, and that discussion on the latter remains proportionate and balanced.
7. Arqiva believes that children and young people should be able to enjoy the benefits of digital technology
and the internet, whilst having the right to be safe online. Indeed, the commitment to ensure child safety online
is a matter of grave importance—and should be a shared responsibility between parents, businesses and the
wider communications industry including ISPs, web publishers, WiFi providers and government. Therefore,
Arqiva is pleased to be a member of the Internet Watch Foundation (IWF) and to have fully engaged with the
UK Council for Child Internet Safety (UKCCIS), as well as government; key stakeholder organisations,
businesses, retailers and manufacturers on this important issue.
Protecting Children through filters
8. As one of the six main public WiFi providers (with BT, Nomad, Sky, Virgin Media and O2), we provide
internet access through WiFi, which is used by the public in a variety of means. Arqiva’s recent contracts
include the provision of WiFi to UK airports (including Heathrow; Stansted; Glasgow; Edinburgh; Manchester
and Southampton); several hotels and retail environments (including Premier Inns and Enterprise Pubs); as
well as several local authorities, including Manchester City Centre; and several London Boroughs.
9. As a member of the Internet Watch Foundation we are committed to blocking access to child sexual abuse
content and other illegal material. We are also committed to blocking terrorist content, as well as intolerance
and hate, and criminal activity. This applies across our network.
10. We believe that it is particularly important to have strong protection in public places particularly where
unsupervised children might reasonably be found on a regular basis. Therefore, in the absence of a specific
request from our business customers, we commit that our standard public WiFi offering will also include
filters to block pornography. This will help prevent children and young people from accessing inappropriate
pornographic material themselves, and should also limit accidental exposure to inappropriate material that
nearby adults might be looking at in public.
11. Children gain access to the Internet from many devices: PCs, smartphones, tablets and games consoles;
both their own and those owned by their friends and parents. Many of these devices gain access to the Internet
via multiple networks: fixed broadband at home, mobile broadband and Public WiFi outside of the house. In
simple terms, there are two main ways of filtering the Internet content presented to children:
I. The network can filter content, for all users, or for children only, where the network can firmly
identify the user and the age of that user.
II. The device can be set to filter content, with device passwords to prevent children from changing
these filter settings.
Arqiva believes that both measures are required to minimise the adult content available to children. The first
would benefit from a consistent approach to content filtering for all Internet access networks that provide
service in public places. The second requires education for parents and children, as well as a policy that all
devices (and internet access software on those devices) should be supplied with adult content filtering turned
on by default.
12. It is unfortunate that, on occasion, filters can also mistakenly “over-block” and prevent access to sites
which parents would be content to let their children browse. Arqiva recognises this can sometimes be frustrating
for consumers.
13. However, it is also important to guard against a false sense of security. Parents must be made aware that
turning filters on does not immediately make the Internet “safe”. Parents should be encouraged to talk to their
children about what they do online and offline.
14. Filters can only block domains or websites—they cannot filter content or behaviour within sites such as
social media platforms. Given that these platforms are a major driver of young people using the internet, it is
important for these platforms to be vigilant themselves at combating cyber-bulling or anti-social behaviour.
29
30
31
Do we have safer children in a digital world? A review of progress since the 2008 Byron Review—Professor Tanya Byron
March 2010
Ofcom’s Children and Parents: Media use and attitudes report (Oct’12)
Speech by Rt Hon David Cameron MP, 22 July 2013—delivered to NSPCC
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w21
Protecting Consumers through education
15. Young people’s use of the internet should be an incredibly rewarding experience—both within the school
gates, and during their extra curricula home and social life. Activities that harness new technologies can make
a valuable contribution to the wider school curriculum and to children’s learning. Arqiva recommends that
government place greater emphasis on internet safety themes across the educational curriculum. We believe
that policy makers should do more to research and improve online protections for more vulnerable children;
define inappropriate content and improve the means for identifying it online.
16. Arqiva are proud to work with government, UKCCIS and others to establish clear, simple benchmarks
and classifications for parental control solutions. We believe parents should be empowered with the skills and
knowledge to protect their children while going online at home. In particular, parental controls can help limit
the chances of children being exposed to inappropriate online content. Arqiva welcomes parental control
features that may be included in new technology such as personal computers, computer and video games,
computer software, mobile phones and digital television services. These controls can be used to limit access to
only age appropriate content, to set usage times and limits and to monitor activity.
17. There are many ways of accessing and downloading music, film, TV and video safely online and it is
important that children and young people understand how to download content legally. Copyright law applies
to downloading, sharing and streaming—just as in the world of physical CDs and DVDs. Improved “media
literacy” should make explicit that those who make music, film or TV content available to others on a filesharing network, download from an illegal site, or sell copies without the permission of those who own the
copyright, are effectively breaking the law and could face serious penalties. In addition, illegal file-sharing
programmes and websites pose greater risks to your computer or mobile phone than legitimate sites. Users
often unwittingly download viruses or spyware and can inadvertently share personal computer files and
information. Some files are purposely misnamed on file-sharing and peer-to-peer networks to trick people into
downloading them.
18. When it comes to online safety, we also believe that media and communications companies, including
the public service broadcasters, play a key role in educating the public. The BBC should be praised for regularly
commissioning innovative, multi-platform content which inspires and educates audiences. The BBC’s sixth
Public Purpose includes “Delivering to the public the benefit of emerging communications technologies and
services:” Indeed, the BBC’s content and services are perhaps the most powerful means to drive adoption as
well as to educate the public as to the risks of emerging technologies. The BBC is a trusted guide to the digital
world for the inexperienced or unsure, a safe place to be for the young, a reliable and accurate on-air and
online source for the information seeker, and a challenging and involving partner for the more advanced user.
Ultimately, “online safety” should be regarded as an important plank of the BBC’s commitment to support the
media literacy of all its audience.
The need for research
19. There is a growth in using alternative portable devices (including mobiles and portable media players)
to access online content in a variety of places and without parental supervision. Use of mobile phones continues
to grow among young people—around 90% of young people over the age of 11 own a mobile phone. There is
also a popular perception that online behaviour on shared devices in the family living room is “safer” than
behaviour on mobiles, personal tablets or portable devices. However, Arqiva notes that some experts suggest
there is little evidence on the links between using more portable devices and the online risks young people
face in using such devices.32 Therefore, it is essential for policy makers and educators to continue to research
and understand the many different ways that children and young people are using internet enabled devices,
especially in the home and within social environments. Similarly, Ofcom has estimated that 43% of children
have an active social network (Facebook, Myspace or Bebo) profile, and across social networking sites it is
estimated that children aged 8–11 years have 92 friends—and have not met 12% of them face-to-face, and
those aged 12–15 years have 286 friends—and have not met 25% of them face-to-face.33 More research is
needed to explore the links between where, with what, how often and with whom, children access the internet
and the likelihood of online risks.
September 2013
32
33
NFER Report “Children’s online risks and safety—a review of the available evidence” (2010)—commissioned by UKCCIS
Ofcom’s Children and Parents: Media use and attitudes report (Oct’12);
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w22 Culture, Media and Sport Committee: Evidence
Written evidence submitted by James Griffiths
1. I am writing as a concerned parent and individual in relation to the current issue being considered by the
Government relating to the safety of children from the corrupting content of adult material that can be viewed
on the internet. I have seven children and along with my wife have sort to bring the children up in a safe and
loving household while facing the everyday challenges of family life. This current threat is both disturbing and
very harmful to the fundamental basics of family life and to the moral fibre of a right foundation for our
young people.
2. We recognise that the internet has brought considerable benefits to both children and society in being used
as a tool for gathering information especially in helping them with their studies. Our older children didn’t have
the same wealth of information when they were in education as our three younger siblings have now (aged 15,
13 and 11) but whilst we have blocks and filters in place ourselves there are many many other children who
don’t have this protection which leaves them very venerable to unsuitable and corrupting content.
3. We believe that much more needs to be done by the Government, industry, parents and young people to
put in place measures to protect the youth of today and make the internet a safer place for children. It is a
basic necessity that our children are safe from cruelty and harm wherever they are and that includes when they
are on the internet.
4. When we talk about child safety online we believe two areas of greatest concern and areas that need to
be focused on are:
— Keeping children safe from harmful content online (eg grooming and sexting).
— Protecting young eyes from legal content that’s not appropriate for their age.
5. It is not a new issue facing the Government as many persons and organisations have been calling for
years for policy changes and investment of resources in both technology and policing to better protect children
and young people from online dangers
6. Along with many other concerned parents we are calling on the Government and especially this Committee
with its responsibilities in this particular matter to bring in legislation that would require default blocking of
all adult content to be a standard setting on all internet enabled devices sold in the UK. This would a way of
stopping children from seeing harmful adult content and give peace of mind to parents when the internet is
being used. There is also a need of the development and use of better age-verification software.
7. The constant attempt to water down this action being taken in the name of freedom of choice is a very
poor argument indeed and surely recent statistics from Ofcom revealing two key facts—“81% of children aged
14–16 have viewed ADULT material online” and “Only 46% of parents have filters in place on their home
internet”—only serve to illustrate the urgent need of action as to this very serious issue.
8. My wife joins with me in my grave concerns as to this matter and we would urge you as a committee to
take account of these concerns along with many other parents like ourselves in moving forward to implement
the necessary actions to give our very treasured possessions and the hope of the future in this country—our
young impressionable children—the protection and moral support they need to prosper.
September 2013
—
—
—
—
—
Written evidence submitted by Christopher G. H. Thomas
I wish to make a submission to the Committee as an individual, a parent and grandparent, a practising
Christian and as a user of the internet.
I understand the great advance that the internet has brought to young people in accessing the immense
wealth of information available to assist them in their studies and other areas of normal life.
However, it has unquestionably brought with it very great moral dangers and much material which
damages children and adolescents. Included with this is access to images and information which is
inappropriate for those in developing years. They are just not able to handle such things at that age.
I recognise that parents have a major responsibility to protect their children from contact with
damaging or corrupting matter on the internet, but would respectfully submit to the Committee that
the Government also has the responsibility to use its power to provide suitable protection.
Measures I would submit to the Committee as things the Government should do are:
1. Make it a legal requirement that all internet-accessible equipment and devices sold in the UK have a
default setting as standard that all adult material is blocked.
2. Software to verify the age of users should be improved, and the Government should promote its use
strongly.
3. The Government should use its authority and influence with manufacturers to ensure that they put into
practice such measures. They should not be diverted by suggestions that they are impinging on individual
freedom.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w23
THE PRESERVATION OF BRITISH YOUTH IS AT STAKE IN THIS MATTER.
September 2013
Written evidence submitted by John Edgar Cameron Lowe
ONLINE SAFETY OF CHILDREN
Thank you for this enquiry and allowing some of us to make a submission.
I write in a personal capacity as one who, with my wife, have constant contact with children through
the church.
Sex is a very powerful force in many people’s lives.
When in its proper setting of raising children it is a good and right thing.
When out of its setting it is like fire out of control and does damage.
I have personally been damaged by adult material although this was in an age before the advent of the
internet. I have felt the effects of this damage throughout the rest of my life.
Today the chances of children being damaged are enormous compared to years ago. I therefore fully agree
with David Cameron, Michael Gove and others, that every possible effort should be made to get government,
parents, internet providers, educators and charities to work together to provide all the protections necessary to
safeguard them from the many different types of sexual corruption available online.
It is surely a crime to shatter the innocence of children and young people.
There is a solemn piece on the last page of the Bible that speaks of those that love and make a lie.
Those that make a lie are those that set up and publish these damaging and vile things.
Those that love a lie are those that delve into and watch them.
As responsible citizens, as many as possible need to do everything they can to address this massive problem.
May you all be helped in your deliberations.
September 2013
Written evidence submitted by Norfolk Library and Information Service
Summary
— Internet access is possible in a range of environments—a one size filtering solution will not fit
them all.
— Most library services already have internet filtering mechanisms in place.
— National standard needed to determine what type of information is acceptable in age bands—schools,
parents, internet industry and filtering software providers will all need to sign up to it.
— Standards will need to apply to volunteer run libraries, who may not have the necessary expertise.
Evidence
1. There are
—
—
—
—
—
—
a number of possible environments where children can access information online:
Home.
School.
Library Computer or authenticated library wifi access.
Internet café.
Open access public wifi hotspot (including in libraries).
Authenticated access public wifi hotspot.
Any recommendations will need to reflect on the conditions that apply in each environment—not all
conclusions can or should necessarily be applicable in all environments.
2. Further comments will be restricted to internet access in libraries.
3. Library services should be able to identify who is logging onto a public computer or authenticating onto
their library wifi access. In Norfolk this is achieved through the use of a borrower umber and PIN combination.
This means that it should be possible to require library customers to agree to abide by an Acceptable Use
Policy, which among other things would include agreement not to access sites or material that is “is obscene,
racist, defamatory or otherwise unlawful, or that would cause harassment or gross offence to others”.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w24 Culture, Media and Sport Committee: Evidence
4. It is then possible to withdraw internet access privileges if this agreement is breached.
5. To prevent breaches, most library services subscribe to an internet filtering service, which is configured
to prevent access to sites falling into a range of categories. In Norfolk, this is the same product that is used for
filtering staff internet access, so any product must be capable of handling more than one filtering regime profile.
6. We use the same filtering product for our public wifi access in libraries as we do on static computers,
although we believe this is by no means a common feature in other library authorities.
7. Different filtering products categorise sites differently, so application of a national standard will be
difficult—filtering service providers regard their classification methodologies as commercially confidential.
8. Filtering is applied at the site level, not at the content level. It is therefore not currently possible to block
adult video content on YouTube without blocking all video content.
9. YouTube is the default hosting service for all Norfolk County Council public video content, so any idea
of a national block on providing access to YouTube in libraries would have significant impact on the delivery
and distribution of information about council services.
10. Norfolk Library and Information Services would like to see all websites that do or may host adult
material to be required to hold it in an easily identifiable section of their site, so filtering software rules can
be applied.
Norfolk Library and Information Service do not currently make any distinction about what a customer can
access, within the bounds of our acceptable use policy, based on age.
This is largely because there is no national standard about what is and isn’t acceptable within each age
range, or even what those age ranges should be.
Our current age categories for borrowing books are:
Under 5
5–11
12–15
16–18
Over 18
11. Norfolk Library and Information Service does not view it as the responsibility of a local library service
to determine what is acceptable at each age, nor do we have the resources to do so—that should be set
nationally and applied consistently across all filtering services so no local variation is possible.
12. Norfolk Library and Information Service does currently view it as the responsibility of the parent or
guardian to monitor and control the internet activity of their children while they are on library premises. Our
safeguarding policy states that no children under the age of 8 should be in a library unaccompanied.
13. Having said that, use of computers in libraries in Norfolk by children and young people is not that high,
according to our usage statistics. However, as more free public wifi is rolled out across Norfolk, we may find
the proportion of children accessing the internet across our wifi network (and hence our filtering regime) but
on their own devices increases.
14. Access to social media sites is one of the most popular uses of uses of public computers and is used by
customers to keep in touch with their friends and family across the world. Any recommendation to filter out
use of social media sites in public libraries would have a severe impact on the positive benefits of social
networking for library customers. For this reason we would be opposed to the principle of blocking internet
access to social media sites in libraries.
15. Management of a filtered internet service requires resources and a degree of technical expertise to provide
it. This is a skill set that will probably be unavailable to most groups who may wish to take over the running
of a local library service on a voluntary basis.
There should therefore be a commitment that the local authority will make sufficient funding available to
continue to provide a comprehensive and secure IT infrastructure in volunteer run libraries. This will also help
to support the Government’s Digital by Default policy.
September 2013
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w25
Written evidence submitted by Phil Alexander
Children nowadays have the ability to access large quantities of very useful information effectively replacing
the large and space-taking encyclopaedias of days passing.
What ruins the good that we appreciate is the low down, corrupting content that even adults are safer without.
As a parent of four children I would welcome a large padlock on all content that was unsuitable to children,
the key to which would also be locked away.
Please could the Government urgently developing a secure firewall and legislate a mandatory download that
blocked such content and the owners of computers that feel they must defile themselves, would have to apply
to a Government watchdog?
Why do less than half of our parents have blocks on home computers? If the parents aren’t bothered, pity
help the children.
I appeal to the pity there is in the House, to have the heart to get this cracked.
September 2013
Written evidence submitted by Ben Hardwick
I am delighted to hear of your committee’s laudable efforts to look into the subject of the damage that
can be done to the younger generation by uncontrolled internet access—thank you very much for launching
this enquiry.
The ordered, peaceful and generally prosperous outward-looking UK society with its opportunities,
comprehensively organised education, healthcare, aid, infrastructure and defence/security arrangements are
a priceless result of the daily work of Government—something that is rarely recognised or valued like it
should be.
These benefits also extend far beyond Britain’s shores and the UK is still a beacon of tolerance and crucial
lynchpin of global peace and security as well as being an example that is frequently followed by other
Governments and States.
Little wonder then that Government is undoubtedly an instrument of God—1st Epistle of Peter, Chapter 2,
V13; it is a “giver” and “provider” as well as a “controller”—again, the Bible states in the Epistle to the
Romans Chapter 13 “for he bears not the sword in vain” along with a good deal else.
Government’s first duty is the protection of its citizens—this is an area at which HMG excels whether in
defence and intelligence matters or law-enforcement that stabilises and enhances the condition and experience
of society and its citizens and the whole global community.
I seek by these few paragraphs to set out simply how absolutely right it is for you to persist with better
control measures on internet viewing and that you may be encouraged and set forward in doing what is clearly
your duty.
I am a father of four children from three to 10 years old and revel in watching their rapid development and
willingness to absorb knowledge at an apparently impossible rate. They are very impressionable and it is a
very great responsibility as parents that my wife and I have in protecting and nurturing the next generation
into responsible citizens.
Affectionate control and unselfishness is absolutely critical in this and it is really tragic to hear figures from
OFCOM that more than 75% of 14–16 year olds have accessed adult material.
This is a situation that will cause permanent mind-scarring and a lack of fidelity and stable relationships on
which true happiness is built in our society—to say nothing of the cost to Government and the taxpayer of
broken homes because of so-called freedom to look at this content when in one’s tender teens.
HMG should move forward fearlessly with comprehensive and efficient control measures of what is currently
a lawless “wild west” of media so as not to fail the potential of following generations.
It is a most revolting and immoral thought that there are actually publishers and persons making millions
out of degrading us and ruining the moral compass of the young in our society.
I believe that there should be default blocking of all adult material on internet enabled devices sold in the
UK as soon as possible.
I wish you all the best and courage in your deliberations on this vital subject.
September 2013
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w26 Culture, Media and Sport Committee: Evidence
Written evidence submitted by John R. Edwards
Summary
— The present position of unlimited access to online material is morally dangerous and requires control;
— Some form of censorship should be established to determine what should be classified as “Adult
Material”;
— Parameters should be established as to what classifications of material should be eliminated from
online availability;
— Grades should be established to determine what material is acceptable for access by young people;
— Age levels should be established to determine when access should be allowed to “Adult Material”;
— Controls should be instituted which bar young people from access to material deemed unsuitable
for them;
Submission
1. The Inquiry into this subject arises from the fact that, at the present time, there is no restriction placed on
the Internet in respect of juvenile access to material widely accepted as unsuitable for them. Furthermore, such
online material is not graded and in the eyes of many much of it is regarded as being in the nature of cruel,
salacious and/or obscene. Some would go so far as to say that at least some of it is potentially harmful even
to adult viewers.
2. Concern has been increasingly evident that action is required to ensure that some kind of control is
established that will at least protect young people in particular from access to material regarded and classified
as “Adult Material”.
3. As a lifelong committed and practicing Christian, and as a parent, grandparent and great-grandparent of
a total of just on 50 children, grandchildren and great-grandchildren, I have an intense interest in the moral
safety of young people and am concerned to witness to the Committee my views, with which many would, I
am sure, join, that action is urgently required to bring in a measure of protection for young people against the
existing access to the volume of unsuitable online material that is currently so freely available to them.
4. Children are the potential for the oncoming generation. We need to be thinking and planning now for the
wellbeing of that new generation. If we permit freedom of access to and do not control the supply to our
children of, morally damaging material, the next generation will, inevitably, be basically corrupt. If we are
planning and endeavouring to ensure a well educated, law-abiding, honest, principled and upright next
generation, we must act now to make as certain as we reasonably can that they are morally equipped to be
convinced of and to maintain an upright way of life and to suitably educate their next generation. If, on the
other hand, we allow the standard of life to drift whichever way it may go, the result in the oncoming generation
could be intolerable, bordering on anarchy. The fact is that people are only law-abiding because they are taught
so to be. If the reins of law become slack, the quality of life deteriorates to the point where people do what is
best in their own interests regardless of the effect it may have on others.
5. Consideration needs to be given to the ambitions and character that should characterize the oncoming
generation. Parental influence needs to be promoted and parental responsibility encouraged so that they teach
their children the basic principles of an acceptable way of life.
6. Sex education, whilst theoretically justifiable, inevitably stimulates many children to enquire into what,
in effect, is pornography. The basic thinking behind the conception of sex education is clearly the concern
that children need to understand the inherent dangers that ignorance involves. Whilst this train of thought is
understandable, the real responsibility as to sex education devolves on the parents and the fact that the
responsibility has been included in the state education system is an acceptance of and pointer to the fact that
parents have been neglecting their duty in this regard.
7. The onset of the IT age has resulted in the computer becoming an essential to life. Children become
addicted to the machine to the point where they may feel compelled to use it even to do simple calculations
that in times past would have been regarded as mental arithmetic. It has become an avenue for exploration and
inevitably children use it to satisfy curiosity in various avenues, many of which are very beneficial. As a tool
for subject research it can clearly be advantageous and valuable. Regrettably, however, the freedom of access
and the almost infinite range of material available currently enable children to attempt to satisfy their innate
curiosity by gaining access to material that could be morally damaging. This is clearly the concern of the
Inquiry.
8. The United Kingdom has established and maintained a rule of law based on democracy and the freedom
of speech. As political views have changed and the availability of information has increased and become more
widespread, it has become increasingly necessary to ensure that established media standards are maintained.
This has been generally accepted by the population and political parties have, in the main, worked together to
ensure the maintenance of the principles governing media standards.
9. In current law regarding media, there are clear standards concerning what may and what may not be
published in newspapers, journal and films. On the other hand, the development of the internet has resulted in
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w27
a virtually uncontrolled flow of material, quite an assortment of which would not be permitted in a newspaper
or journal. What is even worse is that there is no control on the access to such material, whether by adults or
children. It is indeed questionable whether such cruel, salacious and/or obscene material, which could not
legally be published in a newspaper, should be available at all, whether to adults or children, on the internet.
10. Children generally learn the benefits of access to the internet without necessarily being taught any form
of control, leaving it to their judgment as to the limits that they should apply to their access to such material.
Indeed, such is the availability of material, that, in the absence of control, it is incredibly simple, once one
learns the method of access, for them to gain entry to such sites and to spend as much time as they wish in
digesting the stuff.
11. Voluntary controls are available. Thoughtful parents, concerned with the upbringing of their children,
can install filters on their computers that limit the availability of access. We are told that 46% of parents have
taken advantage of such methods of control and they are to be commended for their action, but that leaves
54% that have not done so, and whose children are therefore vulnerable to the range of such unsuitable material
that is so simple to access. Indeed, we are told that 81% of children aged 14 to 16 years have viewed what
would be regarded as “Adult Material” online—a very worrying fact.
12. The current law in this country grades films for public showing according to the composition of the
audience, in an endeavour to protect young people. The concern behind this Submission would be as to whether
it would be possible to apply that principle to the internet. In presenting this Submission, the concern is to
underline that the present position is morally unacceptable, to determine what controls need to be established
in order to bring about essential control and to encourage the institution of the necessary legislation as quickly
as is reasonably possible. The present position is serious and needs urgent attention. The present Inquiry is
valuable and places a huge responsibility on the members of the Committee. The concern is to support the
Inquiry and those who are appointed by drawing attention to how the present position is regarded by persons
who have for a long time carried a serious concern as to the subject.
13. When such legislation is prepared for Parliamentary consideration and enquiry, it is inevitable that some
persons and organizations will complain that people’s freedom of expression is being limited. The present law
as to libel, slander and sedition already puts certain limitations on the freedom of expression and what is being
considered is not incompatible with that. The recommendation would be that the present legal framework
should be extended to cover and include such matters as corrupting views, not only sexually explicit words
and pictures, but also violence, cruelty and the promotion of terrorism.
14. I submit therefore:
(a) That the present position of the unlimited access of young people to online material is morally
dangerous and requires a reasonable level of control;
(b) That some form of censorship should be established that reviews material available online with
a view to determining what material should be classified as “Adult Material” (and therefore
limited to viewing by adults) and what should be freely available;
(c) That the consideration of (b) above should extend to the setting of parameters as to what may
be published on the internet and to establish the necessary legal framework to ensure that
certain types of explicit sexual material and of violent, cruel, terrorist-promoting and obscene
literature should be eliminated from the internet;
(d) That, as a result of (b), grade levels should be established that determine what is acceptable
and what is not acceptable for access by young people, in respect of such material;
(e) That age levels should be established, probably in line with film viewing levels, that will
determine at what age young people may access such material;
(f) That controls should be established which bar young people from online access to material
deemed dangerous for their consumption.
September 2013
Written evidence submitted by Chris Evershed
Our young people are the future of the nation. Without them there is no future. How careful we need then
to be in caring for and educating them so that a generation emerges that the country can be proud of.
The internet has brought considerable benefits to children, society and the business world. However it can,
and is, being used to corrupt society and the very children that it should benefit. It is a sorrowful statistic as
stated by Ofcom, that 81% of children aged 14 to 16 have viewed adult material online, whilst only 46% of
parents have filters in place on their home internet.
Pornographic images can be viewed on the internet by young persons, so warping their minds. This can only
lead to a rapid decline in moral standards, increased delinquency, dysfunctional families, and a proliferation of
adults with no moral structure or ability to discriminate between right and wrong, good and evil. Because they
see adults engaging in such actions it must be OK.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w28 Culture, Media and Sport Committee: Evidence
The government needs to act now and decisively, to address this extremely serious problem. Default blocking
of all adult content should be the standard setting on all equipment sold in the UK that is able to access the
internet, combined with robust age verification. Without this, children will continue to be damaged irretrievably.
There is also a tremendous increase in child pornography polluting the internet, and research by the Child
Exploitation and Online Protection Centre (CEOP) has alleged that more than half of those who view child
abuse images go on to commit abuse themselves. This is a proof of the vicious spiral of moral decline that
sets in when such activity is allowed. Internet Search Providers should be held responsible to track down child
pornography and remove such images from the internet.
It is well to remember that the decline and fall of the Roman Empire was not due to any lack in the military
might, by which their empire was formed, but by moral decline and degeneracy.
September 2013
Written evidence submitted by EE
Introduction
1. EE welcomes the opportunity to provide comments to the Culture, Media and Sports Committee inquiry
into Online Safety.
2. We treat this issue with the utmost seriousness and we recognise that, as the ways in which we all get
connected to the internet develop, we have a responsibility to ensure a safe experience for our customers and
for their children. This is a challenge which will continue indefinitely as online safety and the control of
harmful content requires unremitting effort by individual companies, as well as collective endeavour.
3. EE is a British company which provides service to 27 million customers on EE, T-Mobile and Orange
plans. In addition to voice and text, we host a variety of content services including music, games and video
clips which are available within our mobile portals (Orange World, Web n’ Walk and T-Mobile) as well as
providing general internet access. We also run a fixed ISP (EE Home) which currently has around 750,000
customers and a corporate WiFi service.
4. Turning to the particular issues raised by the Committee, this paper comments on the best approach to
protecting children from accessing inappropriate content, recognises that technical solutions are only part of
the solution and that education for children and parents is key and that both network and software filtering can
meet the requirements of family friendly and effective filtering. It then details EE’s own policies on e-safety
and addresses the issue of blocking other harmful content.
How Best to Protect Minors from Accessing Adult Content
5. EE strongly believes that the best way to protect children from accessing inappropriate content is by a
combination of parental controls and education. Blocking websites maybe sufficient protection for a young
child but will not deal with the needs of young teenagers who will use social media and other personal
communications. Teenagers in particular will always try and find material that their parents or teachers would
rather they did not see, and it’s the industry’s role to provide educational material, as well as the tools, to
ensure children and parents are well equipped to make informed choices and do not inadvertently come across
material that is not appropriate for them.
Parental Controls
6. EE supports the recommendation that came out of the Byron review of “active choice” or “unavoidable
choice” where parents are forced to make a decision on whether they wish to apply a filter, as opposed to an
across the board “default on”. The principles of “unavoidable choice” as set out in the Government’s white
paper “Connectivity, Content and Consumers” include the availability of family friendly filters that block adult
content for both new and existing customers, the need for customers to make a choice about whether they want
filtering at first connection to the internet or at point of sale, age verification of the person setting up or
removing the filter and communication of the benefits of filtering. These principles can all be met through
either network based filtering or software based filtering on devices.
7. There is no “one” right answer for all customers, across all platforms or all devices. EE supports an
approach that allows for a portfolio of network based and device based solutions to meet the needs of the tens
of millions of people in the UK that access the Internet, through PCs, mobile devices, games consoles and,
increasingly, TVs.
8. There are very good reasons why mobile and fixed operators have chosen to take different approaches to
the delivery of parental controls for mobile and home internet. On mobile phones operators filter adult content
on the network and it is on by default for all customers and can only be lifted on proof of age. On the fixed
home internet many ISPs offer all customers free parental controls and provide unavoidable choice on set up.
This difference of approach is because the way people purchase mobile and home internet differs:
— all home internet customers sign a contract and are adults;
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w29
—
—
the home is likely to contain a number of PCs for each family member, each requiring different
levels of filtering, and some no filtering at all; and
households attempting to accommodate this mix of PCs via a single network filtering profile
are likely to experience dissatisfaction, and may well turn off filtering altogether. Allowing
different PCs to be protected via their own filtering software provides the necessary degree
of flexibility.
9. For mobile phones we filter content at a network level and it is set as on by default. This is because:
— Mobile phones unlike home computers are personal devices and so a single filtering profile
is effective.
— Filtering software is not available for all mobile devices. Different operating systems (Android,
Apple, RIM) each require a different solution which is why operators implemented a network
based approach.
— Mobiles are often bought by children who are unlikely to request at point of sale that parental
controls software should be activated.
— Many mobiles are sold through independent retailers where operators have little or no control
over how the devices are sold or who they are sold to. The purchasing decision for mobile is
very different for fixed with many indirect channels of distribution (such as supermarkets) who
are not set up to discuss the benefits of parental controls.
10. Given that the majority of children are highly technically proficient and routinely share tips and
information, it is considered that such controls would offer little or no practical protection. We continue
to evaluate options for offering greater controls on mobile devices and will revisit this if viable solutions
are identified.
11. As we have explained there are good reasons for taking different approaches for filtering mobile and
fixed traffic. In addition, we believe mandating network filters for all fixed ISPs in order to have filters which
are on by default, will be a huge expense and take considerable time to implement. The initial investment of
providing network filters on the mobile network costs millions of pounds and there are also ongoing costs for
maintenance and for the categorisation and filtering suppliers.
12. More information on the parental controls available across EE’s mobile brands, fixed business and Wifi
network is set out in Appendix I.
Education and Advice
13. Parental education and engagement is vital. As Dr Tanya Byron set out in her review on child safety,
parents have a responsibility to understand what their children are viewing and to educate themselves and their
children on how to assess the appropriateness of content.
14. Education is so important particularly for teenagers and older children who want more choice in what
they are able to access. Since access controls were put in place, content on the internet has continued to develop
and more players are involved in its delivery. This means that even with the best filters available it is
increasingly hard to block inappropriate content particularly on social networking, interactive sites and app
stores. Raising awareness of e-safety through diverse channels therefore remains of paramount importance.
15. Our research shows that parents often feel overwhelmed and unable to keep up with the pace of change
and consequently what their children are doing online.
16. Parents need to be made aware of the risks and given clear, simple signposting to sources of information
and tools to help them manage their children’s online life. This would include clear messages about what the
risks are and tools such as help with setting up parental controls.
17. We also advocate a “talk & trust” approach whereby parents are given guidance on how to open up
dialogue and have conversations with their kids so that they understand the risks. We don’t believe that blocking
and banning on every level is a sustainable solution as children will always rebel and find new ways to express
themselves. It is far more important to arm them with confidence and trust and provide them with information
to help them manage the risks as well as their parents.
18. As part of our rebranding, we recently updated our online safety information and launched EE’s Digital
Living website with dedicated pages on keeping families safe online. http://explore.ee.co.uk/digital-living/
keeping-children-safe
19. This includes a series of six short films together with activity guides, so parents can talk about e-safety
issues with their children. The films cover a variety of issues including cyberbullying, digital footprint and
posting information online and are EE branded versions of films provided in our previous award winning
Orange education programme which had reached 67% secondary school penetration by the end of 2012.
20. Our network of employee volunteers (e-Safety Champions) reached over 10,000 young people directly
with e-safety messages from 2009–12 as part of an employee engagement programme.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w30 Culture, Media and Sport Committee: Evidence
21. We continue to keep our material up to date and have recently updated our advice for parents leaflet
which provides guidance on a range of issues including setting parental controls, cyberbullying, sexting, privacy
settings, social networking, downloading illegal copyright, how to report abuse and controlling costs. The
leaflet “Advice for parents” is now in all 600 plus EE stores nationwide and is also available from our Digital
Living website.
22. By the end of the year all our frontline employees in retail and customer services will receive online
training on E-Safety and Active Choice highlighting the tools and materials on offer across the brands. This
module will also be included for new employees within their induction training.
Promoting Online Safety in Schools
23. In EE’s opinion more emphasis is needed in PSHE as part of the curriculum and specific attention given
to e-safety issues. We feel that as a network operator we have a duty of care to provide information and
guidance on how to use our technology in a safe and responsible way. We can share these messages with
existing and new customers and our free education programme which is used in 67% of secondary schools
shows that we go beyond this and out into the community. However the education system plays a vital role in
equipping young people with life skills. With less emphasis on PSHE in recent years we feel the issues are not
getting the traction they need. It should not be solely up to the network providers to fill this gap but rather for
the education system to embrace the materials and resources that our organisations can offer and build them
into daily school life. EE will continue to develop new materials on e-safety in the future for our website, but
the real key to success is the way these, and other organisations’ resources can be slotted into a relevant and
robust curriculum, so that teachers can take advantage of them.
Consistent Application of Active Choice
24. ISPs no longer control all aspects in the value chain. We firmly believe that content providers and
increasingly device manufacturers have a significant part to play in child safety if controls are to remain
effective. Government should also be ensuring that the responsibility for child protection applies across the
wider value chain with manufacturers, software providers, WiFi providers, search providers and social
networks, as well as ISPs. Not only will this provide more channels to market for child safety messages and
products but it will also ensure there is a level playing field in terms of implementation costs and responsibility.
25. Furthermore WiFi providers and manufacturers of consumer electronics products such as games consoles,
smart TVs and set top boxes, which are offering an ever richer and open online experience, should also be
required to offer active choice.
Filtering Out Extremist Material
Tackling Child Abuse Images and the IWF
26. EE has been an active member of the IWF for nine years and is one of the companies that contributes
the maximum amount to the IWF. We fully support the recent extension to the IWF’s remit to allow proactive
reporting by the IWF (rather than only investigating reports made through the hotline). This will lead to a more
effective use of the IWF’s resources. Any subsequent proposals put forward by the IWF should always be
supported by detailed plans and agreed through the normal budgeting process.
27. EE has never received a notice to take down content for any of our hosted services. We have strict
guidelines on the type of content hosted on our mobile portals and our interactive services such as chatrooms,
message boards and commenting facilities are all moderated.
28. We implement the IWF’s child abuse blocking list across all our platforms each day and return a splash
page to those customers looking for or stumbling upon illegal images. We also participate in the quarterly
testing programme to confirm that the block is operational.
Blocking other Forms of Content
29. There have been increasing calls by Parliament and lobbying groups for other types of illegal or
inappropriate content such as extreme pornography, racist or anorexia sites to be subject to the same treatment
as child abuse images with ISPs blocking offending websites for all its customers. For EE there are practical
and policy concerns with extending full blocking into other areas. The challenge comes in areas where there
is no consensus on the appropriate role of providers. Asking ISPs to block content that is offensive to some
people but is clearly legal, places ISPs as arbiters in deciding what customers can access.
30. We believe it is ultimately for Parliament, not ISPs, to take a view on unpalatable content and to
determine what is illegal or whether the legislative framework requires revision. The current process with
online copyright is that the courts through a court order decides whether a website is facilitating access to
illegal material and which IP addresses should be blocked and on what terms.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w31
Radicalisation/Terrorism
31. EE believes that sites that promote radicalisation and terror sites (as defined in the Terrorism Act 2006)
are already filtered by our mobile parental controls under the “criminal skills” or “hate” or “violence
categories”. However to be certain we require the Home Office to send across a list of the websites so we can
check how they are categorised by our filtering supplier.
Preventing Abusive or Threatening Comments on Social Media
Social Media
32. Social media and interactive services have revolutionised the way people connect with each other and
offers tremendous benefits. However EE recognises there are concerns over potential access to inappropriate
content, bullying and grooming. Due to the sheer volume of postings moderation of all content is impossible.
However providers could publicly commit to respond to all reports of potential child abuse, grooming and
bullying with an agreed time (such as 48 hours). Furthermore greater transparency of the safeguards available
on sites should be available as these vary greatly between services. We also believe that some moderation of
content that has potentially the most harm could be implemented.
33. All user generated content that EE hosts on our portals (T-Zones, Web n’ Walk and Orange World)
including chatrooms, video sharing, picture galleries, message boards and commenting services are moderated
by trained staff working to a documented and defined set of standards that define who is able to use the service
and what constitutes acceptable behaviour. There are slightly different policies in place for different services
and between the brands however all images are pre vetted against agreed guidelines before they are posted
online. In addition we have clear policies on dealing with complaints and on notice and take down.
34. Our message boards are moderated by EE Digital Staff and we also have a Site Feedback form where
people can contact the editorial staff to flag any issues. This is monitored between 9am -5pm every day of the
week. If there is an instance where we need to take down content that is already live we aim to do it in less
than 30 minutes. However the frequency of complaints is very low.
35. For article commenting we enable customers (or anyone who is viewing the website) to report abuse.
This is sent to a live dashboard which our moderation company monitor.
Cyberbullying
36. EE’s Customer Services provide advice to customers who are concerned about bullying by text or on
calls. If the problem persists customer services can offer to change the customer’s phone number and the most
serious incidents are referred to the Nuisance Call Bureaus where dedicated staff help customers who wish to
report cases of harassment or malicious calls to the police. The Nuisance call bureaus receive around 300
enquiries per month and of these around 50 will be referred onto the police and the remaining 250 will be
provided with appropriate advice and guidance.
APPENDIX I
EE’S PARENTAL CONTROLS
Parental Controls—Mobile
All three mobile brands (EE, Orange and T-Mobile) offer the highest level of protection:
—
All prepay and contract customers are offered parental controls free of charge and this is applied
by default for new customers across all mobile devices (handsets, dongles and tablets). Age
verification (via credit card, name and address check or instore) is required to lift the controls.
—
All networks now offer three parental control settings: Strict, Moderate and Off. We are the
only mobile networks to offer a Strict setting as we recognise that parents of younger children
may want to filter some content which is not classified as “18” but that they don’t wish their
children to view.
The three settings manage visual content (both commercial content and websites) as follows:
—
Moderate—this is the default setting which will be on for all customers. It allows customers
to view user generated content but blocks “18” content as defined by the BBFC (British Board
of Film Classification) based on the BBFC’s Classification Guidelines which are the result of
public consultations held every four to five years. This is wider than pornography and also
includes criminal skills, drugs, gore, hacking, hate, self harm, suicide and violence. For more
information see http://www.bbfc.co.uk/what-classification/mobile-content
—
Strict—on request customers can select this setting which blocks all “18” content as above
plus content that is deemed unacceptable for younger children including dating, cyberbullying,
unmoderated user generated content, chat, glamour, occult, weapons and cults.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w32 Culture, Media and Sport Committee: Evidence
—
Off—on proof of age (such as credit card or name and address check) customers can switch
off filtering which gives access to the open internet. The Internet Watch Foundation list is still
filtered for all customers.
EE has invested, and will continue to spend, millions of pounds in setting up and operating our parental
control systems offering the best solutions that are available to the industry. The current filtering arrangements
have been in place for nine years and work well with very few complaints about children viewing 18 classified
content on their mobile devices. Customers who believe content or websites have been misclassified can email
[email protected]. This email box is forwarded to the Safeguard Product Manager and issues are dealt with
within 48 hours.
Our Digital Living website provides detailed information on our parental controls, the full list of the
categories blocked for each setting and how to report abuse or misclassified content http://explore.ee.co.uk/
digital-living/keeping-children-safe
Parental Controls—Fixed
EE Home (our fixed business) now offers all new and existing broadband customers McAfee Family
Protection software which is free of charge for 12 months. Customers can install the software on up to five
devices are protected even when they’re outside the home as the software sits on the computer.
It offers five predefined age settings (under 5, 6–8, 9–12, 13–15 and 16–18) and parents can also customise
the settings, add or delete websites and set time restrictions. It also provides activity reports detailing web
activity and it will send email or text alerts if access to inappropriate websites is attempted. It will also report
on what information is being shared on social media sites and by instant messaging and blocks the downloading
of music with explicit lyrics. The software is easy to install and we have developed a help video with McAfee
which is available on our safety website.
We actively promote parental controls at point of sale and at all points in the customer’s journey including
in the Welcome Pack, in the router guide, in the customer’s first email after activation, in the Home Mover
Pack and Online on our Digital Living Website. Twice a year, as part of the EE Broadband Newsletter, we
include information on the Family Protection Software to existing customers. We will also send a dedicated
email to all customers every six months highlighting the benefits of online security and parental controls.
By the end of the year all new broadband customers will be presented with a “splash page” when they open
their browser for the first time which will force them to make a choice on whether to download the parental
controls software. If they choose to opt for parental controls they just click on the link provided which will
take them straight to the McAfee download page.
Parental Controls—Wifi
EE offers both branded and white labelled Wi-Fi services to a range of corporate customers. We are engaged
in the UKCCIS (UK Child Safety Council) working group and we are committed to offering a Family Friendly
Wi-Fi service. All our hotspots currently block illegal child abuse images on the IWF list. By the end of the
year we plan to set Adult Content filtering as “on” by default. It is then up to the venue owners to actively
choose to disapply the filter if they want to.
September 2013
Written evidence submitted by BCS, The Chartered Institute for IT
The Institute promotes wider social and economic progress through the advancement of information
technology science and practice. We bring together industry, academics, practitioners and government to share
knowledge, promote new thinking, inform the design of new curricula, shape public policy and inform the
public.
As the professional membership and accreditation body for IT, we serve over 70,000 members including
practitioners, businesses, academics and students, in the UK and internationally. We deliver a range of
professional development tools for practitioners and employees.
A leading IT qualification body, we offer a range of widely recognised professional and end-user
qualifications.
www.bcs.org
Headline Points
— The public, both collectively and individually, hold conflicting views on the balance between privacy
and protection;
— It is impossible to prevent knowledgeable determined people accessing material, however “illegal”,
if that material is anywhere on the Internet;
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w33
—
—
—
—
However, it is possible to make it less likely for naïve users to stumble across material that some
agency (which? and how?) has deemed to be offensive;
The question of what is offensive is one that calls for human judgement, is often context-dependent
and that human judgement may be highly controversial;
Individual cases prompt knee-jerk reactions, but “hard cases make bad law”;
Preventing someone accessing something that they regard as desirable is likely to encourage them to
adopt evasion technologies, which nullify *all* filters, not just those for material thought undesirable.
Background
1. In order to access a document/film/video across the internet one has first to know where it is (discovery)
and then have its data on one’s device (delivery). Delivery is the business of Internet Service Providers (ISPs),
while discovery is generally the business of search engines, but also of catalogues, such as Facebook links,
recommendations, word of mouth and rumour in school playgrounds.
2. The Internet Service Providers are doing business in this country, and know who their customers are (in
the sense of who is paying their bills). They are subject to various forms of UK regulation and legislation. The
search engines may or may not be based in the UK. They may or may not have any office or business in the
country and may not be subject to legal or moral pressure by the UK authorities. The Prime Minister’s speech
was somewhat confused when he said; “the search engine shouldn’t be involved in finding out where these
images are because the search engines are just the pipe that delivers the images, and that holding them
responsible would be a bit like holding the Post Office responsible for sending illegal objects in anonymous
packages”—the ISPs are the analogue of the Post Office, not the search engines.
3. It is important to understand that all documents (such as films and videos) are delivered across the internet
as if they were live broadcasts: no intermediary holds the entire document for analysis. A good analogy is that
of a service reading books aloud. There are then two fundamentally different approaches to censorship (which
is the main issue).
4. One is black-listing: analogous to “I’m not going to read you that book because it’s called ‘Lady
Chatterley’s Lover’, and I’ve been told not to read that book” or “I’m not going to show you that film because
it’s classified 18”. The major problem with this approach is that a vanishingly small proportion of the internet
has been examined for banning/classification.
5. The other is content-based filtering: analogous to “I’m going to stop reading that book now because I’ve
come across this banned word in it”, or “I’m going to stop showing this film because the last frame was more
than 40% ‘pink’ (sexualised imagery)”. There are two problems with this approach. The first is that, under the
Regulation of Investigatory Powers Act, it is probably illegal for ISPs to do. The second is that of false
positives: many books contain occasional “banned” words, and a frame may well be “pink” because of a
sunset, or a swimming gala.
6. The same objections apply to hybrid approaches, such as “I’m going to stop reading that book now
because I’ve come across a mention of a gamekeeper called Mellors”.
7. These difficulties should not be minimised. It would be no defence under the Regulation of Investigatory
Powers Act for an ISP to argue that the consumer (typically the parent of the child actually using the
connection) has consented: both participants to the connection have to consent to the interception, and it is
hard to see how an automatic web-server can consent.
8. Equally, the Chinese had a significant research project as part of what is generally termed “the Great
Firewall of China” to search for “inappropriate” skin tones, but that project has apparently been abandoned.
9. A further problem is that what is “abhorrent”, or even “illegal images of children”, is context-sensitive.
It is easy to say “medical textbooks are a special case, and are not meant to be read by minors anyway”, but
the problem is far broader than that. In 1995 there was a flurry (eg http://www.independent.co.uk/news/juliasomerville-defends-innocent-family-photos-1538516.html) of stories about photograph developers reporting
parents over image of their children. The switch to digital photography has merely meant that such images do
not need developing. Many of them then find their way onto social media.
“How Best to Protect Minors …”
10. The Select Committee asks about “minors”. There is no known technology which will determine if a
computer, or other device, is being used by a minor. For a home broadband connection, it would be possible
for the purchaser to be asked whether there was likely to be a minor using the connection, but of course it is
possible for the purchaser to lie, or even be honestly mistaken, as when suddenly looking after grandchildren.
11. Greater availability of, and publicity about, “parental controls” (which in fact are not parental at all, but
the parent buying into some-one else’s controls), on the lines of that offered by UK Safer Internet Centre,34
would help. However, it is worth recalling two fundamental statements from their site: “filters can be a helpful
34
http://www.saferinternet.org.uk/advice-and-resources/parents-and-carers/parental-controls
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w34 Culture, Media and Sport Committee: Evidence
tool in reducing the chances of coming across something upsetting” and “remember that filtering is only part
of the solution”.
12. A more challenging problem is provided by public WiFi technologies, which are used, often without
accounts, or via accounts with no verification. Public convenience would seem to demand this light-touch
access. It would be technically possible to have parental controls, although BCS does not necessarily
recommend this would affect all users and could lead to a rapid spread of evasion technology.
13. Similar to parental controls on the connection, greater publicity about tools such as Google’s Safe
Search35 would help, but again it is worth noting a fundamental statement: “please note that no filter can
replace a watchful eye”.
“… From Accessing Adult Content”
14. One problem which complicates the issue is that there is no international agreement about what
constitutes “adult” content. Social norms vary widely and it is unlikely that there will be much consensus in
the near future.
15. Content which has been formally rated “adult” in the UK is not a major problem. That content typically
requires purchasing, for instance, via a credit card or PayPal account and the transaction will show up. A
greater problem is pirated copies of such material, which are therefore not formally classified. The worldwide
digital content industry is working hard to combat such piracy and this should be viewed as an international
problem based on the fundamentally international character of the internet. BCS therefore does not see that
UK centric action is likely to be helpful.
16. A particularly worrying development is the prevalence of truly home produced material by apparent
minors. In one four-week period, the Internet Watch Foundation36 (IWF) had 12,224 such images reported.
88% of these were on “parasite” (IWF terminology) websites, ie those that harvested such material from the
website to which it was originally uploaded.
Education
17. The Byron report made a powerful analogy: “At a public swimming pool we have gates, put up signs,
have lifeguards and shallow ends, but we also teach children how to swim”. To this one could well have added
“and we help parents to teach children to swim, and we teach parents to be lifeguards.”
18. The sort of education necessary here for children is not technology education, it is societal education.
For this reason BCS believes that it belongs in the general Personal, Learning and Thinking Skills (PLTS)
category, rather than in ICT- or Computing-specific classes. There is excellent advice at the Get Safe Online
website, and class templates such as https://www.isc2cares.org/safe-and-secure/are available.
19. The IWF’s comment on the home-produced material points again in this direction. “These findings
provide evidence to support the importance of the education work delivered by child protection agencies to
raise awareness of the permanence of information on the internet and the risks inherent to young people in
creating and distributing this type of content.”
20. A challenging question is what and how much education is appropriate for parents. Some advice and
help on “parental controls”, both on the configuration and on tools such as Google Safesearch37 and YouTube’s
SafetyMode,38 most of which have come along since parents first encountered the Internet, is also appropriate.
Similarly, parents need to become aware of the (excellent, and recommended in the Byron report) PEGI rating
system for games and the advice at http://www.askaboutgames.com/, another site which has emerged since
many parents learned about the internet and which recent learners will not necessarily come across. Schools
should probably be encouraged to facilitate and host such classes in the PTA context.
21. Such classes will certainly need to cover technology, but should probably be wider. Parenting guidance
is sometimes sought where parents would like support in terms of how to engage with their children on social
topics that children might explore on the internet. These social topics are all related to what we might class as
“growing up on the internet” and have security facets to them. Topics might include; management of friendships
mediated by the Internet, trolling and other forms of Internet mediated abuse, balancing internet mediated
activities with offline activities, identity and projection of self via the internet etc. Again, Get Safe Online has
good age-specific materials, but these need to be drawn to the attention of parents, and their attention refreshed
as the children grow older.
“Filtering Out Extremist Material, including Images of Child Abuse and Material Intended to
Promote Terrorism or other Acts of Violence”
22. This is already done to a significant extent in the area of child abuse (technically speaking, indecent
images of children, which are illegal to possess under the Protection of Children Act 1978) by use of
35
36
37
38
https://support.google.com/websearch/answer/510?hl=en
https://www.iwf.org.uk/
https://support.google.com/websearch/answer/510?hl=en
https://support.google.com/youtube/answer/174084?hl=en-GB
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w35
blacklisting technology. More support could be given to the people, largely volunteers, who do the initial
reporting, and to the blacklisting process, generally under the auspices of the IWF. It is worth commending the
extent to which ISPs and mobile operators already cooperate with the IWF to manage and apply these blacklists.
23. It should be noted that blacklisting is not free, and has both financial and non-financial costs:
(a) The ISPs need to install and operate substantially more powerful equipment to do filtering than
is needed to pass through requests unexamined;
(b) The ISPs, very largely, fund the IWF;
(c) There is a risk of error and “false positives”: one such prevented edits to the whole of
Wikipedia;
(d) It is difficult to get right: the Irish study of filtering in schools;39 showed that 50% of schools
reported that filtering occasionally blocked valid educational sites, with a further 20% reporting
that it regularly did so;
(e) Filtering encourages the use of bypasses.
24. There has been much resistance to the Internet Watch Foundation’s widening its remit to the other
material in the Select Committee’s question, and BCS does not believe that this is the way forward.
25. Some people say that more should be done, and imply, without saying so, that content-based filtering
should be used, so that more such material could be blocked. This would require a major change in society’s
attitude to censorship, as well as primary legislation to enact fundamental changes to the Regulation of
Investigatory Powers Act. BCS does not believe that this is either feasible or desirable.
Preventing Abusive or Threatening Comments on Social Media
26. There has been much publicity recently around “online bullying”. Again, the problem is largely societal
rather than technological. The recent publicity has forced some developers into adding “report abuse” buttons,
but that is of little use unless the victims have the courage to do so. Hence this really comes down to an
education question, see above.
27. Even in cases where there is not a “report abuse” button, it is important that social media service
providers provide clear guidance and support for victims of distressing communications. These should detail
methods for locating support and information on how to report the incident(s). Where possible and appropriate,
providers should maintain regular contact with support and criminal justice agencies.
28. It is vital however, that the distinction is understood between those communications which:
(a) Amount to threats of violence.
(b) Targets an individual, resulting in harassment or stalking within the meaning of the Protection
from Harassment Act 1997 (including two new criminal offences of stalking added as sections
2A and 4A to the Act by the Protection of Freedoms Act 2012).
(c) Amount to a breach of a court order.
(d) Those not covered by the provision above but may be considered grossly offensive, indecent,
obscene or false.
29. In particular, given the evidence of the significant impact of the first three categories (including recent
evidence on the impact of cyber stalking) we must ensure that such actions are not simply considered as grossly
offensive and are dealt with under the appropriate legislation.
30. These categories are discussed in the Guidelines on prosecuting cases involving communications sent
via social media from the Director of Public Prosecutions published on 20 June 2013.
31. Without the public having a clear understanding of the differences in these communications the problem
is unlikely to diminish. Digital Natives have embraced technology but unfortunately without appropriate
training and education they struggle to understand the social norms of internet communication and behaviour.
32. There is a clear issue around anonymity and perceived anonymity (as well as untraceability) in social
media.
(a) In cases where senders of malicious communications have anonymity and (practical)
untraceability there can be difficulty in bringing justice and technological and legal changes
may be needed.
(b) In cases where senders have a (mistaken) perception of anonymity or untraceability they may
display fewer inhibitions and feel beyond reproach. It is important that all those who can assist
in making communications more traceable, particularly by giving up log information, do so
fully and readily when requested by those in the criminal justice system. The recent changes to
the Protection from Harassment Act 1997 do give police further powers and this is a welcome
change.
39
http://tnc2007.terena.org/programme/presentations/show4fa0.html
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w36 Culture, Media and Sport Committee: Evidence
(c) Where receivers of abusive messages perceive (rightly or wrongly) anonymity or untraceability
of the senders they may feel there is little point in reporting the communication, even when it
has significant impact.
33. It is important all stakeholders consider the vulnerability of the victim in cases of abusive or
threatening messages.
Conclusion
34. As a charity whose mission is “to enable the information society”, BCS looks forward to being a
significant player in society’s development of answers to these questions and welcomes continued dialogue.
September 2013
Written evidence submitted by Dudley Friend Clayson
I write as an independent & individual contributor of evidence to this Inquiry.
I am the retired Chief Executive of a company supplying UK industry with packaging machinery &
consumables. My wife is a retired doctor. Our evidence will cover the following issues:
— How our family has taken the initiative in safeguarding against the danger of our children
accessing pornographic material online.
— How members of our church have coped with this danger and cared for those who have been
mentally and morally damaged by it.
— Measures that Government should put in place to ensure that minors are seriously protected
from accessing pornographic material.
— Regulatory initiatives which OFCOM should take to regulate & police social media which are
currently responsible for a significant volume of material totally inappropriate for exposure
to minors.
— The internet industry must unite to develop technology to prevent the covert proliferation of
such material amongst minors.
1. How did we cope? My wife & I raised our family of four children without a computer in our home. As
practising Christians we could not, with a good conscience, expose our children to the possibility of coming
in contact with the poison of adult pornographic material, which is morally destructive in its consequences and
“utterly without redeeming social value” (1966 Supreme Court ruling in the obscenity case re. Fanny Hill,
defining pornographic material).
2. Our children are now married, and have their own families. Both they and we all now have computers in
our homes which are effectively filtered to eliminate any possibility of inadvertent contact with adult material
by anyone in the household.
3. Implicit in this Inquiry is the Government’s acknowledgement that the exposure of minors to adult material
is harmful. Such exposure entails grave consequences for the child, the family and for society at large. We are
now seeing the social norms & stabilising values which have underpinned our customary way of life in this
country being threatened by the emergence of a juvenile generation whose attitudes in respect of sexual
gratification, sexual crime, violence, promiscuity and marriage are being distorted & perverted by exposure to
online pornography. Sadly we have had some first-hand experience of damage to just a very few of the young
people in our church who have temporarily had contact with this depraved material. It has required ongoing
pastoral care & counselling to help them to get through the moral distress and mental disturbance resulting
from this exposure. Pornography leaves a scarring in the mind. We trust that the effects will be short-term; but
for some it could be a lifelong trauma.
4. It is heartening to see that Government is working with internet providers to put in place measures which
will effectively protect minors from easy & instant access to adult material. The Prime Minister himself has
given an undertaking that pornography will be automatically blocked on new broadband internet accounts,
leaving viewers with the option to remove the filter if they so decide. Mr Cameron has also made a commitment
in respect of existing users: the principal internet providers will have an obligation to contact customers with
an “unavoidable decision about whether or not to install family friendly content filters”. He has indicated this
will happen by end-2014. OFCOM will be responsible for implementation of these measures. The Prime
Minister has promised to take further action if, following these steps, there is evidence that children are still
not being effectively protected. We welcome his robust stance on this issue, which will mean that the many
parents who would like to be prompted or reminded will get that reminder and will be shown very clearly how
to put on family friendly filters. This is a positive move and a valuable commitment at the highest level
in Government.
5. OFCOM should prioritise regulation of social media, such as Facebook, and chatrooms, where the content
is not infrequently pornographic. OFCOM must ensure that the regulation & monitoring of these sites are
significantly upgraded, with mandatory training & compliance regimes put in place. Social media should be
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w37
placed under specific obligation to delete offensive material swiftly whenever a viewer complaint is received.
This is in essence a Health & Safety legislation issue and every other industry in which I have been active has
been subject to the strictest compliance with HSE. The internet should be no exception.
6. OFCOM should recognise that social media are exploited by paedophiles, prostitutes and sex traffickers
and should create an appropriately strong regulatory framework to block these criminal and anti-social abuses
of the internet. OFCOM should also require social networking sites to implement effective & inescapable age
verification checks to prevent minors accessing inappropriate material.
7. The internet industry should devote resources to develop technology to prevent the covert and currently
uncontrollable circulation of adult material within minors’ peer groups (such as school friend communities).
In conclusion, we are grateful for this Inquiry. The current situation is dire, with a reported 81% of 14–16
year olds having already accessed pornographic material online. We believe that < 50% of UK families
currently have family friendly content filters in their home computers. The Government’s commitments in
response to this situation are encouraging. We and a wide circle of our friends in this country will be
scrutinising progress and developments closely. We are very deeply concerned that the situation is rapidly
brought under effective, ongoing control—for the sake of every young person and child in this country.
September 2013
Written evidence submitted by PAPYRUS Prevention of Young Suicide
Summary
PAPYRUS has campaigned since 2004 for better regulation of the online promotion of suicide.
Amongst the many dangers posed by the internet, its use to promote suicide receives relatively little attention.
Young people, for whom the internet is a primary source of information, are particularly vulnerable to
websites and chatrooms promoting suicide. For those young people who are suicidal or are vulnerable because
of a mental health concern, such platforms can pose a particular danger.
The Government has made clear that the law which prevents encouraging or assisting suicide applies online
as much as offline, but no prosecution for online assistance has been made in this country.
ISPs should block websites and chatrooms which deliberately promote suicide. In the absence of this parents
should have to “opt in” to suicide sites in the same way as is proposed for other sites posing a danger to
children and young people.
Bodies such as IWF and CEOP should include suicide within their remit—at present both exclude it.
Detail
1. PAPYRUS Prevention of Young Suicide is a national charity dedicated to the prevention of young suicide.
Originally started by parents who had all lost a son or daughter to suicide it is now open to all who share our
aims, although many of its members have personal experience of a young person’s suicide.
2. Suicide is one of the biggest causes of death amongst young people; every year in the UK between 600
and 800 young people between the ages of 15 and 24 take their own lives—a number equivalent to the
population of a small secondary school; under the age of 35, the number rises further. According to the ONS
statistics published on suicides in 2011,40 the total number of those who took their own lives before they
reached 35 years of age was over 1700.
3. In the early years of the century, PAPYRUS became aware of several young people who had killed
themselves after downloading information from websites giving specific information and advice on how to
take one’s own life and/or having participated in chatrooms where advice and encouragement was given on
killing oneself. Since 2004 it has campaigned for the removal, or at least some regulation, of such sites.
4. Much of the public concern about the dangers of the internet has centred on its use to promote
pornography, bullying, eating disorders and self harm; these are important areas of concern both in their own
right and because they may play a significant part in leading vulnerable people to consider and attempt suicide.
But the suicide sites themselves are uniquely dangerous in that they can—and do—lead directly to deaths of
vulnerable young people. They also provide information which is not available from other sources. You will
not find a book in W H Smith’s describing in graphic and lurid detail the various ways of killing yourself; you
can find this (sometimes with pictures) on the internet. And of course this can be accessed by children and
vulnerable young people 24 hours a day, in the privacy of their own home or elsewhere from computers and
other e-devices. My own son took his own life aged 17, after downloading such information from the internet.
5. For many young people the internet has become the primary source of information; coupled with the
pressures of growing up, sometimes exacerbated by mental ill-health, there is a particular confluence of danger.
And although PAPYRUS’s particular concern is for young people many others who are vulnerable, again
40
http://www.ons.gov.uk/ons/rel/subnational-health4/suicides-in-the-united-kingdom/2011/stb-suicide-bulletin.html
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w38 Culture, Media and Sport Committee: Evidence
through mental illness or for other reasons, are also all too easily exposed to the dangers of the sites. When
PAPYRUS began campaigning it knew of four young people who had killed themselves with the aid of the
internet. Some time ago the number of internet-related suicides of which we were aware passed 50 and we are
sure that this is a considerable underestimate—no figures are collected centrally and this figure was obtained
merely through unofficial press monitoring. Many coroners have expressed concern about the impact of the
internet in cases of suicide.
6. The 1961 Suicide Act made it illegal to aid, abet, counsel or procure a suicide. The Coroners and Justice
Act 2009 amended the law by updating the language to refer simply to assisting and encouraging suicide,
whilst making it clear that the law applies to online actions in exactly the same way as it does offline.41
PAPYRUS contends that sites and chatrooms promoting suicide may be illegal in that they can—and do—
assist and encourage others to take their own lives; however no prosecution has taken place in this country.42
We believe that the Internet Service Providers (ISPs) should take action to block such sites in the same way
as they do for other illegal sites, such as child pornography.
7. In his speech on 22nd July the Prime Minister called on the ISPs to provide to all new customers family
friendly filters covering all devices in a household, which can only be switched off by the householder; and to
extend this in due course to all existing customers—in effect an “opt-in” approach despite the fact that this has
previously been eschewed by the Government. PAPYRUS warmly welcomes the Prime Minister’s intervention
and believes that the filters must automatically block dangerous sites and chatrooms which promote suicide.
(TalkTalk has told us that the category most blocked by parents using their existing filters is suicide/self harm.)
We have written to the Prime Minister seeking his assurance that the suicide sites will be encompassed within
these new arrangements. We trust that this will have the support of your committee also.
8. The internet can, of course, also be beneficial in countering suicide and in providing support to those who
are suicidal; PAPYRUS’s own helpline43 provides professional advice and information by telephone, email or
text messaging, to those concerned that a young person may be at risk of suicide and to young people who are
themselves having thoughts of suicide. It has sometimes been argued than any blocking of harmful suicide
sites must of necessity also block helpful sites; we have always believed that it is perfectly possible to retain
access to the helpful sites and were pleased to note that the Prime Minister acknowledged this in his speech
and has asked UKCCISS to lead work to ensure that this is so. We stand ready to assist in this work.
9. Although these recent advances have focussed on protecting children and the young (and we warmly
welcome them as such) we believe that the dubious legality of the websites and chatrooms provides sufficient
grounds for ISPs to block them for all users. There is also a moral justification for doing so; anyone seeking
access to them is likely to be vulnerable, either generically or at that particular time.
10. One of the proven ways of reducing suicides is to limit access to the means of killing oneself. This is
highlighted in the current National Suicide Prevention Strategy in Area for Action 344 and a recent study has
confirmed the effectiveness of limiting the numbers of paracetamol sold in individual packets.45 The National
Suicide Prevention Strategy for England states (para3.3) that “The internet is a ready source of detailed
information concerning the use of lethal suicide methods.” The young and vulnerable can and should be
protected from easy access both to information on the means of killing themselves and indeed to the means
themselves, through suicide paraphernalia and kits sold online.
11. One of the difficulties in pursuing the dangers of suicide promotion online is that there is no central
body to which relevant concerns can be expressed. Neither the industry body—the Internet Watch Foundation
(IWF)—nor the Government’s Child Exploitation and Online Protection Centre (CEOP) regards suicide as
falling within their remit. We have asked repeatedly for this to change and believe that real benefits would
ensue if this were to happen.
September 2013
Written evidence submitted by EU Kids Online
Preamble
We note with interest that the Culture, Media and Sport Committee has decided to investigate a number of
aspects of online safety that are currently raising concerns, in particular:
— How best to protect minors from accessing adult content.
— Filtering out extremist material, including images of child abuse and material intended to
promote terrorism or other acts of violence.
— Preventing abusive or threatening comments on social media.
41
42
43
44
45
Ministry of Justice Circular 2010–03
A former nurse has been convicted in Minnesota of assisting suicide online, including that of one person in England
HOPELineUK 0800 068 41 41
https://www.gov.uk/government/uploads/system/uploads/attachment_data/file/216928/Preventing-Suicide-in-England-A-crossgovernment-outcomes-strategy-to-save-lives.pdf
http://www.bmj.com/content/346/bmj.f403
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w39
This response is written by Professor Sonia Livingstone, on behalf of the EU Kids Online network.46 This
response may be made public. The address, for a copy of the Committee’s report, is S105, Department of
Media and Communications, London School of Economics and Political Science, Houghton Street, London
WC2A 2AE.
Much of EU Kids Online’s work has focused on scoping the nature and consequences of the online risks of
harm encountered by children aged 9–16 years old in Europe. The detailed recommendations below relate to
the European findings detailed on our website.47 For UK-specific findings, see Livingstone, S et al (2010).
Risks and safety for children on the internet: The UK report. At http://eprints.lse.ac.uk/33730/
Summary
The strength of a cross-national study is that it reveals how the UK measures against other European
countries. In many respects, the UK has been a leader in protecting children against online risk of harm.
However, EU Kids Online’s cross-national analysis also reveals that this protection has often come at the cost
of empowering children to benefit from the digital skills and opportunities of new online technologies.
Specifically, we found that European countries divide into four main groups, based on children’s risk profiles:
Supported risky explorers; Semi-supported risky gamers; Protected by restrictions and Unprotected networkers.
In Belgium, France, Germany, Greece, Ireland, Italy, Portugal, Spain, Turkey, and the UK—the “protected
by restrictions” countries, parents tend to overprotect their children, significantly reducing their online
opportunities. Researchers are concerned that both too much parental restriction and the lack of support for
children’s online use might lead to higher levels of harm when risk is encountered. Hence the EU Kids Online
network consistently emphasises that policy to encourage opportunities should accompany policy to reduce
risk of harm.
Main Points
1. Children have the right to protection and safety online but they must also take responsibility for keeping
safe and respecting the rights of others online
1.1 New means of internet access, less open to adult supervision, are increasingly evident in young people’s
internet use. Nearly half of all children in Europe go online in their own bedroom where it is unrealistic to
expect parents to monitor their safety.
1.2 Children need to be encouraged to develop self-governing behaviour in which they take greater
responsibility for their own safety in the use of the internet.
1.3 Awareness-raising should emphasise empowerment rather than restriction, and appropriate, responsible
behaviour with regard to technology use.
2. A new focus is needed on internet safety for younger users
2.1 It is important to balance protection of younger users with opportunities. It is important not just to
make the online world safe by stopping their use of internet services, but also to ensure their participation in
safe ways.
3. Safety messages should be adapted to new modes of access
3.1 33% of children now go online via a mobile phone or handheld device. Laptops, mobile phones, game
consoles and other mobile devices allow children to go online anywhere, anytime, away from parental
supervision. Emerging services (such as location-based ones) may lead to new risks.
4. Children’s online opportunities and skills need human and financial investment
4.1 Not only do younger children and girls not progress as far up the “ladder of opportunities” as teenagers
and boys, many never reach the final set of activities at all. Only half of 9–10 year olds progress further than
basic content-related activities.
4.2 Promoting more creative and skilled applications is essential to ensure all children avail of online
opportunities.
4.3 Schools play a pivotal role in digital skills development, mitigating forms of digital exclusion. However,
teachers are often inadequately resourced and trained to carry out the functions entrusted. Country differences in
online skills point to the need for targeted educational interventions where there is evidence of a digital divide.
4.4 Since opportunities and risks online go hand in hand, efforts to increase opportunities may also increase
risks, while efforts to reduce risks may restrict children’s opportunities. A careful balancing act, which
recognises children’s online experiences “in the round”, is vital.
46
47
http://www.lse.ac.uk/media@lse/research/EUKidsOnline/Home.aspx
http://www.lse.ac.uk/media@lse/research/EUKidsOnline/Home.aspx
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w40 Culture, Media and Sport Committee: Evidence
5. Positive online content for children should be made a policy priority
5.1 Provision of appropriate—high quality, diverse content online should be a priority.
5.2 The “European Award for Best Children’s Online Content” is a valuable step in this direction, but such
provision could also be supported by high profile national initiatives.
6. Digital safety skills are needed to build resilience online
6.1 Inequalities in digital skills persist—in terms of SES, age and, to a lesser degree, gender, so efforts to
overcome these are needed.
6.2 Younger age groups need to be a particular priority for parents and teachers. Secondary level schools to
date have been the main providers of ICT skills training but new interventions are required at the primary level.
6.3 Encouraging children to do more online will also improve their digital skills as well as their overall
confidence and/or increasing children’s beliefs in their abilities to use the internet. Similarly, teaching safety
skills is likely to improve other skills, while teaching instrumental and informational skills will also improve
safety skills.
6.4 Given uneven digital skills, particularly safety skills, across Europe and the discussion among
stakeholders about the need to identify more precisely the kinds of skills required, an inventory and agreed
framework for digital safety training would provide a valuable resource for educators, awareness-raising and
civil society groups.
6.5 Schools are uniquely placed to reach the maximum number of children. They are regarded by parents as
the most trusted source of information and, as the second most common location for going online, they provide
children with a very important point of access.
7. Social networking service providers need to ensure that maximum protection is provided for the accounts
of minors
7.1 If SNS age restrictions cannot be made effective, the de facto use of SNS by young children should be
addressed so as to ensure age-appropriate protection.
7.2 Privacy/safety settings and reporting mechanisms should be far more user-friendly. If they remain difficult
to use, privacy/safety settings should be enabled by default.
7.3 Digital skills to protect privacy and personal data should be strongly supported among children of
all ages.
7.4 It should also be noted that one in three parents (51% of parents of 9–12 year olds, 15% of parents of
13–16 year olds) did not wish their child to use SNS.
7.5 The review of data protection legislation at a European level needs to be considered from the point of
view of children’s privacy.
8. Awareness-raising in relation to online risks should be balanced and proportionate, and targeted at those
most at risk of harm
8.1 Children are concerned about a wide range of online risks. Efforts to manage these risks, and to support
children in coping with them, should maintain a broad and updated view of these risks.
8.2 As 9% of 9–10 year olds have been bothered or upset by something on the internet in the past year, it
is important to promote awareness-raising and other safety practices for ever younger children.
8.3 Awareness-raising among teenagers and their parents and teachers remains a priority since upsetting
experiences rise with age and the array of risks keeps changing.
9. Parental awareness of risks and safety online needs to be enhanced
9.1 Without being alarmist or sensationalist, parents need to be alerted to the nature of the risks their children
may encounter online. Awareness-raising should try to encourage dialogue and greater understanding between
parents and children about young people’s online activities.
9.2 Increasing parental understanding of risks is particularly important in those countries where awareness
of children’s risk experience is lowest.
10. Responses to young people’s exposure to online sexual content needs to be proportionate and should
focus on those most likely to be distressed or harmed by such exposure
10.1 Although public concern over online sexual content is justified, the extent of children’s exposure should
not be exaggerated, and nor should it be assumed that all children are upset or harmed by such exposure.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w41
10.2 Although the internet makes sexual content more readily available to all, with many children reporting
exposure via accidental pop-ups, the regulation of more established media (television, video, magazines, etc)
remains important.
10.3 Private access also matters—children who go online via their own laptop, mobile phone or, especially,
a handheld device are more likely to have seen sexual images and/or received sexual messages. Similarly,
those who go online in their bedroom, at a friend’s house or “out and about” are more likely to see sexual
content online. The early advice that parents should oversee children’s internet use must be revised, and new
safety tools are needed.
10.4 It seems that popular discourses centred on teenage boys’ deliberate exposure to sexual content makes
it harder for parents and others to recognise the distress that inadvertent exposure may cause girls, younger
children and those facing psychological difficulties in their lives.
11. Sensitive responses to bullying are required with equal attention to online and offline occurrence
11.1 In countries where there is more bullying overall, there tends to be more bullying online. This suggests
that as internet use increases, so will bullying online. Thus anti-bullying initiatives should accompany efforts
to promote internet use.
11.2 Online and offline bullying should be seen as connected, part of a vicious cycle in which perpetrators
reach their victims in diverse ways and victims find it hard to escape.
11.3 Those who bully may also be vulnerable, and they are often victims themselves, so sensitive treatment
is required.
11.4 Although children have a range of coping responses, cyberbullying upsets them, and more support and
awareness-raising is needed. Fewer than half tell a parent or other adult, and fewer than half know how to
block the person or delete their messages.
12. Parents need to be more aware of the practice of offline meetings with contacts first made online
12.1 It is important to distinguish making new contacts online—a common occurrence—from going to meet
new online contacts offline. It is equally important to recognise that for the most part, meeting online contacts
offline is harmless, probably even fun.
12.2 But for a minority of children, meeting online contacts offline is harmful, and these children tend
already to be the more vulnerable.
12.3 Since their parents are often unaware of what has happened, awareness-raising efforts should be
increased so that parents of younger and/or more vulnerable children recognise the risk, but without this
undermining the chance for most children to have fun making new friends online.
13. Policy makers need to be alert to new risks that affect children and young people, especially arising from
peer-to-peer contact
13.1 As well as conducting surveys, qualitative work based on listening to children is vital to learn what
new risks they are experiencing.
13.2 Addressing risks associated with peer-to-peer conduct (user-generated content and personal data misuse)
poses a critical challenge to policy makers.
13.3 While younger children have fewer resources to cope with online risk, they are also more willing to
turn to parents for help. Meanwhile, teenagers face particular risks that worry them and that they may struggle
with alone, so they need particular coping strategies and support.
14. Awareness-raising should highlight effective coping strategies in safety messages, emphasizing social
supports such as talking to parents, friends and teachers, as well as the use of online tools
14.1 Policy makers have long advised children to tell someone if they’ve been upset online, and it seems
such messages have been heard.
14.2 Children try some proactive strategies more than others and few are fatalistic. This suggests a desire to
cope as best they can and a readiness to adopt new technical tools if these are accessible.
14.3 When asked which strategies really helped the problem, children say that reporting the problem to an
ISP was effective with sexual images but less so for sexual or bullying messages: this suggests that better
solutions are needed for peer-to-peer risks.
14.4 Mostly, children said the approach they chose helped in up to two thirds of cases, but this leaves room
for provision of better support and/or tools.
14.5 It seems that efforts to promote children’s digital citizenship—in terms of online safety and good
practice—are bearing some fruit, and should be extended. There may be many reasons why solutions children
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w42 Culture, Media and Sport Committee: Evidence
try, when upset, do not help the situation, but one possibility is that the technical tools are flawed or difficult
to use, and another is that adults—professional or personal—are unprepared or unable to help children.
14.6 The “knowledge gap” phenomenon—in which the information-rich learn from available advice and
guidance more rapidly than the information-poor—means that efforts to promote digital citizenship will
disproportionately benefit the already-advantaged. Targeting less privileged or more vulnerable children is
a priority.
14.7 Overwhelmingly, children tell a friend, followed by a parent, when something online upsets them.
Rarely do they tell a teacher or any other adult in a position of responsibility. Their apparent lack of trust in
those who may have more expert solutions is a concern.
15. Practical mediation skills for parents should be a part of the overall effort to build awareness among
parents of risks and safety online
15.1 Parents appear to have got the message that it is valuable for them to engage with their child’s internet
use, and they employ a wide range of strategies, depending partly on the age of the child. But there are some
parents who do not do very much, even for young children, and there are some children who wish their parents
to take more interest. Targeting these parents with awareness-raising messages and resources is thus a priority.
15.2 Cynicism that what parents do is not valued, or that children will evade parental guidance, is
ungrounded: the evidence reveals a more positive picture in which children welcome parental interest and
mediating activities while parents express confidence in their children’s abilities. It is important to maintain
this situation as the internet becomes more complex and more embedded in everyday life.
15.3 Parental restrictions carry a significant cost in terms of children’s online opportunities and skills, but
they may be appropriate if children are vulnerable to harm. Parental efforts to empower children online seem
to enhance their opportunities and skills, though there is little evidence that they reduce risk or harm. Since
there are no easy answers, parents should be supported in judging what best suits their child.
16. Filtering technologies and parental control software need to be far more usable and transparent and take
into account the needs of parents in order to improve uptake
16.1 Across the 25 countries surveyed by EU Kids Online, less than one third (28%) of parents were found
to filter the websites visited by their child. It is clear that many parents find them such software either too
complicated or ill-suited to their needs.
To be effective, parental controls need to incorporate all of the issues that concern parents about their
children’s internet use. Thus, in addition to filtering out adult or unsuitable online content for children, controls
may also need to include features such as the amount of time spent online, filtering of user-generated content
and blocking of commercial content.
While there continues to be debate about the appropriateness of parental controls in all situations, they
continue to be a valuable resource particularly for those who may lack skills or knowledge in advising on and
guiding their children’s internet use.
Parental controls are also available as an integral element of some internet services and do not need to be
separately installed. An industry-wide agreement on the design and features of safety and parental controls built
into web-based services could provide parents with better opportunities to consider adopting them. Training in
the use of tools should also be made readily available to deal with lack of confidence and knowledge on the
part of parents.
17. Levels of teacher mediation are high but could be higher, as a large minority of children are not reached
by teacher guidance. Since schools have the resources to reach all children, they should take the biggest
share of the task of reaching the “hard to reach”
17.1 The youngest children (9–10 years) report the least mediation from teachers: as this age group now
uses the internet widely, primary schools should increase critical and safety guidance for pupils.
17.2 The benefits of supporting peer mediation are easily neglected but could be constructively harnessed,
especially as children are most likely to tell a friend if something bothers them online. Peer mentoring schemes
have a valuable role to play.
17.3 When something has bothered them on the internet, 36% of children said a parent helped them, 28% a
friend and 24% a teacher. The ideal may for children to have a range of people to turn to, depending on the
circumstances. A minority of children has no-one to tell when something upsets them.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w43
18. Industry needs to be much more proactive in promoting internet safety awareness and education. In order
to increase trust, the management of safety, identity and privacy settings of internet services used by children
needs to be transparent and independently evaluated
18.1 The overwhelming majority of parents would like to receive information and advice about internet
safety. Most, however, get it from firstly from family and friends (48%) rather than from the providers of
internet services. Traditional media (32%) and the child’s school (27%) are the next most common sources of
information about internet safety. Internet service providers (22%) and websites (21%) are much less evident
as sources of advice.
18.2 There is considerable scope, therefore, for industry to improve its own awareness raising and provision
of safety advice. Internet safety advice should be provided in an accessible and user-friendly way at the point
of access on web services used by young people. Internet service providers (ISPs) should play a more prominent
role in providing online safety resources for parents as the primary account holders.
18.3 Traditional media sources—press, radio and television—also have a major role to play in promoting
online safety awareness as supporting greater media literacy among the adult population. They are best
positioned to reach all adults and, crucially, are influential in forming parents’ attitudes towards opportunities
and risks on the internet.
18.4 Evidence repeatedly shows that children still struggle with user tools, safety devices, privacy settings
and policies, reporting mechanisms, etc, even though the industry claims they have been improved and made
easier. Independent evaluation of progress by the industry is crucial, to measure whether improvements have
been made (against benchmarks) but more importantly, whether those improvements work—ie are they actually
sufficient for children to manage their safety, privacy, identity and risk online?
September 2013
Written evidence submitted by Rolf Smith
Background to Submission
Firstly, I would like to thank the Government for giving me the opportunity to provide this submission to
the Committee.
I am 33 years old and am happily married to my wife; of whom we have two children aged seven and five.
I am writing therefore in my capacity as a father with some experience, as well as with feelings for all fathers
and mothers of children throughout the UK and across the globe. As a parent I am acutely aware of the many
potential threats to our children’s safety, of which the internet is an ever present and increasing one; hence the
reason for this submission.
Whilst I am not a professional or seasoned campaigner, I can honestly say that the following information is
given in good faith and is believed to be correct. Any data provided was taken from Ofcom reports.
Personal Experiences as Father of a Young Family
Computers and the internet are a fundamental part of life for our children, as they are for any children in
the UK today. In a way that we adults did not, they are growing up with computer technology all around them.
Indeed, it is astounding how quickly—almost naturally—they pick it up. My daughter has just turned five and
cannot yet read, but even she can navigate with remarkable speed through websites she knows to find the part
she wants, simply by memorizing the position of the links and buttons! It is abundantly clear and proven that
the internet appeals to a child’s inquisitive nature and their exploratory learning style; certainly that has been
our experience, and it won’t be long before they are better at it than us.
There is no doubt that the internet can be useful for children, particularly we have found it helpful for
improving parts of our older child’s education. But we are very thankful for the strong filters we have in place
which allows access to approved sites only. This gives my wife and I the opportunity to check the content of
sites carefully before allowing them to be visited by our children. Even so, our children are only allowed to
access the internet for timed, monitored periods and we certainly wouldn’t leave them alone during this time.
I know and they know—as we have taught them—that the internet has hidden dangers that they are not able
to cope with.
In addition, too much computer activity tends to make our children less interested in outdoor activities and
other necessary skills like writing and drawing.
The National Situation Regarding Online Safety
The Committee will no doubt be aware of the many statistics published by Ofcom on the subject of children
and safety online. Therefore it is not necessary for me to go into too much detail regarding the national
situation; however I must point out that it is very clear from data gathered that many parents in the UK, unlike
myself and my wife, are relatively unaware of how potent the hidden dangers of the internet are. While most
parents acknowledge that there should be some form of parental control, many do not view the potential
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w44 Culture, Media and Sport Committee: Evidence
dangers as particularly serious so are not pro-active in controlling their children’s use; therefore they allow
their children to spend long periods online alone without supervision and without controlling software. This
key fact could have a serious negative impact on UK society in the future if it is not properly addressed now.
This is of great concern to me. A few key Ofcom findings bear this out:
—
54% of parents in the UK do not use parental control software on their home internet.
—
46% of parents agree that their children know more about the internet than they do.
—
81% of children aged 14–16 have viewed adult material online.
—
Only 17% of parents say they are concerned about the internet.
—
25% of 8–11 year olds and 34% of 12–15 year olds are likely to communicate with people not
directly known to them via social networking sites, these figures rose greatly in 2012 report.
Why Online Safety for Children is an Issue which needs Action
It is undoubtedly a necessity that children are kept safe from harm wherever they are, and this must include
when they are on the internet. The proliferation of social networking, video sharing, personal smartphones etc
has made it even more vital that this issue is viewed seriously. The reality is that practically every UK child is
using the internet, and 80% of 12–15 year olds in the UK have an online profile. Therefore children must be
actively protected from the following (among others):
—
Access to explicit adult material.
—
Online grooming and other harmful communications.
—
Cyber bullying.
—
Online stranger danger.
—
Uncontrolled spending.
It seems from Ofcom data that the issue is twofold:
(a) While most parents know that the internet has dangers, many do not consider them to be serious
enough to pose a significant risk to their children. Therefore there is a lack of awareness.
(b) This has meant that parental control software, which is essential to maintain safety for children
on all online devices, is not used nearly as widely as it needs to be.
Suggestions of Actions that could be Taken to Improve Online Safety
As is the case for the care of children generally, the ultimate responsibility for ensuring children stay safe
online lies with their parents. Parents need to speak to their children about online safety, impose restrictions
and controls, and maintain supervision at all times. However, there are two things that I feel Government must
do fulfil their part in addressing the issue of Online Safety for children:
(1) Increase parental awareness—the dangers of the internet do often seem remote and
hypothetical, whereas they are in fact very real and present. Government can use its
considerable influence to help parents across the UK to realize this. Parents need to be helped
to understand that it is a basic responsibility to protect their children online, as much as it is
basic to teach them how to cross the road safely.
(2) Insist on parental control software—rather than leaving parents to investigate and install some
kind of control software, Government must insist that every online device sold in the UK for
personal use must come with parental control software pre-installed. Ofcom’s findings
absolutely support this suggestion. They found that the primary reasons given for parental
controls not being used were as follows:
(a) a lack of awareness or gaps in understanding;
(b) the perceived complexity of parental controls;
(c) the degree of effort that was expected to be required; and
(d) busy lives coupled with the absence of a specific trigger.
In addition to this, Ofcom reports that “when parents were asked whether they would install parental controls
if the process was made very easy, almost all parents said they would do so.”
I look forward to hearing the results of the Committee’s inquiry, and look forward to hearing about the
Government action which is sure to follow.
Many thanks again for the opportunity to provide this submission.
September 2013
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w45
Written evidence submitted by Mediawatch-UK
A. Executive Summary
1. We welcome recent government initiatives designed to protect minors from accessing adult content but
are of the opinion that three key areas must still be addressed in order to provide the best possible protection.
These are:
— Education.
— Age verification.
— Statutory backing.
2. We welcome discussion of how best to prevent abusive and threatening behaviour using social media.
This is a behavioural issue and unlikely to be effectively dealt with by filtering but by collaboration between
industry and other stakeholders. We believe there are three key areas in tackling this problem:
— Education.
— Regulation.
— Real sanctions, with statutory backing if necessary.
B. Mediawatch-UK
3. Mediawatch-UK is a voluntary organisation, established in 1965 by Mary Whitehouse CBE and formerly
known as the National Viewers’ and Listeners’ Association. Mediawatch-UK has thousands of subscribing
members throughout the UK. Our membership consists of people of all ages, occupations and backgrounds
who are concerned about the overall influence of the media on the individual, the family and wider society.
4. Our Director and other spokespersons appear regularly in the media (press and broadcast) on behalf of
our members. Further information about Mediawatch-UK can be found on our website:
http://www.mediawatchuk.org.uk/
C. How Best to Protect Minors from Accessing Adult Content
5. We believe that parents and businesses have a shared responsibility for children’s online safety. Parents/
guardians have the ultimate responsibility for their children but businesses and government have a responsibility
to provide parents with the tools they need to keep their children safe.
Education
6. Many of our members have reported difficulty with installing content filters and keeping up to date with
filters across the plethora of internet-enabled devices in their homes. Coupled with this many have reported
confusion in knowing where to go to access reliable information and education on internet safety.
7. This is best illustrated by a family which recently contacted us for help. The family fostered children
whose chaotic backgrounds placed them at particular risk. They were long time carers for a child who had
suffered a history of serious sexual abuse. They were aware of the potential risks online and had taken advice
from social services and their ISP. They had set up filters on the internet enabled devices in their home and
believed their mobile devices would be protected under The Mobile Broadband Group Code of Practice. Their
child was introduced to pornography on the phone of a child at school and she soon became addicted and was
regularly accessing pornography using the wi-fi at the homes of friends using her phone. Because of this child’s
history she is particularly vulnerable to sexual exploitation and misadventure.
8. At the time of the Parliamentary Enquiry into Online Child Protection in 2011 we asked our members (a
group likely to be more than averagely engaged with this issue) whether their ISP had ever contacted them
with details of the parental controls available as part of their package.
— 20% reported that their ISPs had informed them about parental controls.
— 80% said their ISPs had never done this.
— Of those who had been told about parental controls 61% were with TalkTalk. Approximately
60% of TalkTalk subscribers had been alerted to their provider’s new HomeSafe service.
— Of members who subscribed to other services only 11% had been told about the parental
controls available.
Of some concern were a number of BT customers who were under the impression that installing parental
controls would cost them a further subscription.
9. There is a great need for integrated and accurate advice for parents and carers on the potential dangers of
accessing inappropriate online content and how best to protect their children. There is much good work being
done in this area by a variety of organisations but further co-ordinated work is required to ensure that all
parents are made aware of and receive access to information and assistance.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w46 Culture, Media and Sport Committee: Evidence
10. We believe there should be space in sex and relationship education in schools to educate young people
about the potential dangers of accessing adult content online, how to avoid doing so and strategies to enable
them to understand what they may have seen.
Age verification
11. We welcome the Government’s efforts in this area but believe no mechanism (“active choice”, “defaulton” or “opt-in”), will work without “robust age verification”.
12. We commend efforts made in the internet gambling industry to all but eradicate underage online gambling
and we would like to see similar measures implemented with regard to adult material with appointed agencies,
possibly OFCOM or ATVOD, regulating.
Statutory backing
13. We also remain convinced that the best approach to ensure the Government’s pledges are actually
delivered on is by changing the law. The current “default-on” proposal is a voluntary agreement between the
major ISPs and, although four currently dominate the market, the Government’s plan does leave around 10%
of the market unaccounted for. If filters were legislated for then all ISPs and MPOs, regardless of size, would
have to conform which would offer a greater degree of protection.
14. Statutory backing would also remove anomalies between different providers and contracts. For example:
many consumers believe that, as a result of the Mobile Broadband Network’s code, child filters are activated
as a default for all UK mobile phone customers. This is not the case. Some providers require the filters to be
activated and provision varies dependent on whether a phone is pay-as-you-go or contract, provided by a
Mobile Phone Operator or a Mobile Phone Virtual Operator.
D. Preventing Abusive or Threatening Comments on Social Media
Education
15. There is a great need for integrated and accurate advice for users, parents and carers on the potential
pitfalls of social media use and how best to protect themselves and their children. There is much good work
being done in this area by a variety of organisations but further co-ordinated work is required to ensure that
all parents are made aware of and receive access to information.
16. We believe there should be space in PSHE education in schools to educate young people about safe and
responsible social media use including the consequences of anti-social behaviour and sources of advice and
support in case of abuse.
Regulation
17. The former Director of Mediawatch-UK, John Beyer, during the course of his work attracted much
criticism including an abusive Facebook paged entitled “I Hate John Beyer”. When this was brought to our
attention we reported it Facebook and asked for it to be removed. We reported the page three times but heard
nothing back from Facebook. Eventually, after approximately three months, the page was removed although
we were not formally informed of its removal but discovered it ourselves.
18. Although the issue was dealt with it took too long and there was not enough communication in the
process. Neither were we directed to sources of support and advice. These are issues which need to be addressed
in a code of practice for social media sites.
19. Social media is now a mature medium and it is our opinion that it should be regulated accordingly.
Recent events (such as the suicide of Hannah Smith following bulling on ask.fm) have demonstrated the
potential effect of online abuse. We believe that abusive and threatening behaviour on social media is now an
issue of health and safety and we would like to see it regulated accordingly via an industry code of practise.
20. We would like an independent regulator to ensure that sites are sufficiently moderated and that complaints
are dealt with swiftly, efficiently and within an agreed timeframe. Such a regulator could ensure the
development and implementation of an industry code of practice with the results promoted to parents and
children. This code should include:
— Clearer and simpler reporting mechanisms, especially where a service is marketed at and
provided to under 18s, making it easier for users to report abuse.
— Improved transparency and communication of protocols followed when reports of abuse are
made, including average response times, so that reporting users are able to know the timescale
for action with regards to when and if the problem will be dealt with and if and when the
content will be removed.
— Increased moderation of user-generated content. This moderation is especially important where
a service is proactively promoted and used by children.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w47
—
—
Prominent signposting to sources of expertise, advice, support and help for users affected by
threats or abuse.
A code of practice relating to anonymous use of sites.
21. We would also like to see service providers working more closely with the organisations dealing with
the consequences and aftermath of online threats and abuse taking place through their services, providing both
support and funding.
Real sanctions with statutory backing if necessary
22. Given the anonymity afforded online there can be a perception that abusive behaviour on social media
is without consequence. We welcome the clarification from the CPS on this issue but we would like to see
increased sanctions for abusive or threatening behaviour online which reflect those which exist offline and in
other forms of media. For maximum efficacy we believe these need to have statutory backing.
September 2013
Written evidence submitted by South West Grid for Learning (SWGfL)
Summary: With over 10 years of pioneering work in online safety and as part of the UK Safer Internet
Centre, we are fully aware of emerging technologies and potential risks, the issues children and young
people, professionals and parents face, and the gaps that exist in their support; as well as stakeholders
and partner organisations working in this field. In the response below we outline strategies we feel would
improve online safety in the UK:
— For point I. How best to protect minors from accessing adult content we recommend
parent/filtering tools together with wide education campaigns, better content rating
mechanisms and more research to identify gaps in support and provision.
— For point II. Filtering out CAI material we recommend providers should be members of
IWF and the increased use of alert tools.
— For point III. Preventing abusive comments on social media we recommend a combination
of prevention and empowerment strategies such as education and information about
reporting abuse, and clear and simple reporting mechanisms from industry.
1. Background to SWGfL:48 Our esafety team of experts has national and international reputation in
safeguarding children and young people online. Our expertise is around e-safety policy guidance and
improvements and training in schools to staff, parents and pupils. We produce award winning resources such
as 360 degree safe49 used by over 4,000 schools to assess their esafety provision, pinpoint their weaknesses
and make appropriate improvements. We also collaborated with OFSTED to inform the inspection of esafety
standards in schools and early years settings.
2. Role within UK Safer Internet Centre: As one of the three partners comprising the UK Safer Internet
Centre,50 SWGfL operates a unique helpline51 for professionals to help resolve online issues about themselves
or the young people they work with. The centre also coordinates Safer Internet Day every February in the UK.
To illustrate our work from the last 12 month here is an infographic:
48
49
50
51
http://www.swgfl.org.uk/Staying-Safe
http://www.360safe.org.uk/
http://www.saferinternet.org.uk/
http://www.saferinternet.org.uk/about/helpline
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w48 Culture, Media and Sport Committee: Evidence
I. How best to protect minors from accessing adult content
3. Adult content presents a range of harmful risks to children and young people and a number of strategies
are required to best protect them. In this context, we are considering legal adult content. Children and young
people have many influences on their lives- their age, social & economic environment, location, family
situation, parental and peer groups etc; hence a single strategy is ineffectual. A broad complement of strategies
would include:
4. Filtering/parental controls: There has been much discussion recently in relation to filtering or parental
controls and whilst we support “Active Choice”, we think that this is not a panacea. Parental controls have
advantages to help prevent accidental access to adult content, however we know it will not prevent determined
access, especially for “tech savvy” teenagers. Over the past 14 years we have been providing network level
filtering into schools in South-West England and continue to see “loop holes” that children discover to
circumvent and bypass the filtering in place. As we move towards multi device access in the home, our
preference is for network level filters compared to device specific ones. We encourage parents to use parental
controls and advocate age appropriate filtering, however always point out that these should not be used in
isolation but alongside other strategies—filtering and parental controls are not a “fire and forget” solution.
Consoles already use content ratings like PEGI and BBFC but these need to be more sophisticated on cable,
TV services, mobile devices and app devices.
5. Safe Search: Linked to filtering, we encourage parents to use safe search functionality and age appropriate
settings for online searches. It would be worth considering applying “active choice” to these services? For
example safe search could be enabled as the “default” configuration, allowing users to “opt out” if they wish?
6. Education for young people: Education is the most important strategy. We should provide children with
age appropriate, reflective, non-moralising and non-sensationalist educational programmes, mapped across the
wider curriculum and covering the broader aspects of digital life, alongside parental support initiatives.
Integrating discussions around sexting and impact of online pornography into RSE (Relationship & Sexual
Education) is in our view the most suitable approach, ensuring that children understand the context around
online adult content. Our fears relate to the influence of adult content on children’s expectations and normal
healthy sexual development. Whilst these are high level objectives, the effective success will require additional
education and support for those raising these issues with children.
7. Education for parents: In our experience parents can often feel intimidated when discussing technology
with their children and extending this to adult online content can exacerbate the situation. Many parents will
not appreciate the extent and extremes of adult content online, well beyond the sort of content they witnessed
as children. Support in this regard will be required, perhaps using a storyline in popular “soap operas” can
help to raise the subject and allow parents to both progress discussions and to provide some messaging to
children directly.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w49
8. Education for teachers: Our evidence suggests that online safety training for teachers is consistently the
weakest part of a school’s wider online safety practice. The impact of adult content should be integrated into
staff training and should extend to all staff. Pedagogical methods of integrating into RSE and online safety
sessions should be provided to teaching staff.
9. Social Workers: Social workers (and social care professionals) who work with society’s most vulnerable
children should be able to access materials, resources and support in this regard. Combating CSE (Child Sexual
Exploitation) is clearly a key priority in this sector and online adult content has a significant impact. Social
workers should be able to recognise the impact of this content in particular cases.
10. Service Providers: For those service providers who do not allow or host adult content (as defined by
the terms and conditions), they should be clear about how they monitor and enforce their policy, together with
clear and transparent reporting procedures for users to report questionable content. For those service providers
who do permit adult content, it should be clearly marked as such. Perhaps the use of age rating (ie 12, 15 and
18) could have a part to play? Providers who host user generated content should consider implementing user
or community ratings.
11. Public Health: Public health campaigns should be considered to raise general awareness of online adult
content, perhaps extending or integrating with existing teenage pregnancy campaigns, with which there is an
obvious impact.
12. Research: Further research is needed to better understand the experiences, perceptions and knowledge
of children and young people when it comes to online adult content together with assessments and evaluations
of the impact of this variety of formal and informal educational programmes and resources.
II. Filtering out extremist material including CAI (Child Abuse Images) and material intended to promote
terrorism and/or other acts of violence
13. IWF membership: UK ISPs should by mandatory requirement be members of the IWF (Internet Watch
Foundation). Combating CAI in the UK through the work of the IWF has proved a particularly successful
model and one that could be considered also for extremist material.
14. Using alert tools: SWGfL has been working with the IWF and South West police forces since 2006 to
alert for any attempted access to websites containing CAI. This pilot project has been managed by CEOP with
Home Office approval and due for evaluation shortly. The objective of the project is to flag intelligence to
police forces if anyone in a school (specifically) attempts to access a website containing CAI. This intelligence
prompts the police to then undertake their normal investigation routes resulting in a number of school staff,
who have been identified and removed as a direct result of this simple alerting process. We believe we should
use technology better to identify those accessing CAI (and extremist material) and we are involved in a project
with Plymouth University and IWF to extend and refine our existing alerting capability.
III. Preventing abusive or threatening comments on social media
15. Background to Professionals Online Safety Helpline: The Professionals Online Safety Helpline, based
at SWGfL is regularly called upon to comment on abusive online behaviours, both via printed and broadcast
media and at public speaking events across the UK. The Helpline is part of the Safer Internet Centre, and is a
dedicated service for professionals who work with children and young people, including teachers, social
workers, police officers, youth workers, early years professionals and more. We have seen a large increase in
calls from parents frustrated at the lack of practical help to get abusive comments removed online.
16. How the helpline operates: The helpline is unique in its approach to resolving social media issues, with
a direct contact route in all of the major sites. Our current partners include Facebook, Twitter, Ask.FM,
Instagram, Moshi Monsters, Tumblr, Apple, 192.com, Omegle, Snapchat and others. These contacts allow us
in the first instance to ensure that the advice we give to callers about how to report, is accurate, and secondly,
where necessary, to escalate the matter directly to the policy and safety personnel in each organisation, and
have inappropriate, abusive content removed. In some cases this also leads to the perpetrator being suspended
for their behaviour.
17. More guidance on how to report and remove content: Many services already exist to provide
therapeutic support, particularly for children. What has been lacking is the practical guidance for users on
reporting issues and ability to remove content. As such the helpline has addressed this gap and through its
unique network of direct contacts to providers, supports professionals, and increasingly parents, in how to
report abusive content appropriately in order to ensure or speed its removal.
18. Prevention strategies: Prevention is more challenging, and needs to be a collaboration between everyone
involved. Schools, and even pre-school and early years settings, need to educate children about online
behaviour much earlier. Generally where we see social media discussed and embedded in curriculum, there
tend to be fewer issues and children take a more responsible approach. We believe children need to build a
level of resilience, rather than being sheltered from experiences (of course we do not mean being subject to
online abuse!). Schools should talk about how to respond to abuse, for example around blocking and reporting,
and where possible introduce confidential reporting routes.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w50 Culture, Media and Sport Committee: Evidence
19. Strategies to empower young people and professionals: We know that children often do not “tell”
when they are being bullied or abused. Allowing young people to report anonymously, either for themselves
or their friends, can be very empowering. This can be simple, for example a “Whisper”52 alerting system via
SMS, or email, or even a locked comments box. Other initiatives such as Anti Bullying Ambassadors also help
discourage abusive behaviour. Schools need to act quickly and decisively when there are issues offline, often
the online abuse follows a period of face to face bullying. If this is picked up earlier, better interventions can
be put in place to support the child at risk.
20. Parental support: Parents need access to more up to date advice about specific popular sites, in how to
support their child if the worst happens, and what they can expect from the child’s school. Schools need to
alert parents of any bullying incidents at school as this can be an indicator that there or may be issues online
which the parent can assist with. Having regular, open conversations with their children about their online
activities can help to identify potential risks before they happen, help children feel more able to discuss
concerns, and also alert parents if their children are themselves engaging in bullying or abusive behaviours.
Parents need to take a more proactive response when they discover their children are perpetrators of online
abuse, including working with the school to ensure that they can support any victims who may not have already
come forward.
21. Role of industry: Industry needs to ensure that they have sufficient safeguards in place to protect users.
Our experience has been very varied, and some sites are better at dealing with issues than others. The most
effective systems use a combination of automated systems, such as language filters (although these are in no
way foolproof), automated moderation of specific phrases or search terms, and human moderation. It has been
our experience that with very rapidly growing sites, including Twitter and Ask.FM, user safety isn’t considered
as one of the priorities at the start. We are not overly critical of this, because having some flexibility around
policy decisions allows for regular review and improvements to be implemented, however it can be dangerous
to launch a site without some basic safeguards in place from the offset. An ideal approach would be a more
proactive moderation, rather than a purely “report”-based moderation. We do not support censoring of the web,
nonetheless making users aware that sites are monitoring behaviour may act as a deterrent for online abuse.
22. Clearer privacy settings and simple reporting routes: As well as working to prevent harassment, sites
must adhere to their own community standards when reports are made to them. We would like to see much
clearer advice on how to set up privacy on each site, report abuse, and clarification on what will happen once
a report has been made, eg timescale for resolution. Reporting should be a very simple process, and we are
pleased to see improvements to both Twitter and Ask.FM with the introduction of “in tweet” reports and a
more prominent report button soon coming to Ask.FM. Matters must be resolved quickly, and with decisive
sanctions, such as temporary suspension of accounts while investigations are underway. Our helpline is
undertaking a national role of mediation with providers in reporting and removing abusive content and could
be considered for national recognition like in New Zealand.
23. Accountability for anonymous services: We would also like to see more accountability to sites that
allow anonymous questions, such as Little Gossip. Although most media stories would say otherwise, Ask.FM
have been fairly proactive in dealing with cases reported to them, we have a named contact and they have kept
us informed of policy and procedures. Little Gossip however do not respond to any requests for contact and
act on abuse reports. This site allows users to report an abusive message by posting an equally abusive message
in its place, leading to significant harm to the young people named and shamed on the site. We would like
these sites to have an “opt in” to being anonymous, rather than “anonymous” being the default setting, and to
have warning messages on their landing pages.
24. Support links and help: Sites should have better links to support services which are relevant to each
country, for example Tumblr has links to self harm and suicide support services, but only in the US. Syndicated
support would be more effective, where different providers point to consistent sources of support—for example
to ChildLine, Samaritans etc.
25. Supporting organisations like the UK Safer Internet Centre: One area we would like improved is in
financial support from industry to organisations such as the UK Safer Internet Centre, to produce resources
and support operation of the Helpline, which directly tackles the issues users are facing. At present Industry
contribute financially to the IWF in tackling child abuse images, but there is no co-ordinated approach to
funding the services which resolve other problems such as removal of abusive content and in general, wider
education and awareness-raising initiatives.
26. Role of Government: Government has an integral part to play, statutory funding should be in place for
services working to resolve online harassment, and there needs to be clear direction and consistency between
departments. Too often we are given conflicting areas of focus from the Department of Justice, DCMS, DfE,
and even the Prime Minister. It would be more effective to streamline these through one distinct channel or
Department. Government should also be more active and consistent in engaging with social media sites when
there are issues in the public interest. There should be clearer guidance for the public on what legal action can
be taken, and under what circumstances.
52
http://boost.swgfl.org.uk/about.aspx
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w51
27. Role of media: One area often overlooked is the responsibility that Media holds. While it’s important
to openly discuss online issues, we would like to see a more responsible use of language and reporting on
behalf of the media. Often stories will have attention grabbing headlines, and very inflammatory language, for
example “Facebook bully probe after death” and “How many more teenage girls will commit suicide because
of Ask.fm”. In nearly all the cases we have been involved in, media have made sweeping statements and
assumptions about the circumstances of cases, without knowing the facts, and often while a Police investigation
is still underway. These messages are very unhelpful, distressing for the families involved and some even
contain inaccuracies about the sites which may lead to further problems. We would like to see the media
adhering to a general code of conduct about how they respond to issues involving young people, and adopting
more supportive safety advice and links to relevant resources.
28. Users responsibility: Finally, we need to consider the users themselves. In cases involving adults
particularly, we need to be clear that if they break the law they may be prosecuted, that just as inappropriate
offline behaviour may have legal consequences so may behaviour online. Young people’s issues however are
sometimes more complex, and we don’t support criminalising children, however there does need to be a point
when young people are held accountable for their behaviour. Awareness of those laws could be raised through
an advertising campaign similar to TV licence/Road safety etc.
September 2013
Written evidence submitted by British Sky Broadcasting Limited (‘Sky’)
1. Introduction
1.1 We welcome the Committee’s Inquiry into Online Safety as a further contribution to a topic that Sky
has been heavily engaged in.
1.2 In June, we attended the summit chaired by the Secretary Of State for Culture Media and Sport, and
have attached a letter we sent to her ahead of this summit as an appendix to this submission.
1.3 Over the last 20 years BSkyB has established itself as a successful and responsible consumer business.
Since entering the broadband market in 2006, we have become the second largest provider in the UK with
almost 5m broadband customers.
1.4 Our growth has come by providing customers with the products and services they want, and recognising
the trust that our customers place in us to act responsibly. Over this time we have created a track record
where we have helped families keep their children safe when enjoying the content we provide them on-screen
and online.
1.5 The internet has become one of the most important sources of information, education and entertainment.
Access to the internet is now an integral and fundamental part of people’s lives, bringing with it a wide range
of benefits but also a range of challenges.
1.6 There are many positive aspects to life in the online world, but just as in the offline world, there are
services and websites that provide access to material that is not suitable for children. We have taken our
philosophy of giving parents market leading tools to protect their families on screen, and applied it to our
broadband services.
1.7 As a result, we have developed a range of tools to help parents manage how their families access
content on whichever platform they choose to use, and are looking forward to enhancing our offering in the
coming months.
1.8 There is a very important distinction to be made between content that is legal but potentially unsuitable
for children, and other content which is illegal. It is vital that this distinction is not only understood by
policymakers but highlighted at every opportunity to help ensure that there is more clarity in the debate, and
in the way in which it is reported.
1.9 Sky’s commitment to provide families with tools to choose how much of the internet gets through to
their home, is separate to any legal or regulatory responsibilities we may have. Where content is deemed illegal
by legislators or the Courts we will do everything required of us to prevent or restrict access.
2. Protecting Minors from Accessing Adult Content
2.1 As a leading consumer brand, our customers have high expectations of Sky. They expect us to provide
effective filtering products, presented clearly in an easily manageable way, which will enable parents to protect
their families from legal, but potentially inappropriate content.
2.2 In addition to the work that we are doing to use technology to help parents, we believe that believe that
industry, government and educators can work together to provide parents with the necessary information and
advice to be able to make the most appropriate choices for their children, and to guide children and young
people to make the right choices for themselves.
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w52 Culture, Media and Sport Committee: Evidence
In home fixed line broadband
2.3 We offer all of our broadband customers free parental controls which are easily installed on customer’s
PCs or laptops. We use a category based web filtering package, which can be configured by age.
2.4 The internet is being accessed via an increasing number of connected devices including connected
televisions, gaming consoles, tablets and smartphones. As internet use and the number of connected devices in
customers’ homes increase, so do the challenges for protecting families. Whilst the existing controls are an
important tool for parents, as technology has developed, there is more that we are able to do.
2.5 We are in the process of developing an enhanced parental control product which will protect all the
devices in customers’ home. Whilst there remains a role for device manufacturers and operating system
providers, we believe it is right for us to take responsibility and empower our customers to protect their families
so they have confidence that the internet can be enjoyed in a safe environment.
2.6 Over the past few years, there have been policy questions as to the best way of delivering parental
controls. Through our involvement in UK Council for Child Internet Safety we have engaged in Government
initiatives such as the Bailey Review and the Department for Education’s 2012 consultation, and the delivery
of our new product is consistent with the features described in the Prime Minister’s July speech on online safety.
2.7 Our enhanced parental control product will be free and available to all customers from the autumn. In
addition to making the product available we are keen to encourage take up, and we will do this in a number
of ways.
I. From launch, when customers obtain new routers (by upgrading their router, moving to Sky
Fibre or joining Sky as a new customer) they will be presented with the choice as to whether
or not to implement our parental controls product. Furthermore, the “yes” button will be
preticked. This means that unless customers opt-out, and click “no”, the customer journey will
see parents install and configure parental controls to meet the requirements of their family. We
have also introduced measures to ensure that an “age verified” adult is in control of the
installation process and any subsequent amendments to the configuration of these controls.
II. We will present all existing customers with a choice whether or not to apply the filters, and
ensure that they have all made this decision by the end of 2014.
III. We will support our new product with a major marketing campaign to encourage the adoption
of the filters.
IV. We are working with other ISPs and internet companies to highlight the importance of online
safety, the use of parental controls, and to provide parents with more information on how they
can protect their families in the online space. This will take the form of a major campaign
launched early next year.
Out of home WiFi
2.8 Increasingly connectivity extends beyond the home. Sky’s WiFi service, The Cloud, for example,
provides customers with access to over 20,000 WiFi hotspots across Britain’s High Streets, in venues such as
Pizza Express and Greggs, and other public spaces like the London Overground railway stations.
2.9 When children are accessing the internet outside of the home, it is far more difficult for parents to
monitor their activity, and our research tells us this is of concern to parents.
2.10 The Cloud pioneered offering its commercial customers the opportunity to automatically filter out
access to adult websites in public WiFi hotspots, being the first WiFi provider to do so in 2007.
2.11 The increase in children accessing internet outside the home through connected devices, has led to
parents seeking greater reassurance. That is why from September 2012, we decided to apply these filters by
default, so that sites which contain content only suitable to adults are inaccessible in the public places served
by The Cloud. As a leading WiFi provider, we think this is an important contribution to ensuring children are
protected when using the internet outside of their homes, providing reassurance for parents.
3. Illegal Content
3.1 Separate from the issue of providing tools to customers to filter legal but potentially inappropriate
content is our commitment to meet any obligations imposed on us to restrict access to content as a result of a
legal determination.
3.2 Sky, as an ISP, is not in a position to judge the legality of a website or content within a website. Instead,
Sky works with the bodies responsible for making such determinations and complies with any obligations it
may have.
3.3 Content determined to be illegal tends to fall under one of three categories, namely child abuse imagery,
material intended to promote terrorism and content infringing copyright.
cobber Pack: U PL: CWE1 [O]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Culture, Media and Sport Committee: Evidence Ev w53
Child abuse imagery
3.4 Sky is a member of the Internet Watch Foundation (“IWF”), which is the UK Hotline for reporting
online child abuse imagery. Its key role is to get content removed at source. Where content is hosted in the
UK, the IWF is typically able to get it removed in under an hour. Where content is hosted overseas, it takes
longer to get it removed. However, the IWF provides its members with a list of URLs which contain illegal
content. Sky ensures that requests to access such webpages, regardless of where they are hosted, result in a
“splashpage” being served that informs the user that they attempted to access a page that contains illegal
content.
3.5 Sky is a strong supporter of the IWF. We are currently engaging with the IWF to see how we can
increase our support to it, both in terms of supporting the work they do and in terms of our financial support,
where we are already part of the “top tier” of financial contributors. In particular, we are keen for the IWF to
be proactive in seeking out illegal content rather than rely on public reporting.
3.6 It is worth noting that we have recently changed the splashpage served after requests from government
and after consultation with the IWF, the Child Exploitation and Online Protection Centre (CEOP), and the
Lucy Faithfull Foundation (LFF), which is a charity dedicated to reducing the risk of children being sexually
abused. The splashpage now is more explicit in highlighting the consequences of viewing child abuse imagery
linking to the LFF to provide further information.
Material intended to promote terrorism
3.7 We are keen to play our role in working with other Government departments, such as the Home Office,
and its Counter Terrorism Internet Referral Unit (CTIRU) to ensure that illegal content is removed and/or
filtered.
3.8 CTIRU was set up as part of the PREVENT strategy, to ensure unlawful online content is removed from
the internet, thereby disrupting and limiting access to extremist material. Since its establishment in 2010, this
unit has removed approximately 6,000 pieces of online terrorist content.
3.9 Sky does not provide webpage hosting, and as a result has are not been requested to remove this type
of content. However, we are keen to liaise with the government to explore whether there are other opportunities
to work together.
Copyright infringement
3.10 Sky is keen to support rightsholders in seeking to restrict access to content that breaches copyright.
3.11 Although Sky is not in a position to determine which sites are making illegal content available, we
welcome decisions by the Courts to require blocking of infringing sites, and have built systems to ensure we
can comply with blocking requirements of resulting court orders.
3.12 We believe that the established process broadly works and has been accepted by our customers. Indeed,
evidence from Ofcom’s research into online copyright infringement suggests a reduction in the people using
infringing sites,53 implying that blocking is a proportionate contribution to the fight against piracy.
September 2013
Written evidence submitted by Big Brother Watch
The Culture, Media and Sport Committee has decided to investigate a number of aspects of online safety
that are currently raising concerns, in particular:
1. How Best to Protect Minors from Accessing Adult Content
Firstly, it should be asked—is it the role of Government to prevent minors accessing adult content, or
parents?
Critically—any measure must consider the privacy and freedom of speech issues involved. Given the
importance of trying to help young people appreciate the importance of protecting their personal data and to
consider what they are sharing online, intrusive “big sister” approaches should be resisted. If we raise young
children with an expectation that they have no privacy, we cannot in later life then hope they discover the
importance of privacy.
With nine in 10 parents aware that internet filters are freely available, it is clear that internet controls need
to be made easier to use rather than taking the control away from parents and handing it to the government.
53
http://stakeholders.ofcom.org.uk/binaries/research/telecoms-research/online-copyright/w4/OCI_MAIN_REPORT_W4_
FINAL.pdf
“Thirty-two per cent of those who consumed any content illegally claimed to use “peer-to-peer” (P2P) services…..This was the
lowest level seen during the 13 months of tracking (it was 37% in [the last set of research])”
cobber Pack: U PL: CWE1 [E]
Processed: [17-03-2014 16:07] Job: 035306 Unit: PG01
Ev w54 Culture, Media and Sport Committee: Evidence
The first problem associated with online safety is not only are there issues trying to define adult content, but
there are differing views on what is unacceptable. For instance, some households would see topless photographs
as adult content whereas others would not. This is w