Better protections from harmful online videos
People who use online video-sharing sites and apps should be better protected from harmful content, as Ofcom issues new guidance for tech companies today.
- Ofcom sets out guidance to help video-sharing platforms keep users safer
- One in three video-sharing users come across hate speech
- Our approach balances protection of users with upholding freedom of expression
Video sharing platforms (VSPs) are a type of online video service where users can upload and share videos with other members of the public. They allow people to engage with a wide range of content and social features.
VSPs established in the UK – such as TikTok, Snapchat, Vimeo and Twitch – are required by law to take measures to protect under-18s from potentially harmful video content; and all users from videos likely to incite violence or hatred, as well as certain types of criminal content.
Ofcom research (PDF, 4.6 MB) shows that a third of users say they have witnessed or experienced hateful content; a quarter claim they’ve been exposed to violent or disturbing content; while one in five have been exposed to videos or content that encouraged racism.
Today’s best practice guidance is designed to help companies understand their new obligations and judge how best to protect their users from this kind of harmful material. We have already begun discussing with platforms what their responsibilities are, and what they are doing to comply with them.
What platforms should do to protect users
Ofcom’s job is to enforce the rules set out in legislation and hold VSPs to account. Unlike in our broadcasting work, our role is not to assess individual videos. And the massive volume of online content means it is impossible to prevent every instance of harm.
Instead, the laws focus on the measures providers must take, as appropriate, to protect their users – and afford companies flexibility in how they do that. To help them meet their obligations to protect users, our guidance sets an expectation that VSPs should:
- Provide clear rules around uploading content. Uploading content relating to terrorism, child sexual abuse material or racism is a criminal offence. Platforms should have clear, visible terms and conditions which prohibit this – and enforce them effectively.
- Have easy reporting and complaint processes. Companies should implement tools that allow users to flag harmful videos easily. They should signpost how quickly they will respond, and be open about any action taken. Providers should offer a route for users formally to raise concerns with the platform, and to challenge their decisions. This is vital to protect the rights and interests of users who upload and share content.
- Restrict access to adult sites. VSPs that host pornographic material should have robust age-verification in place, to protect under-18s from accessing such material.
Plans and priorities for the year ahead
One of our five priorities for the year ahead – as set out in our workplan (PDF, 685.8 KB) – is to work with VSPs to reduce the risk of child sexual abuse material.
The Internet Watch Foundation reported a 77% increase in the amount of “self-generated” abuse content in 2020. Adult VSPs carry a heightened risk of child sexual abuse material and the rise in direct-to-fans subscription sites specialising in user-generated adult content has potentially made this risk more pronounced. Given this heightened risk, we expect that VSPs’ creator registration processes and subsequent checks are strong enough to significantly reduce the risk of child sexual abuse material being uploaded and shared on their platforms.
Over the next 12 months we will also prioritise: tackling online hate and terror; ensuring an age-appropriate experience on platforms popular with under-18s; laying the foundations for age verification on adult sites; and ensuring VSPs’ processes for reporting harmful content are effective.
Our approach to enforcement and reporting
We will take a rigorous but fair approach to our new duties. As for TV and radio, we will balance protecting people from harm, with rights to freedom of expression.
If we find a VSP provider has breached its obligations to take appropriate measures to protect users, we have the power to investigate and take action against a platform. This could include fines, requiring the provider to take specific action, or – in the most serious cases – suspending or restricting the service.
We also have a broad range of new powers to collect information from providers about what they are doing to tackle user safety on their services.
In autumn next year, we will publish a first-of-its-kind report providing transparency for users and the wider public on the steps VSPs are taking to protect children and other users from harm.
Dame Melanie Dawes, Ofcom Chief Executive said:
Online videos play a huge role in our lives now, particularly for children. But many people see hateful, violent or inappropriate material while using them.
The platforms where these videos are shared now have a legal duty to take steps to protect their users. So we’re stepping up our oversight of these tech companies, while also gearing up for the task of tackling a much wider range of online harms in the future.
Notes to Editors
- We commissioned bespoke consumer research to inform our approach to VSP regulation. Fieldwork for this research was conducted between September and October 2020, and findings relate to the three months prior to interview. We have also published two reports from leading academics from The Alan Turing Institute, covering online hate; and from the Institute of Connected Communities at the University of East London, on protection of minors online.
- Ofcom’s video-sharing platform framework: a guide for industry
Latest News from
Making sure you’re on the best deal for you is vital01/12/2021 12:25:00
Ofcom’s Director of Telecoms Consumer Protection, Cristina Luna-Esteban, explains why making sure you’re on the best phone or broadband deal for you has perhaps never been more important – and how we’ve made it easier to do so.
Telecoms customers saving millions as Ofcom rules bed in30/11/2021 15:10:00
More broadband and mobile customers are shopping around and signing up to better deals – and saving themselves millions of pounds – following Ofcom rule changes.
BBC Three can return as TV channel26/11/2021 13:15:00
BBC Three is set to return to TV screens for the first time in six years, after Ofcom yesterday approved its relaunch as a broadcast television channel.
The time has come for strong, external oversight of social media companies25/11/2021 09:20:00
Our Chief Executive, Melanie Dawes, explains why we can’t rely on tech media companies responding to users’ outrage when serious harm takes place. Instead, we need a clear set of rules.
Preparing for our new online safety duties23/11/2021 13:10:00
We’re looking forward to taking on greater responsibilities for helping people lead a safer life online, and are starting to prepare for taking on this new role and developing relationships with a new set of companies that we haven’t previously regulated.
New powers for Ofcom to oversee security of telecoms networks22/11/2021 10:15:00
The government’s Telecommunications (Security) Act recently become law last week (19 November 2021), introducing new powers for Ofcom to help make sure that the UK’s telecoms networks are safe and secure.
Latest telecoms and pay-TV complaints revealed16/11/2021 13:15:00
Ofcom has today published the latest league tables on the complaints we receive about the UK’s major home phone, broadband, mobile and pay-TV firms.
How to protect yourself from ‘slamming’15/11/2021 10:15:00
Ofcom recently (12 November 2021) fined two telecoms firms – Guaranteed Telecom and Met Technologies – a combined total of £35,000 after they took over the home phone services of more than a hundred people without their permission – a practice known as ‘slamming’.