Get all your news in one place.
100’s of premium titles.
One app.
Start reading
The Independent UK
The Independent UK
National
Martyn Landi

Ofcom lays out first plans to enforce new online safety rules

PA Wire

The biggest social media platforms will be required to protect children online by keeping them off suggested friend lists to stop them being contacted by groomers, Ofcom has said.

The new online safety regulator has published its first draft codes of practice under the Online Safety Act, which was signed into law last week.

The first codes focus on illegal material online – such as child sexual abuse material, grooming content and fraud.

Under the code, the largest platforms will be required, by default, to ensure that children on their sites are not presented with lists of suggested friends, do not appear in other users’ lists, that their location information is not visible to other users and that people outside their agreed connections cannot direct message them.

Regulation is here, and we’re wasting no time in setting out how we expect tech firms to protect people from illegal harm online, while upholding freedom of expression
— Dame Melanie Dawes, Ofcom

Ofcom is set to publish further codes in the coming months on other areas of online safety, such as guidance for adult sites on keeping children away and on protecting children from harmful content promoting things such as suicide or self-harm.

Each of the draft codes will have a consultation period before requiring final approval from Parliament.

The regulator’s own timetable says it hopes to begin enforcing its first codes of practice by the end of 2024.

The illegal content code also encourages larger sites to use hash matching technology to identify illegal images of sexual abuse and use automated tools to detect websites that have been identified as hosting abuse material.

On fighting fraud and terrorism, Ofcom says services should use automatic detection systems to find and remove posts linked to the sale of stolen financial information and block all accounts run by proscribed terrorist organisations.

The codes of practice also propose that tech firms nominate an accountable person who reports to senior management on compliance around illegal content, reporting and complaints duties, ensure their content moderation teams are well resourced and trained, offer easy reporting and blocking tools to use, and carry out safety tests on recommendation algorithms.

Dame Melanie Dawes, Ofcom’s chief executive, said: “Regulation is here, and we’re wasting no time in setting out how we expect tech firms to protect people from illegal harm online, while upholding freedom of expression.

Children have told us about the dangers they face, and we’re determined to create a safer life online for young people in particular.”

By working with companies to set out how they can comply with these duties, the first of their kind anywhere in the world, the process of implementation starts today
— Technology Secretary Michelle Donelan

Technology Secretary Michelle Donelan said the publication of the first codes marked a “crucial” step in making the Online Safety Act a reality by “cleaning up the wild west of social media and making the UK the safest place in the world to be online”.

“Before the Bill became law, we worked with Ofcom to make sure they could act swiftly to tackle the most harmful illegal content first,” she said.

“By working with companies to set out how they can comply with these duties, the first of their kind anywhere in the world, the process of implementation starts today.”

Ofcom said it had been and would continue working with social media and other in scope platforms over the coming months to help ensure they were in compliance with the proposed codes when they come into full force.

Campaign groups have backed the first proposals from the regulator.

Susie Hargreaves, chief executive of the Internet Watch Foundation, said: “We stand ready to work with Ofcom, and with companies looking to do the right thing to comply with the new laws.

It’s vital companies are proactive in assessing and understanding the potential risks on their platforms, and taking steps to make sure safety is designed in
— Susie Hargreaves, Internet Watch Foundation

“It’s right that protecting children and ensuring the spread of child sexual abuse imagery is stopped is top of the agenda.

“It’s vital companies are proactive in assessing and understanding the potential risks on their platforms, and taking steps to make sure safety is designed in.

“Making the internet safer does not end with this Bill becoming an Act. The scale of child sexual abuse, and the harms children are exposed to online, have escalated in the years this legislation has been going through Parliament.

“Companies in scope of the regulations now have a huge opportunity to be part of a real step forward in terms of child safety.”

Sign up to read this article
Read news from 100’s of titles, curated specifically for you.
Already a member? Sign in here
Related Stories
Top stories on inkl right now
One subscription that gives you access to news from hundreds of sites
Already a member? Sign in here
Our Picks
Fourteen days free
Download the app
One app. One membership.
100+ trusted global sources.