The landscape of digital content regulation is undergoing significant scrutiny as technology firms face mounting pressures to curb the spread of illegal material. A recent proposal from Ofcom, the UK’s telecoms regulator, emphasizes the need for enhanced online safety measures aimed at protecting vulnerable users, particularly children. This includes urgent demands for tech companies to implement mechanisms that would prevent the viral dissemination of illegal content.
On June 30, 2025, Ofcom initiated a consultation to gather insights from key stakeholders on various strategies designed to fortify online safety regulations. With the primary goal of safeguarding citizens, especially minors, the proposed measures encompass additional requirements for tech platforms to assess their capabilities in proactively identifying and removing terrorist-related content before it spreads. The initiative underscores the importance of responsiveness to “constantly evolving” online risks, as articulated by Oliver Griffiths, the director of the online safety group at Ofcom.
In detailing the proposed measures, Griffiths indicated that the strategies are an extension of existing UK online safety rules. The consultation identifies three critical areas of focus: preventing illegal content from reaching broader audiences, addressing root causes of online harm, and enhancing protections for children navigating these digital spaces. These measures serve to hold technology platforms accountable, paving the way for more assertive enforcement actions where there are identified concerns about user safety.
Moreover, Ofcom’s proposals address a holistic array of issues that span from intimate image abuse to the dangers of live-streamed physical harm. For instance, platforms would be required to establish systems that enable users to report live-streams showcasing imminent risks of bodily harm, targeting user-to-user sites that allow individual users to broadcast to many viewers. Notably, these recommendations only apply to larger platforms presenting heightened risks of associated harms, underscoring a tailored approach to online safety regulation.
While some advocates welcome additional safety measures, there are calls for a more robust response to systemic failures within existing legislation. Ian Russell, chair of the Molly Rose Foundation, criticized Ofcom’s proposals as lacking in ambition. He argues that without comprehensive solutions, regulatory efforts will fall short, failing to keep pace with emerging trends in online threats such as suicide and self-harm.
Russell’s comments stem from personal tragedy; his daughter Molly took her life after being exposed to harmful online content, and he stresses that the current regulatory framework does not adequately compel companies to proactively identify and mitigate the risks posed by their platforms. His appeal for interventions from top government officials reflects a desire for a fortified Online Safety Act capable of addressing preventable harm directly.
As part of the consultation process, which remains open until October 20, 2025, Ofcom aims to engage a diverse array of contributors, including service providers, civil society organizations, law enforcement, and the general public. The initiative aligns with broader efforts by tech platforms to align their services with the UK’s comprehensive online safety regulations, with companies like TikTok and YouTube already enacting changes to restrict livestreaming for minors to mitigate risks of exploitation and harmful interactions.
For example, TikTok recently instituted a policy raising the minimum age for users to livestream from 16 to 18, following disclosures about concerning content linked to child exploitation. In a similar vein, YouTube has adjusted its own minimum requirements for livestreaming, seeking to foster a safer online environment for younger audiences.
In a world increasingly reliant on digital interactions, these regulatory steps taken by Ofcom and the conversations they provoke are vital for safeguarding users and constructing a framework that not only reacts to existing harms but anticipates new risks as they emerge. As the landscape evolves, the dialogue surrounding online safety will remain crucial, demanding commitment from all stakeholders involved in the tech ecosystem.