UK Targets Non-Consensual Intimate Images with New Digital Tagging and Tough Penalties
February 18, 2026
Ofcom is weighing the use of digital tagging to automatically remove non-consensual intimate images when they are reposted, with guidance to internet providers to block access to sites hosting such content, including rogue sites beyond the Online Safety Act.
Non-consensual intimate images would be treated as a priority offence under the Online Safety Act, placing them on par with child sexual abuse and terrorism content.
Hash-matching technology would detect and remove non-consensual intimate images across platforms, potentially enabling automatic takedowns similar to procedures for child abuse and terrorism content.
Supporters, including lawyer Hanna Basha, welcome the progress but argue 48 hours may still be too slow and call for clearer reporting contacts and stronger accountability.
Context includes related political events, cabinet appointments, and ongoing debates about gender equality and institutional reform.
Public reaction is mixed, with strong support from women’s rights groups and criticism from free-speech advocates and some tech figures over censorship and feasibility, including concerns about arbitrary removals.
The mandate would impose significant costs on major platforms, requiring enhanced AI detection, more human moderators, and revised workflows, with penalties up to 10% of global revenue or UK blocking for non-compliance.
The announcement follows a prior confrontation with X over AI-generated images of real women, which led to removal of certain features for users.
The initiative shifts responsibility from victims to perpetrators and platforms, signaling forthcoming amendments to strengthen enforcement against AI-enabled abuse.
Prime Minister Keir Starmer underscored the urgency, referencing his past role as director of public prosecutions in addressing intimate image abuse.
Political momentum reflects a broader trend to hold tech giants accountable for online harms, aligning the UK with global regulatory movements on AI-enabled content misuse.
A July 2025 government report on digital violence and survivor-centric tools is cited to illustrate the evolving nature of intimate image abuse and remedies.
Summary based on 14 sources
Get a daily email with more World News stories
Sources

BBC News • Feb 19, 2026
Tech firms will have 48 hours to remove abusive images under new law
The Guardian • Feb 18, 2026
Tech firms must remove ‘revenge porn’ in 48 hours or risk being blocked, says Starmer
BBC News • Feb 18, 2026
Tech firms will have 48 hours to remove abusive images under new law