AI-generated child sexual abuse images could flood the Internet. A watchdog is calling for action


Computer-generated child sexual abuse images made with artificial intelligence tools like Stable Diffusion are starting to proliferate on the Internet and are so realistic that they can be indistinguishable from photographs depicting actual children, according to a new report. — AP

NEW YORK: The already-alarming proliferation of child sexual abuse images on the Internet could become much worse if something is not done to put controls on artificial intelligence tools that generate deepfake photos, a watchdog agency warned on Oct 24.

In a written report, the UK-based Internet Watch Foundation urges governments and technology providers to act quickly before a flood of AI-generated images of child sexual abuse overwhelms law enforcement investigators and vastly expands the pool of potential victims.

Save 30% OFF The Star Digital Access

Monthly Plan

RM 13.90/month

RM 9.73/month

Billed as RM 9.73 for the 1st month, RM 13.90 thereafter.

Best Value

Annual Plan

RM 12.33/month

RM 8.63/month

Billed as RM 103.60 for the 1st year, RM 148 thereafter.

Follow us on our official WhatsApp channel for breaking news alerts and key updates!

Next In Tech News

Musk says steps to stop Russia from using Starlink seem to have worked
French tech company Capgemini says selling US subsidiary
Indonesia lets Elon Musk's Grok resume, lifting ban over sexualised images
I'm a parent, how worried should I be about AI?
Elon Musk's Grok generated 3 million sexualised images in just 11 days, new analysis finds
After robotaxi hits child, Waymo says its software prevented worse
Elon Musk says ‘singularity’ is here – What to know about AI threats to humanity
Waymo seeking about $16 billion near $110 billion valuation, Bloomberg News reports
Bitcoin falls below $80,000, continuing decline as liquidity worries mount
SpaceX seeks FCC nod for solar-powered satellite data centers for AI

Others Also Read