DeepNude app exploiting women in S’pore






Disrespectful act: One of the photos that was posted on the sex forum by a user who asked for it to be doctored. 
— The Straits Times/Asia News Network

Disrespectful act: One of the photos that was posted on the sex forum by a user who asked for it to be doctored. — The Straits Times/Asia News Network

SINGAPORE: She posted an innocuous selfie on social media more than a year ago. Last week, her photo appeared on a sex forum with a startling difference – she is naked.

On Tuesday, the 27-year-old woman, who wanted to be known only as Rose, was aghast when The New Paper told her that her photo had been doctored using artificial intelligence to show her in the nude.

She is not alone. Over the past week, dozens of women in Singapore have had their pictures on social media stolen, doctored and uploaded to the sex forum.

Some of these pictures have been compiled and recirculated on pornographic sites, with more additions every day.

The photos are believed to have been doctored using a version of the DeepNude app, which was launched several months ago.

It uses artificial intelligence to make women appear naked.

Its creators, who listed their location as Estonia, shut down the application last month following an uproar on social media.

They said the app was meant only for entertainment and they had not expected it to go viral.

Several versions of the software have since surfaced online. Versions of the app have been shared via download links on the sex forum, which has a high number of visitors from Singapore.

Forum users could also submit photos and request for them to be doctored by those who have the software. The doctored pictures would then be uploaded and circulated.

Rose believes circulation of her doctored photo had led to a recent surge in followers on her social media accounts.

She said: “This is so disgusting, disrespectful and perverted. What if someone else did this to their mothers, sisters, wives or girlfriends?”

She now intends to make her social media accounts private.

Lawyers said that while there has not been any reported prosecution here over the use of the DeepNude app, using it to doctor photos to make people appear naked is a criminal offence under the law.

Lawyer Fong Wei Li said that in the eyes of the law, creating such pictures is no different from taking an actual nude photo.

He said: “It is fundamentally the same, causing the same kind of harassment and alarm. At the end of the day, the content is obscene by objective standards.”

While the use of artificial intelligence is a seemingly new aspect, the broad definitions of current law allow for prosecution, even if those responsible hide behind anonymous usernames online, Fong added.

“Anonymity makes it difficult but not impossible to identify them. With their resources, the police can break through the barrier of anonymity to identify the people responsible,” he said.

Lawyer Gloria James said that under the Films Act, anyone who creates such pictures can be fined up to S$40,000 (RM121,556), jailed for up to two years, or both.

The culprits can also be charged with insult of modesty, for which they face a jail term of up to a year, a fine, or both.

Both lawyers said that victims can use the Protection from Harassment Act to take out protection orders against online users, even if they appear to be anonymous.

Fong said: “As long as the person is still identifiable via a username, the victim can still make an application.

“There are limitations, but it doesn’t mean nothing can be done.” — The Straits Times/Asia News Network

Singapore