Facebook adds new tools to fight online child exploitation


The social media giant updated its guidelines to make it clear it will remove Facebook or Instagram accounts dedicated to sharing images of children posted along with captions, hashtags or comments containing innuendo or inappropriate signs of affection. — AFP

SAN FRANCISCO: Facebook on Feb 23 said it is stepping up its fight against child abuse with new tools for spotting such content and tighter rules about what crosses the line.

“Using our apps to harm children is abhorrent and unacceptable,” global head of safety Antigone Davis said in a blog post.

“We are developing targeted solutions, including new tools and policies to reduce the sharing of this type of content.”

The social media giant updated its guidelines to make it clear it will remove Facebook or Instagram accounts dedicated to sharing images of children posted along with captions, hashtags or comments containing innuendo or inappropriate signs of affection.

“We’ve always removed content that explicitly sexualises children, but content that isn’t explicit and doesn’t depict child nudity is harder to define,” Davis said.

“Under this new policy, while the images alone may not break our rules, the accompanying text can help us better determine whether the content is sexualising children and if the associated profile, page, group or account should be removed.”

New tools being tested included one that triggers pop-up messages in response to search terms associated with child exploitation, warning of the consequences of viewing such material and suggesting people get help changing the behaviour.

Facebook is also testing a safety alert that informs people sharing child exploitation content about the harm it causes and the legal consequences, according to Davis.

Along with removing content violating Facebook rules, such posts are reported to the National Center for Missing and Exploited Children (NCMEC).

“We are using insights from this safety alert to help us identify behavioural signals of those who might be at risk of sharing this material,” Davis said.

An analysis of illegal child exploitive posts shared with the NCMEC late last year found that more than 90% of it was the same or very similar to previously reported content, according to Facebook.

Just six videos accounted for more than half the content reported in that period, Davis said.

Facebook worked with the NCMEC and other groups to glean the apparent intent of people sharing such content.

It was concluded that more than 75% of the sharing scrutinised did not appear to be malicious, but was done for reasons such as expressing outrage or in poor attempts at humour, according to Davis.

Facebook has sparked concerns among law enforcement agencies with its plans to provide end-to-end encryption at all its messaging platforms in a move that police say could let criminals hide communications. – AFP

Article type: metered
User Type: anonymous web
User Status:
Campaign ID: 46
Cxense type: free
User access status: 3
   

Next In Tech News

Porsche to set up joint venture with German battery maker
Australian research finds 'pervasive' privacy breaches on health apps
Are games in children's learning apps harmless or addictive?
Learning a musical instrument by app: How well does it really work?
Like it or not, Facebook is still around but who's still on it?
Wall of sound: Ikea's new Sonos speaker is literally a piece of art
AI-powered Mayflower, beset with glitch, returns to England
Pressure on Pornhub grows after 34 women sue
As Cyberpunk reboots, can unloved games win an extra life?
After user exodus, WhatsApp promises new privacy features

Stories You'll Enjoy


Vouchers