Why does Facebook fail to fix itself? It’s partly humans

  • TECH
  • Friday, 10 May 2019

FILE - In this April 11, 2018, file photo, Facebook CEO Mark Zuckerberg testifies before a House Energy and Commerce hearing on Capitol Hill in Washington. Facebook gives the impression that it’s stopping the vast majority of extremist posts before users ever see them. But a confidential whistleblower’s complaint to the Securities and Exchange Commission obtained by The Associated Press alleges the social media company has exaggerated its success. (AP Photo/Andrew Harnik, File)

SAN FRANCISCO: The question comes up over and over, with extremist material, hate speech, election meddling and privacy invasions. Why can’t Facebook just fix it? 

It’s complicated, with reasons that include Facebook’s size, its business model and technical limitations, not to mention years of unchecked growth. Oh, and the element of human nature. 

The latest revelation: Facebook is inadvertently creating celebratory videos using extremist content and auto-generating business pages for the likes of ISIS and Al Qaida. The company says it is working on solutions and the problems are getting better. That is true, but critics say better is not good enough when mass shootings are being livestreamed and online mobs are spreading rumours that lead to deadly violence. 

“They have been frustratingly slow in dealing with everything from child sexual abuse to terrorism, white supremacy, bullying, non-consensual porn” and things like allowing advertisers to target categories such as “Jew hater”, simply because some users had listed the term as an “interest”, said Hany Farid, a digital forensics expert at the University of California, Berkeley. 

As new problems crop up, Facebook’s formula has been to apologise and promise to make changes, sometimes also noting that it did not anticipate how malicious actors could so readily misuse its platform. More recently, the company has also emphasised just how much it is improving, both technically in its use of artificial intelligence to detect problems and in terms of focusing more money and effort on fixing them. 

“After making heavy investments, we are detecting and removing terrorism content at a far higher success rate than even two years go,” Facebook said Wednesday in response to the revelations about the auto-generated pages. “We don’t claim to find everything, and we remain vigilant in our efforts against terrorist groups around the world.” 

It has seen some success. In late 2016, CEO Mark Zuckerberg infamously dismissed as “pretty crazy” the idea that fake news on his service could have swayed the election. He later backtracked, and since then the company has reduced the amount of misinformation shared on its service, as measured by several independent studies. 

Zuckerberg has also, by and large, avoided similar gaffes by conceding mistakes and delivering apologies to the public and to lawmakers. 

But even as the company bats down one problem, others pop up. The reason for that might be baked into its DNA. And that’s not just because its business model relies on as many people as possible using it as much as possible, leaving behind personal details that can then be targeted by advertisers. 

“Almost everything Facebook has designed has been designed for good people. People who are nice to each other, who have birthdays to celebrate, who have new puppies and generally like to treat others well,” said Siva Vaidhyanathan, director of the Centre for Media and Citizenship at the University of Virginia. “Basically Facebook is made for a better species than ours. If it were made for golden retrievers, everything would be great.” 

But if just 1% of the 2.4 billion people on Facebook want to do terrible things to others, that’s 24 million people. 

“Every couple of weeks, we hear about Facebook knocking down troublesome pages, making promises about hiring more people, building AI and so on,” Vaidhyanathan said. “But at Facebook’s scale, none of that will matter. We are basically stuck with all this garbage.” 

Chris Hughes, a co-founder of Facebook, called for a breakup of the social media giant in a Thursday op-ed. Vaidhyanathan also thinks strong government regulation could be the answer, such as laws that “limit companies’ ability to suck up all our data and use it to target advertising”. 

“We really should be addressing the back end of Facebook,” he said. “That’s what you have to attack.” – AP

Article type: metered
User Type: anonymous web
User Status:
Campaign ID: 1
Cxense type: free
User access status: 3

Did you find this article insightful?


Next In Tech News

Uber urges Hong Kong government to rethink clampdown, claiming livelihoods of 14,000 drivers at risk
Opinion: Threat of the digital giants and their unfettered trade monopolies
Google’s Messages app rolls out global support for successor to SMS
Contact lenses worthy of James Bond to zoom and film in a blink of an eye
Will remote working be the new norm in tomorrow’s world?
Facebook’s AI mistakenly bans ads for struggling businesses
HTTPS only: Firefox now blocks websites without encrypted connections
PlayStation 5 giveaways: Real deal, phishing or data scam?
For Big Tech, Biden brings a new era but no ease in scrutiny
Twitter has flagged 200 of Trump’s posts as ‘disputed’ or misleading since the US election day – does it make a difference?

Stories You'll Enjoy