Facebook to scrutinize abusive content on the social network
SCI & TECH
| November 16th 2018
Facebook announced Thursday it is creating an independent body to make potentially precedent-setting calls on what content should be yanked from the social network.
The announcement came as Facebook reported it has ramped up its ability to quickly detect "hate speech" and other posts violating community rules, with the leading social network under pressure from regulators in various countries and activists to root out abusive and inappropriate content.
"I have come to believe that we shouldn't be making so many decisions about free expression and safety on our own," Facebook chief executive Mark Zuckerberg said in a media briefing.
Content spied by artificial intelligence software or reported by users is now reviewed by an internal system that Facebook has been ramping up.
An independent body to be constituted in the coming year will act as a "higher court" of sorts, considering appeals of content removal decisions made by the social network, Zuckerberg said.
The composition of the appeals body along with how to keep it independent while remaining in line with Facebook principles and policies was to be determined in the coming year.
Facebook also planned next year to begin releasing content removal summaries quarterly in a tempo on par with earnings reports, according to executives.
"We have made progress getting hate, bullying and terrorism off our network," Zuckerberg said.
"It's about finding the right balance between giving people a voice and keeping people safe."
Challenges faced by the California-based social network include the fact that people naturally tend to engage with more sensational content that, while perhaps at the edge of violating Facebook policies, are unhealthy for civilized discourse, according to Zuckerberg.
"We see this in cable news and tabloids too," Zuckerberg said.
"A lot of our work is to insure that borderline content that comes close to violating our content gets less attention not more."
Bullying represents a tougher challenge to AI systems, because it tends to be personal and subjective. For example, someone might playfully mock a friend in a post that could also be interpreted to be mean.
Detecting bullying or hate can also require understanding of the gamut of languages used at Facebook, along with cultural contexts.
"We are getting better at proactively identifying violating content before anyone reports it, specifically for hate speech and violence and graphic content," Facebook said in the new transparency report.
"But, there are still areas where we have more work to do."
Facebook said that since its last transparency report, the amount of hate speech detected proactively, before anyone reported it, has more than doubled.
"The single biggest improvement comes from AI and machine learning," said product management vice president Guy Rosen.
Ghana’s budget deficit easing to 4.2 percentGhana budget deficit is expected to narrow from 4.5 in 2019 from 4.5 pct 2018 target with an estimated inflation of 8 percent.
Restoring Nairobi’s iconic librariesBook Bunk is turning public libraries into what they call ‘Palaces for The People' while introducing technology in every aspect.
Former MP John Serut battling cancer, daughter seeks aid to offset Sh20m bill
By Jael Mboga
- How suspected child killer Masten Wanjala easily left police station
By Kamore Maina
- Men, here are the natural foods to increase your ‘stamina’
- Independent candidate beats UDA and Wiper candidates to Nguu/Masumba Ward seat
- Suspected child killer Masten Wanjala is dead
- Lawyer Evans Monari amuses mourners at funeral with tribute to himself