wordpress blog stats
Connect with us

Hi, what are you looking for?

On Facebook adding 3000 people to monitor flagged content

The global country that is Facebook will hire 3000 people by the end of this year to “review the millions of reports we get every week, and improve the process for doing it quickly”, its founder (and head of state) Mark Zuckerberg said in a post today.

Facebook reviews content according to its community guidelines, which disallow hate speech and child exploitation. “And we’ll keep working with local community groups and law enforcement who are in the best position to help someone if they need it -either because they’re about to harm themselves, or because they’re in danger from someone else.”

Policing content via community is going to be tough

Facebook doesn’t actively police content, but relies on its community to flag content, which is then reviewed. This is, in a manner of speaking, reactive policing. This essentially absolves Facebook of prior knowledge before content is posted, and allows it to play the role of an intermediary and a tech platform merely allowing others to publish content. If it were to screen content, it would, like media publications, have liability over that content.

Facebook’s challenge with live streaming is far greater than policing hate speech, and it includes suicide, murder, and to people considering committing suicide.  Zuckerberg, in his post, says “Just last week, we got a report that someone on Live was considering suicide. We immediately reached out to law enforcement, and they were able to prevent him from hurting himself. In other cases, we weren’t so fortunate.”

Earlier last month,  Justin Osofsky, Facebook’s VP global operations, wrote in a blog post that Facebook is reviewing its reporting flows and reviewing its reporting process, but he also indicated a timeline of how the community on Facebook reacted, following an incident in Cleveland:

Advertisement. Scroll to continue reading.

Timeline of Events
11:09AM PDT — First video, of intent to murder, uploaded. Not reported to Facebook.
11:11AM PDT — Second video, of shooting, uploaded.
11:22AM PDT — Suspect confesses to murder while using Live, is live for 5 minutes.
11:27AM PDT — Live ends, and Live video is first reported shortly after.
12:59PM PDT — Video of shooting is first reported.
1:22PM PDT — Suspect’s account disabled; all videos no longer visible to public.

As you can see, depending on people to report is not very effective.

Even with and addition of 3000 people by the end of the year, Facebook is clearly understaffed: it will have around 7500 people to monitor flagged content from a daily active user (citizen) base of 1.3 billion (and growing), with millions of instances of content flagged a week. Facebook will thus also build “better tools to keep our community safe.”


These tools, according to Zuckerberg, will involve:
– Making it simpler for users to report problems (which could lead to an increase in reports)
– Faster for our reviewers to determine which posts violate our standards (which could, if it works well, helps address the increase in reports)
– Easier for reviewers to contact law enforcement.

I don’t quite remember which company it was, but at the world wide web conference in Hyderabad in 2011, I remember a developer talking about how a Russian video messaging streaming site (which allowed strangers to interact via video) dealt with nudity on its platforms: its image recognitions algorithms were trained to spot instances of nudity, and the stream would shut down. Of course, like I mentioned earlier, the challenges that Facebook are far greater, owing to its scale.

The problem is that with time, and with an increase in incidents, regulators will step in: the more the instances of shocking and explicit video content being streamed live, the more the pressure from courts and governments to get Facebook to prevent live-streaming. Beyond a point, Facebook may not be able to retain its safe harbour from Intermediary Liability: it’s only a matter of time.

Advertisement. Scroll to continue reading.

Written By

Founder @ MediaNama. TED Fellow. Asia21 Fellow @ Asia Society. Co-founder SaveTheInternet.in and Internet Freedom Foundation. Advisory board @ CyberBRICS

MediaNama’s mission is to help build a digital ecosystem which is open, fair, global and competitive.



Do we have an enabling system for the National Data Governance Framework Policy (NDGFP) aiming to create a repository of non-personal data?


A viewpoint on why the regulation of cryptocurrencies and crypto exchnages under 2019's E-Commerce Rules puts it in a 'grey area'


India's IT Rules mandate a GAC to address user 'grievances' , but is re-instatement of content removed by a platform a power it should...


There is a need for reconceptualizing personal, non-personal data and the concept of privacy itself for regulators to effectively protect data


Existing consumer protection regulations are not sufficient to cover the extent of protection that a crypto-investor would require.

You May Also Like


Google has released a Google Travel Trends Report which states that branded budget hotel search queries grew 179% year over year (YOY) in India, in...


135 job openings in over 60 companies are listed at our free Digital and Mobile Job Board: If you’re looking for a job, or...


Rajesh Kumar* doesn’t have many enemies in life. But, Uber, for which he drives a cab everyday, is starting to look like one, he...


By Aroon Deep and Aditya Chunduru You’re reading it here first: Twitter has complied with government requests to censor 52 tweets that mostly criticised...

MediaNama is the premier source of information and analysis on Technology Policy in India. More about MediaNama, and contact information, here.

© 2008-2021 Mixed Bag Media Pvt. Ltd. Developed By PixelVJ

Subscribe to our daily newsletter
Your email address:*
Please enter all required fields Click to hide
Correct invalid entries Click to hide

© 2008-2021 Mixed Bag Media Pvt. Ltd. Developed By PixelVJ