Go to Contents Go to Navigation

Facebook unveils report on enforcement of its community standards

Consumer Electronics 14:26 November 28, 2019

By Kim Han-joo

SEOUL, Nov. 28 (Yonhap) -- U.S. social networking giant Facebook Inc. said Thursday it has released an updated report on how it has been making efforts to detect content that violates community rules to proactively remove it from both Facebook and Instagram.

The report contains data and metrics that show which categories are the most problematic in terms of violations of Facebook rules and the number of cases that violated regulated standards.

Among the wide variety of content were child nudity, child sexual exploitation and terrorist propaganda, as well as illicit firearms, drug sales and suicide, according to Facebook.

"Our content policy team is a global network who work 24 hours and 365 days to monitor the potentially disturbing content," Yoo Dong-yeong, an official at the Facebook Asia Pacific regional office, told reporters in Seoul.

The corporate logo of Facebook provided by the company (PHOTO NOT FOR SALE) (Yonhap)

Facebook said 15,000 people at Facebook review potentially harmful content and provide reviews on reported content in over 50 languages, including Korean.

Facebook said its ability to spot and remove potentially harmful content on both websites and apps has vastly improved.

On Facebook, 99 percent of the removed content was "proactively detected," meaning the rest was reported. On Instagram, 94.6 percent of the removed content was proactively detected.

Some 2.5 million posts concerning suicide and self-harm were taken down from Facebook between the fourth quarter of last year and the first quarter of this year, while 845,000 pieces of similar content were removed from Instagram, according to Facebook.

Yoo said Facebook removes content that depict suicide or self-injury, even graphic imagery that experts deem to lead others to engage in similar behaviors.


Send Feedback
How can we improve?
Thanks for your feedback!