Assessing YouTube, Facebook and Twitter鈥檚 Content Takedown Policies
One year ago, 麻豆果冻传媒鈥檚 Open Technology Institute, as part of a coalition of organizations, advocates, and academic experts who support the right to free expression online, released the .1 The Santa Clara Principles outline the minimum standards tech platforms must meet in order to provide adequate transparency and accountability around their efforts to take down user-generated content or suspend accounts that violate their rules, provide meaningful due process to impacted speakers, and better ensure that the enforcement of their content guidelines is fair, unbiased, proportional, and respectful of users鈥 rights. The principles advocate for greater transparency and accountability by focusing on three key demands鈥攃omprehensive numbers detailing their content moderation activities, clear notice to affected users, and a robust appeals process.
On the first anniversary of the release of the Santa Clara Principles, we aim to assess if and how three of the largest internet platforms鈥擸ouTube, Facebook and Twitter鈥攈ave implemented the recommendations outlined in the Santa Clara Principles. Our findings indicate that although the three platforms have made greater progress in implementing the recommendations related to their 鈥渘otice鈥 and 鈥渁ppeals鈥 efforts, they still fall woefully short when it comes to meeting the standards set forth for the 鈥渘umbers鈥 category.
Over the past year, the online content moderation space has seen a number of important shifts that have sparked the need for such an assessment. In April 2018, Google, via YouTube, released its , which provided information on YouTube鈥檚 content removals based on violations of its Community Guidelines.2 This marked the first time any internet platform published data on this aspect of its content moderation practices, and it was soon followed by similar disclosures from other internet companies.
Shortly after the release of YouTube鈥檚 report, Facebook published a detailed version of its , which provide an overview of the internal guidelines the company uses to make moderation decisions.3 In addition, the company also began permitting users to appeal content takedown decisions made on individual posts, as opposed to only pages and groups. ( is OTI鈥檚 write-up on YouTube鈥檚 initial report and Facebook鈥檚 release of a detailed version of its Community Standards, with specific recommendations for how they could be improved.4) In May 2018, Facebook also released its first which published data on its content removals based on violations of its Community Standards.5
Finally, in December 2018, Twitter followed suit and began issuing in its transparency report. (Here is OTI鈥檚 initial write-up on that report from when it first came out, with specific recommendations for how it could be improved.)
Our assessment is based on data publicly available through the platforms鈥 transparency reports, blog posts, and press releases, as well as through responses to direct questions that we posed to the companies themselves.6 In addition, our assessment is based on the categories of rules-violating content that companies are currently highlighting in their transparency reports and other public materials. However, we recognize that these categories do not cover the full range of rules-violating content on these platforms, and urge companies to continue expanding their transparency efforts to include all of the relevant categories.7
We hope that our assessment here spurs them to do an even better job providing robust transparency reports and notice and appeal processes鈥攁nd that in doing so they will provide a strong example to other content platforms large and small.
Numbers
As demonstrated by the chart above, despite the fact that YouTube, Facebook, and Twitter have all issued transparency reports that cover content takedowns based on violations of their Terms of Service, these reports do not feature many of the granular recommendations around 鈥渘umbers鈥 set forth by the Santa Clara Principles. Out of the 20 recommendations on 鈥渘umbers,鈥 YouTube only fully meets four and partially meets three, Facebook fully meets none and only partially meets two, and Twitter only fully meets two and partially meets two. This demonstrates that these three platforms still have a long way to go before they are successful in providing a more granular level of transparency and accountability around their content takedown decisions.
In addition, as the chart demonstrates, there is a lack of standardization in what companies are currently reporting on in their transparency reports and there is no one recommendation with which all three companies are currently fully complying. This variation in data disclosures鈥攊ncluding the lack of basic data around the total number of flagged and removed posts on each service鈥攑revents meaningful comparison or combination of data across companies, which is ideally one of the key benefits of having transparency reports in the first place. We urge the companies to consider the detailed recommendations on standardization that OTI has offered in its Transparency Reporting Toolkit focused on content takedown reporting. 8
Some platforms have introduced their own unique metrics, such as Facebook鈥檚 prevalence metric, which measures 鈥渢he estimated percentage of views that were of violating content鈥 for each of the reported content categories. These metrics can be valuable in painting a broader picture of content takedown practices on a given platform, and in spotlighting issues related to content takedown practices that companies find important. We certainly encourage innovation in the development and implementation of such unique metrics and hope that it will continue. However, these unique metrics should supplement, not supplant, the basic standard metrics that have been set forth in the principles and described in detail in our toolkit.
Currently, YouTube鈥檚 Community Guidelines enforcement report provide the most comprehensive overview of Terms of Service-related content takedown practices, as the report provides data on channel, video, and comment removals, as well as some limited data on the number of videos flagged, the role automated tools and human flaggers played in flagging and getting content removed, and the categories of rules that were violated by removed content.
Twitter鈥檚 Rules Enforcement report offers the second-greatest level of transparency and granularity, as it provides data on the number of unique accounts that were reported and the number of unique accounts that were actioned. These data points are broken down based on the category of the Twitter Rules that were violated. However, Twitter鈥檚 report does not provide data related to the flagging or removal of individual pieces of content, such as tweets or comments.
Finally, Facebook鈥檚 Community Standards enforcement report provides insight into the amount of content, per category of rule violated, that the company took action on. Facebook鈥檚 report also provides data related to several of their own unique metrics, such as how much violating content Facebook identified before it was flagged to the company, and the previously discussed prevalence metric. However, it is overall the least successful at providing the range of basic data that the principles demand.
By issuing Terms of Service-related content takedown reports, YouTube, Facebook, and Twitter took important steps towards providing transparency and accountability around their content moderation practices. However, these reports fall woefully short when it comes to providing meaningful and granular data. In the future, these platforms should expand the data they report on, and standardize their reporting to enable meaningful comparisons and analyses.
Notice
YouTube, Facebook, and Twitter demonstrated greater progress in implementing the recommendations related to the 鈥渘otice鈥 aspect of the Santa Clara Principles. Out of the seven recommendations put forth, YouTube fully met five of them, Facebook fully met four and partially met one, and Twitter fully met four. In the case of all three companies, the notices provided to users who have had their content removed for violating a platform鈥檚 Terms of Service are relatively detailed and either fully or partially include some key pieces of information: reference to the specific rule that was violated, an explanation of how a user can appeal the takedown decision, and a durable form of the notice.
The main gap that still exists in this section of recommendations, however, is related to providing transparency around how a user鈥檚 content is detected and removed. Given that automated tools are increasingly being used to identify and remove content (e.g. YouTube鈥檚 Community Guidelines enforcement report highlighted that the majority of videos removed鈥6,190,148 in all鈥攚ere detected using automated flagging tools9), this is an aspect of the content moderation process that companies need to provide greater transparency around.
Appeals
YouTube, Facebook, and Twitter demonstrated the greatest level of success with the recommendations set forth in the 鈥渁ppeals鈥 section of the Santa Clara Principles. Out of the four recommendations put forth in this section, YouTube fully met all four, Facebook fully met three, and Twitter fully met three and partially met one.
In the cases where Facebook failed to provide users the opportunity to present additional information during the appeals process and Twitter only partially succeeded in ensuring appeals are reviewed by a different human or panel of persons than the original moderator, both companies had a specific explanation for these gaps, which are rooted in their specific content moderation approaches and policies. In the case of Facebook, for example, the content review process is structured so that moderators purposefully do not have access to external content (i.e. content that would not have been available to them in a traditional content moderation scenario). In the case of Twitter, initial reviews and appeals are typically handled by different teams of reviewers. However, given that some reports鈥攕uch as those related to sensitive policies, language skills, or cultural knowledge鈥攔equire specialized policy teams, this is not guaranteed. Overall, however, the companies have demonstrated a positive commitment to meeting the recommendations set forth in this section of the principles.
Going forward, we urge the companies to provide more information in their transparency reports around how effective their appeals processes are, and about how many users and pieces of content are impacted by these appeals processes. Facebook is already taking steps towards this, as it has committed to including in its future reporting a metric that reflects how often the company corrected its content takedown mistakes. The platform is also working to publish data on their appeals process in its transparency report as well.
Conclusion
Over the past year, YouTube, Facebook, and Twitter have taken leading steps to providing greater transparency and accountability around their content takedown practices. Although these platforms have made positive strides towards providing transparency around their notice and appeals processes, they have made minimal progress when it comes to publishing meaningful data points that illustrate the scope and scale of their content moderation efforts.
In the coming year, we hope these platforms will demonstrate a greater commitment to implementing the recommendations set forth in the Santa Clara Principles. This will not only enable the public to hold these companies accountable for their management of online speech and expression, but it will also encourage other players in the industry to similarly adopt the principles and provide transparency around their content takedown practices (, for example, has begun this process).
Going forward, we also hope that YouTube, Facebook, and Twitter, as three leading content regulating platforms in this space, will adopt greater standardization in the data they are reporting and the approaches they take to content takedowns. This will enable more meaningful comparisons and analyses of their efforts. In addition to this, we also welcome innovation on the part of the platforms themselves or from other organizations and individuals, in terms of the metrics and data points being reported on.
Two years ago, we knew far less than we do today about how major internet platforms are regulating and removing online content. The past year has demonstrated significant progress in shedding a light on these operations and fostering an industry-wide discussion on best practices and standards in this space. We look forward to another year of work on these issues, and hope that by 2020, the recommendations set forth in the Santa Clara Principles, and the values they are based on, are an industry-wide standard.
Acknowledgments
We would like to thank the drafters and signatories of the Santa Clara Principles for their support during the development of this brief. In addition, we would like to thank Kevin Bankston, the Director of the Open Technology Institute, for providing his expertise and editorial support during the review process.
Downloads
Citations
- Santa Clara Principles on Transparency and Accountability Around Online Content Moderation,
- YouTube, YouTube Community Guidelines enforcement, .
- Facebook, "Community Standards," .
- Kevin Bankston and Spandana Singh, "Facebook And Google Finally Take First Steps On Road To Transparency 麻豆果冻传媒 Content Moderation," TechDirt, April 26, 2018, .
- Facebook, Community Standards Enforcement Report, .
- Unless otherwise noted, all YouTube data in the charts was obtained from YouTube鈥檚 Community Guidelines enforcement report or from direct questions posed to the company, all Facebook data in the charts was obtained from Facebook鈥檚 Community Standards enforcement report, Facebook鈥檚 blog post on their Community Standards, Facebook鈥檚 response to an open letter penned by the Santa Clara Principles signatories or from direct questions posed to the company, and all Twitter data in the charts was obtained from Twitter鈥檚 Rules enforcement report or from direct questions posed to the company.
- YouTube鈥檚 report provides data on the following categories: spam, misleading or scams, nudity or sexual content, child safety, impersonation content, content that promotes violence and violent extremism, harassment and cyberbullying, hateful or abusive content, violent or graphic content, and harmful or dangerous content. Facebook鈥檚 initial report provided data on adult nudity and sexual activity, hate speech, terrorist propaganda, fake accounts, spam, violence and graphic content and since then the report has been expanded to include data on bullying and harassment and child nudity and sexual exploitation of children. Twitter鈥檚 report provides data on abuse, child sexual exploitation, hateful conduct, private information, sensitive media, and violent threats.
- Spandana Singh and Kevin Bankston, The Transparency Reporting Toolkit: Content Takedown Reporting, October 25, 2018, source.
- YouTube, YouTube Community.