Saturday, September 26, 2020
Richard Ashby Wilson and Molly K. Land, both of University of Connecticut School of Law, are publishing Hate Speech on Social Media: Towards a Context-Specific Content Moderation Policy in the Connecticut Law Review. Here is the abstract.
For all practical purposes, the decision of social media companies to prohibit hate speech on their platforms means that the longstanding debate in the United States about whether to limit hate speech in the public square has been resolved in favor of greater regulation. Nonetheless, revisiting these debates provides several insights essential for developing more empirically-based and narrowly tailored policies regarding online hate. First, a central issue in the hate speech debate is the extent to which hate speech contributes to violence. Those in favor of more robust regulation claim a connection to violence, while others dismiss these arguments as too tenuous to support regulation. The data generated by social media, however, now allow researchers to begin to empirically test whether there are visible, measurable harms resulting from hate speech. These data can assist in developing evidence-based policies to address the most significant harms of hate speech, while avoiding overbroad regulation that is inconsistent with international standards. Second, reexamining the U.S. debate about hate speech also reveals the serious missteps of social media policies that prohibit hate speech without regard to context. The policies that social media companies have developed attempt to define hate speech solely with respect to the content of the message. As the early advocates of limits on hate speech made clear, the meaning, force, and consequences of speech acts are deeply contextual, and it is impossible to understand the harms of hate speech without reference to local political realities and the power asymmetries between social groups. Regulation that is abstracted from this context will inevitably be overbroad. This Article revisits these hate speech debates and considers how they map onto the platform law of content moderation, where emerging evidence indicates a correlation between hate speech online, virulent nationalism, and violence against minorities and activists. It then concludes by developing specific recommendations to bring greater consideration of context into the policies and procedures of social media content moderation.
Download the article from SSRN at the link.