Cardiff University | Prifysgol Caerdydd ORCA
Online Research @ Cardiff 
WelshClear Cookie - decide language by browser settings

Hate speech, machine classification and statistical modelling of information flows on Twitter: interpretation and communication for policy decision making

Burnap, Peter ORCID: https://orcid.org/0000-0003-0396-633X and Williams, Matthew Leighton ORCID: https://orcid.org/0000-0003-2566-6063 2014. Hate speech, machine classification and statistical modelling of information flows on Twitter: interpretation and communication for policy decision making. Presented at: Internet, Policy & Politics, Oxford, UK, 26 September 2014.

[thumbnail of IPP2014-Burnap.pdf]
Preview
PDF - Accepted Post-Print Version
Download (388kB) | Preview

Abstract

In 2013, the murder of Drummer Lee Rigby in Woolwich, UK led to an extensive public social media reaction. Given the extreme terrorist motive and public nature of the actions it was feasible that the public response could include written expressions of hateful and antagonistic sentiment towards a particular race, ethnicity and religion, which can be interpreted as ‘hate speech’. This provided motivation to study the spread of hate speech on Twitter following such a widespread and emotive event. In this paper we present a supervised machine learning text classifier, trained and tested to distinguish between hateful and/or antagonistic responses with a focus on race, ethnicity or religion; and more general responses. We used human annotated data collected from Twitter in the immediate aftermath of Lee Rigby’s murder to train and test the classifier. As “Big Data” is a growing topic of study, and its use is in policy and decision making is being constantly debated at present, we discuss the use of supervised machine learning tools to classify a sample of “Big Data”, and how the results can be interpreted for use in policy and decision making. The results of the classifier are optimal using a combination of probabilistic, rule-based and spatial based classifiers with a voted ensemble meta-classifier. We achieve an overall F-measure of 0.95 using features derived from the content of each tweet, including syntactic dependencies between terms to recognise “othering” terms, incitement to respond with antagonistic action, and claims of well founded or justified discrimination against social groups. We then demonstrate how the results of the classifier can be robustly utilized in a statistical model used to forecast the likely spread of hate speech in a sample of Twitter data.

Item Type: Conference or Workshop Item (Paper)
Date Type: Completion
Status: Published
Schools: Computer Science & Informatics
Social Sciences (Includes Criminology and Education)
Date of First Compliant Deposit: 30 March 2016
Last Modified: 10 Nov 2023 14:39
URI: https://orca.cardiff.ac.uk/id/eprint/65227

Citation Data

Actions (repository staff only)

Edit Item Edit Item

Downloads

Downloads per month over past year

View more statistics