Resources and benchmark corpora for hate speech detection: a systematic review

F Poletto, V Basile, M Sanguinetti, C Bosco… - Language Resources …, 2021 - Springer
Hate Speech in social media is a complex phenomenon, whose detection has recently
gained significant traction in the Natural Language Processing community, as attested by …

The psychological well-being of content moderators: the emotional labor of commercial moderation and avenues for improving support

M Steiger, TJ Bharucha, S Venkatagiri… - Proceedings of the …, 2021 - dl.acm.org
An estimated 100,000 people work today as commercial content moderators. These
moderators are often exposed to disturbing content, which can lead to lasting psychological …

On the opportunities and risks of foundation models

R Bommasani, DA Hudson, E Adeli, R Altman… - arxiv preprint arxiv …, 2021 - arxiv.org
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …

Dealing with disagreements: Looking beyond the majority vote in subjective annotations

AM Davani, M Díaz, V Prabhakaran - Transactions of the Association …, 2022 - direct.mit.edu
Majority voting and averaging are common approaches used to resolve annotator
disagreements and derive single ground truth labels from multiple annotations. However …

Hatebert: Retraining bert for abusive language detection in english

T Caselli, V Basile, J Mitrović, M Granitzer - arxiv preprint arxiv …, 2020 - arxiv.org
In this paper, we introduce HateBERT, a re-trained BERT model for abusive language
detection in English. The model was trained on RAL-E, a large-scale dataset of Reddit …

Latent hatred: A benchmark for understanding implicit hate speech

M ElSherief, C Ziems, D Muchlinski, V Anupindi… - arxiv preprint arxiv …, 2021 - arxiv.org
Hate speech has grown significantly on social media, causing serious consequences for
victims of all demographics. Despite much attention being paid to characterize and detect …

Social biases in NLP models as barriers for persons with disabilities

B Hutchinson, V Prabhakaran, E Denton… - arxiv preprint arxiv …, 2020 - arxiv.org
Building equitable and inclusive NLP technologies demands consideration of whether and
how social attitudes are represented in ML models. In particular, representations encoded in …

Directions in abusive language training data, a systematic review: Garbage in, garbage out

B Vidgen, L Derczynski - Plos one, 2020 - journals.plos.org
Data-driven and machine learning based approaches for detecting, categorising and
measuring abusive content such as hate speech and harassment have gained traction due …

Misogyny detection in twitter: a multilingual and cross-domain study

EW Pamungkas, V Basile, V Patti - Information processing & management, 2020 - Elsevier
The freedom of expression given by social media has a dark side: the growing proliferation
of abusive contents on these platforms. Misogynistic speech is a kind of abusive language …

A survey of race, racism, and anti-racism in NLP

A Field, SL Blodgett, Z Waseem, Y Tsvetkov - arxiv preprint arxiv …, 2021 - arxiv.org
Despite inextricable ties between race and language, little work has considered race in NLP
research and development. In this work, we survey 79 papers from the ACL anthology that …