• Corpus ID: 237048417

Hatemoji: A Test Suite and Adversarially-Generated Dataset for Benchmarking and Detecting Emoji-based Hate

  title={Hatemoji: A Test Suite and Adversarially-Generated Dataset for Benchmarking and Detecting Emoji-based Hate},
  author={Hannah Rose Kirk and Bertram Vidgen and Paul R{\"o}ttger and Tristan Thrush and Scott A. Hale},
Detecting online hate is a complex task, and low-performing models have harmful conse-quences when used for sensitive applications such as content moderation. Emoji-based hate is an emerging challenge for automated detection. We present H ATEMOJI C HECK , a test suite of 3 , 930 short-form statements that al-lows us to evaluate performance on hateful language expressed with emoji. Using the test suite, we expose weaknesses in existing hate detection models. To address these weaknesses, we… 
A New Generation of Perspective API: Efficient Multilingual Character-level Transformers
This paper presents the fundamentals behind the next version of the Perspective API from Google Jigsaw, and presents a single multilingual token-free Charformer model that is applicable across a range of languages, domains, and tasks.
Handling and Presenting Harmful Text
Practical advice is provided on how textual harms should be handled, presented, and discussed and H ARM C HECK, a resource for reflecting on research into textual harms, is introduced to encourage ethical, responsible, and respectful research in the NLP community.
Dynatask: A Framework for Creating Dynamic AI Benchmark Tasks
We introduce Dynatask: an open source system for setting up custom NLP tasks that aims to greatly lower the technical knowledge and effort required for hosting and evaluating state-of-the-art NLP
Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality
A novel task and dataset for evaluating the ability of vision and language models to conduct visio-linguistic compositional reasoning, which is called Winoground and aims for it to serve as a useful evaluation set for advancing the state of the art and driv-ing further progress in the industry.
Combining Data-driven Supervision with Human-in-the-loop Feedback for Entity Resolution
This case study discusses the human-in-the-loop enabled, data-centric solution to closing the training-production performance divergence, and concludes with takeaways that apply toData-centric learning at large.


Free-Marginal Multirater Kappa (multirater K[free]): An Alternative to Fleiss' Fixed-Marginal Multirater Kappa.
Free-marginal Multirater Kappa (multirater κfree), like its birater free- Marginal counterparts (PAB AK, S, RE, and κm,) is appropriate for the typical agreement study, in which raters’ distributions of cases into categories are not restricted.
Exploiting Emojis for Abusive Language Detection
It is shown that emojis can be exploited for classifying mentions of ambiguous words, such as “fuck” and “bitch”, into generally abusive and just profane usages and generates a lexicon that offers the same performance in cross-domain classification of abusive microposts as the most advanced lexicon induction method.
UN Strategy and Plan of Action on Hate Speech
  • Technical report.
  • 2019
Introducing CAD: the Contextual Abuse Dataset
A new dataset of primarily English Reddit entries which addresses several limitations of prior work, and contains six conceptually distinct primary categories as well as secondary categories, and uses an expert-driven group-adjudication process for high quality annotations.
Annotating Emoticons and Emojis in a German-Danish Social Media Corpus for Hate Speech Research
  • 2020
Beyond Accuracy: Behavioral Testing of NLP Models with CheckList
Although measuring held-out accuracy has been the primary approach to evaluate generalization, it often overestimates the performance of NLP models, while alternative approaches for evaluating models
Evaluating Models’ Local Decision Boundaries via Contrast Sets
A more rigorous annotation paradigm for NLP that helps to close systematic gaps in the test data, and recommends that the dataset authors manually perturb the test instances in small but meaningful ways that (typically) change the gold label, creating contrast sets.
Racist comments after Euro 2020: Saka, Sancho and Rashford racially abused online after England defeat
  • 2020
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
A new language representation model, BERT, designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
2018. BERT: Pre-training
  • 2018