scispace - formally typeset
Open AccessPosted Content

The Hateful Memes Challenge: Detecting Hate Speech in Multimodal Memes

Reads0
Chats0
TLDR
The authors proposed a new challenge set for multimodal classification, focusing on detecting hate speech in multi-modal memes, where difficult examples are added to the dataset to make it hard to rely on unimodal signals.
Abstract
This work proposes a new challenge set for multimodal classification, focusing on detecting hate speech in multimodal memes. It is constructed such that unimodal models struggle and only multimodal models can succeed: difficult examples ("benign confounders") are added to the dataset to make it hard to rely on unimodal signals. The task requires subtle reasoning, yet is straightforward to evaluate as a binary classification problem. We provide baseline performance numbers for unimodal models, as well as for multimodal models with various degrees of sophistication. We find that state-of-the-art methods perform poorly compared to humans (64.73% vs. 84.7% accuracy), illustrating the difficulty of the task and highlighting the challenge that this important problem poses to the community.

read more

Citations
More filters
Posted Content

Learning Transferable Visual Models From Natural Language Supervision

TL;DR: In this article, a pre-training task of predicting which caption goes with which image is used to learn SOTA image representations from scratch on a dataset of 400 million (image, text) pairs collected from the internet.
Proceedings ArticleDOI

Overview of the HASOC track at FIRE 2019: Hate Speech and Offensive Content Identification in Indo-European Languages

TL;DR: The HASOC track intends to stimulate development in Hate Speech for Hindi, German and English by identifying Hate Speech in Social Media using LSTM networks processing word embedding input.
Journal ArticleDOI

Directions in abusive language training data, a systematic review: Garbage in, garbage out.

TL;DR: This paper systematically reviews abusive language dataset creation and content in conjunction with an open website for cataloguing abusive language data leading to a synthesis providing evidence-based recommendations for practitioners working with this complex and highly diverse data.
Proceedings ArticleDOI

TextOCR: Towards large-scale end-to-end reasoning for arbitrary-shaped scene text

TL;DR: TextOCR as discussed by the authors is an arbitrary-shaped scene text detection and recognition with 900k annotated words collected on real images from TextVQA dataset, which can do scene text based reasoning on an image in an end-to-end fashion.
Posted Content

Tackling Online Abuse: A Survey of Automated Abuse Detection Methods

TL;DR: A comprehensive survey of the methods that have been proposed to date for automated abuse detection in the field of natural language processing (NLP), providing a platform for further development of this area.
References
More filters
Proceedings Article

Equal But Not The Same: Understanding the Implicit Relationship Between Persuasive Images and Text

TL;DR: A variety of features that capture the creativity of images and the specificity or ambiguity of text, as well as methods that analyze the semantics within and across channels are developed.
Proceedings ArticleDOI

An empirical study on the effectiveness of images in Multimodal Neural Machine Translation

TL;DR: This article used an attention mechanism to focus on different parts of the source sentence to gather the most useful information before outputting its target word and achieved state-of-the-art results on the Multi30k data set.
Proceedings Article

Grounded Textual Entailment.

TL;DR: The authors compare blind and visual-augmented models of textual entailment and show that visual information is beneficial, but also conduct an in-depth error analysis that reveals that current multimodal models are not performing "grounding" in an optimal fashion.
Posted Content

Grounded Textual Entailment

TL;DR: This paper argues for a visually-grounded version of the Textual Entailment task, and asks whether models can perform better if, in addition to P and H, there is also an image (corresponding to the relevant “world” or “situation”).
Posted Content

Equal But Not The Same: Understanding the Implicit Relationship Between Persuasive Images and Text

TL;DR: In this article, a dataset of advertisement interpretations for whether the image and slogan in the same visual advertisement form a parallel (conveying the same message without literally saying the same thing) or non-parallel relationship is collected, with the help of workers recruited on Amazon Mechanical Turk.
Related Papers (5)