scispace - formally typeset
Open AccessProceedings ArticleDOI

Reporting bias and knowledge acquisition

Reads0
Chats0
TLDR
This paper questions the idea that the frequency with which people write about actions, outcomes, or properties is a reflection of real-world frequencies or the degree to which a property is characteristic of a class of individuals.
Abstract
Much work in knowledge extraction from text tacitly assumes that the frequency with which people write about actions, outcomes, or properties is a reflection of real-world frequencies or the degree to which a property is characteristic of a class of individuals. In this paper, we question this idea, examining the phenomenon of reporting bias and the challenge it poses for knowledge extraction. We conclude with discussion of approaches to learning commonsense knowledge from text despite this distortion.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

COMET: Commonsense Transformers for Automatic Knowledge Graph Construction

TL;DR: This investigation reveals promising results when implicit knowledge from deep pre-trained language models is transferred to generate explicit knowledge in commonsense knowledge graphs, and suggests that using generative commonsense models for automatic commonsense KB completion could soon be a plausible alternative to extractive methods.
Book ChapterDOI

Women Also Snowboard: Overcoming Bias in Captioning Models

TL;DR: The authors proposed a new Equalizer model that encourages equal gender probability when gender evidence is occluded in a scene and confident predictions when gender evidences is present, which can be added to any description model in order to mitigate impacts of unwanted bias in a description dataset.
Proceedings ArticleDOI

Social IQa: Commonsense Reasoning about Social Interactions

TL;DR: Social IQa as mentioned in this paper is a large-scale benchmark for commonsense reasoning about social situations, which contains 38,000 multiple choice questions for probing emotional and social intelligence in a variety of everyday situations.
Posted Content

WINOGRANDE: An Adversarial Winograd Schema Challenge at Scale

TL;DR: The authors introduced WinoGrande, a large-scale dataset of 44k problems, inspired by the original Winograd Schema Challenge (WSC) design, but adjusted to improve both the scale and the hardness of the dataset.
Posted Content

HellaSwag: Can a Machine Really Finish Your Sentence?.

TL;DR: HellaSwag as discussed by the authors ) is a commonsense NLP dataset where a series of discriminators iteratively select an adversarial set of machine-generated wrong answers, and the key insight is to scale up the length and complexity of the dataset examples towards a critical 'Goldilocks' zone where generated text is ridiculous to humans, yet often misclassified by state-of-the-art models.
References
More filters
Book ChapterDOI

Logic and conversation

H. P. Grice
- 12 Dec 1975 - 
Journal ArticleDOI

WordNet : an electronic lexical database

Christiane Fellbaum
- 01 Sep 2000 - 
TL;DR: The lexical database: nouns in WordNet, Katherine J. Miller a semantic network of English verbs, and applications of WordNet: building semantic concordances are presented.
Book ChapterDOI

Logic and Conversation

TL;DR: For instance, Grice was interested in Quine's logical approach to language, although he differed from Quine over certain specific specific questions, such as the viability of the distinction between analytic and synthetic statements.
Proceedings Article

Generating Typed Dependency Parses from Phrase Structure Parses

TL;DR: A system for extracting typed dependency parses of English sentences from phrase structure parses that captures inherent relations occurring in corpus texts that can be critical in real-world applications is described.
Related Papers (5)