Reporting bias and knowledge acquisition
Jonathan Gordon,Benjamin Van Durme +1 more
- pp 25-30
Reads0
Chats0
TLDR
This paper questions the idea that the frequency with which people write about actions, outcomes, or properties is a reflection of real-world frequencies or the degree to which a property is characteristic of a class of individuals.Abstract:
Much work in knowledge extraction from text tacitly assumes that the frequency with which people write about actions, outcomes, or properties is a reflection of real-world frequencies or the degree to which a property is characteristic of a class of individuals. In this paper, we question this idea, examining the phenomenon of reporting bias and the challenge it poses for knowledge extraction. We conclude with discussion of approaches to learning commonsense knowledge from text despite this distortion.read more
Citations
More filters
Proceedings ArticleDOI
COMET: Commonsense Transformers for Automatic Knowledge Graph Construction
TL;DR: This investigation reveals promising results when implicit knowledge from deep pre-trained language models is transferred to generate explicit knowledge in commonsense knowledge graphs, and suggests that using generative commonsense models for automatic commonsense KB completion could soon be a plausible alternative to extractive methods.
Book ChapterDOI
Women Also Snowboard: Overcoming Bias in Captioning Models
TL;DR: The authors proposed a new Equalizer model that encourages equal gender probability when gender evidence is occluded in a scene and confident predictions when gender evidences is present, which can be added to any description model in order to mitigate impacts of unwanted bias in a description dataset.
Proceedings ArticleDOI
Social IQa: Commonsense Reasoning about Social Interactions
TL;DR: Social IQa as mentioned in this paper is a large-scale benchmark for commonsense reasoning about social situations, which contains 38,000 multiple choice questions for probing emotional and social intelligence in a variety of everyday situations.
Posted Content
WINOGRANDE: An Adversarial Winograd Schema Challenge at Scale
TL;DR: The authors introduced WinoGrande, a large-scale dataset of 44k problems, inspired by the original Winograd Schema Challenge (WSC) design, but adjusted to improve both the scale and the hardness of the dataset.
Posted Content
HellaSwag: Can a Machine Really Finish Your Sentence?.
TL;DR: HellaSwag as discussed by the authors ) is a commonsense NLP dataset where a series of discriminators iteratively select an adversarial set of machine-generated wrong answers, and the key insight is to scale up the length and complexity of the dataset examples towards a critical 'Goldilocks' zone where generated text is ridiculous to humans, yet often misclassified by state-of-the-art models.
References
More filters
Journal ArticleDOI
WordNet : an electronic lexical database
TL;DR: The lexical database: nouns in WordNet, Katherine J. Miller a semantic network of English verbs, and applications of WordNet: building semantic concordances are presented.
Book ChapterDOI
Logic and Conversation
TL;DR: For instance, Grice was interested in Quine's logical approach to language, although he differed from Quine over certain specific specific questions, such as the viability of the distinction between analytic and synthetic statements.
Book
Computational analysis of present-day American English
Henry Kučera,W. Nelson Francis,W. F. Twaddell,Mary Lois Marckworth,Laura M. Bell,John Bissell Carroll +5 more
Proceedings Article
Generating Typed Dependency Parses from Phrase Structure Parses
TL;DR: A system for extracting typed dependency parses of English sentences from phrase structure parses that captures inherent relations occurring in corpus texts that can be critical in real-world applications is described.