scispace - formally typeset
Open AccessProceedings Article

Universal Stanford dependencies: A cross-linguistic typology

Reads0
Chats0
TLDR
This work proposes a two-layered taxonomy: a set of broadly attested universal grammatical relations, to which language-specific relations can be added, and a lexicalist stance of the Stanford Dependencies, which leads to a particular, partially new treatment of compounding, prepositions, and morphology.
Abstract
Revisiting the now de facto standard Stanford dependency representation, we propose an improved taxonomy to capture grammatical relations across languages, including morphologically rich ones. We suggest a two-layered taxonomy: a set of broadly attested universal grammatical relations, to which language-specific relations can be added. We emphasize the lexicalist stance of the Stanford Dependencies, which leads to a particular, partially new treatment of compounding, prepositions, and morphology. We show how existing dependency schemes for several languages map onto the universal taxonomy proposed here and close with consideration of practical implications of dependency representation choices for NLP applications, in particular parsing.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings Article

Universal Dependencies v1: A Multilingual Treebank Collection

TL;DR: This paper describes v1 of the universal guidelines, the underlying design principles, and the currently available treebanks for 33 languages, as well as highlighting the needs for sound comparative evaluation and cross-lingual learning experiments.
Journal ArticleDOI

CheXpert: A large chest radiograph dataset with uncertainty labels and expert comparison

TL;DR: CheXpert as discussed by the authors is a large dataset of chest radiographs of 65,240 patients annotated by 3 board-certified radiologists with 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation.
Book ChapterDOI

SPICE: Semantic Propositional Image Caption Evaluation

TL;DR: This paper proposed a new automated caption evaluation metric defined over scene graphs coined SPICE, which captures human judgments over model-generated captions better than other automatic metrics (e.g., system-level correlation of 0.88 with human judgments on the MS COCO dataset, versus 0.43 for CIDEr and 0.53 for METEOR).
Posted Content

SPICE: Semantic Propositional Image Caption Evaluation

TL;DR: It is hypothesized that semantic propositional content is an important component of human caption evaluation, and a new automated caption evaluation metric defined over scene graphs coined SPICE is proposed, which can answer questions such as which caption-generator best understands colors?
Posted Content

CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison

TL;DR: CheXpert as discussed by the authors is a large dataset of chest radiographs of 65,240 patients annotated by 3 board-certified radiologists with 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation and different approaches to using the uncertainty labels for training convolutional neural networks that output the probability of these observations given the available frontal and lateral radiographs.
References
More filters
Proceedings ArticleDOI

Accurate Unlexicalized Parsing

TL;DR: It is demonstrated that an unlexicalized PCFG can parse much more accurately than previously shown, by making use of simple, linguistically motivated state splits, which break down false independence assumptions latent in a vanilla treebank grammar.
Proceedings Article

Generating Typed Dependency Parses from Phrase Structure Parses

TL;DR: A system for extracting typed dependency parses of English sentences from phrase structure parses that captures inherent relations occurring in corpus texts that can be critical in real-world applications is described.
Journal ArticleDOI

The Proposition Bank: An Annotated Corpus of Semantic Roles

TL;DR: An automatic system for semantic role tagging trained on the corpus is described and the effect on its performance of various types of information is discussed, including a comparison of full syntactic parsing with a flat representation and the contribution of the empty trace categories of the treebank.
Journal ArticleDOI

Head-Driven Statistical Models for Natural Language Parsing

TL;DR: Three statistical models for natural language parsing are described, leading to approaches in which a parse tree is represented as the sequence of decisions corresponding to a head-centered, top-down derivation of the tree.
Related Papers (5)