scispace - formally typeset
Open AccessProceedings ArticleDOI

Dependency tree based sentence compression

Katja Filippova, +1 more
- pp 25-32
TLDR
A novel unsupervised method for sentence compression which relies on a dependency tree representation and shortens sentences by removing subtrees and it is demonstrated that the choice of the parser affects the performance of the system.
Abstract
We present a novel unsupervised method for sentence compression which relies on a dependency tree representation and shortens sentences by removing subtrees. An automatic evaluation shows that our method obtains result comparable or superior to the state of the art. We demonstrate that the choice of the parser affects the performance of the system. We also apply the method to German and report the results of an evaluation with humans.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Optimizing Statistical Machine Translation for Text Simplification

TL;DR: This work is the first to design automatic metrics that are effective for tuning and evaluating simplification systems, which will facilitate iterative development for this task.
Proceedings Article

A Monolingual Tree-based Translation Model for Sentence Simplification

TL;DR: A Tree-based Simplification Model (TSM) is proposed, which, to the knowledge, is the first statistical simplification model covering splitting, dropping, reordering and substitution integrally.
Journal ArticleDOI

Problems in Current Text Simplification Research: New Data Can Help

TL;DR: This opinion paper argues that focusing on Wikipedia limits simplification research, and introduces a new simplification dataset that is a significant improvement over Simple Wikipedia, and presents a novel quantitative-comparative approach to study the quality of simplification data resources.
Proceedings Article

Multi-Sentence Compression: Finding Shortest Paths in Word Graphs

TL;DR: Despite its simplicity, the proposed multi-sentence compression method is capable of generating grammatical and informative summaries as its experiments with English and Spanish data demonstrate.
Proceedings ArticleDOI

Controlling Output Length in Neural Encoder-Decoders

TL;DR: This paper proposed methods for controlling the output sequence length for neural encoder-decoder models: two decoding-based and two learning-based methods, which have the capability to control length without degrading summary quality in a summarization task.
References
More filters
Proceedings ArticleDOI

Accurate Unlexicalized Parsing

TL;DR: It is demonstrated that an unlexicalized PCFG can parse much more accurately than previously shown, by making use of simple, linguistically motivated state splits, which break down false independence assumptions latent in a vanilla treebank grammar.
Proceedings Article

Generating Typed Dependency Parses from Phrase Structure Parses

TL;DR: A system for extracting typed dependency parses of English sentences from phrase structure parses that captures inherent relations occurring in corpus texts that can be critical in real-world applications is described.
Journal ArticleDOI

Summarization beyond sentence extraction: a probabilistic approach to sentence compression

TL;DR: This paper focuses on sentence compression, a simpler version of this larger challenge, and aims to achieve two goals simultaneously: the authors' compressions should be grammatical, and they should retain the most important pieces of information.
Proceedings ArticleDOI

The Second Release of the RASP System

TL;DR: The new release of the RASP (robust accurate statistical parsing) system, designed for syntactic annotation of free text, is described, including a revised and more semantically-motivated output representation, an enhanced grammar and part-of-speech tagger lexicon, and a more flexible and semi-supervised training method for the structural parse ranking model.
Journal ArticleDOI

Global inference for sentence compression an integer linear programming approach

TL;DR: This work shows how previous formulations of sentence compression can be recast as ILPs and extend these models with novel global constraints to infer globally optimal compressions in the presence of linguistically motivated constraints.
Related Papers (5)