scispace - formally typeset
Open AccessProceedings ArticleDOI

Abstractive Document Summarization with a Graph-Based Attentional Neural Model

Jiwei Tan, +2 more
- Vol. 1, pp 1171-1181
Reads0
Chats0
TLDR
A novel graph-based attention mechanism in the sequence-to-sequence framework to address the saliency factor of summarization, which has been overlooked by prior works and is competitive with state-of-the-art extractive methods.
Abstract
ive summarization is the ultimate goal of document summarization research, but previously it is less investigated due to the immaturity of text generation techniques. Recently impressive progress has been made to abstractive sentence summarization using neural models. Unfortunately, attempts on abstractive document summarization are still in a primitive stage, and the evaluation results are worse than extractive methods on benchmark datasets. In this paper, we review the difficulties of neural abstractive document summarization, and propose a novel graph-based attention mechanism in the sequence-to-sequence framework. The intuition is to address the saliency factor of summarization, which has been overlooked by prior works. Experimental results demonstrate our model is able to achieve considerable improvement over previous neural abstractive models. The data-driven neural abstractive method is also competitive with state-of-the-art extractive methods.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Bottom-Up Abstractive Summarization

TL;DR: This work explores the use of data-efficient content selectors to over-determine phrases in a source document that should be part of the summary, and shows that this approach improves the ability to compress text, while still generating fluent summaries.
Proceedings ArticleDOI

Don't Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization

TL;DR: This paper proposed a novel abstractive model which is conditioned on the article's topics and based entirely on convolutional neural networks and demonstrated experimentally that this architecture captures long-range dependencies in a document and recognizes pertinent content, outperforming an oracle extractive system and state-of-theart abstractive approaches when evaluated automatically and by humans.
Proceedings ArticleDOI

Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting

TL;DR: The authors proposed a sentence-level policy gradient method to bridge the non-differentiable computation between these two neural networks in a hierarchical way, which achieved state-of-the-art performance on the CNN/Daily Mail dataset.
Posted Content

Don't Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization

TL;DR: A novel abstractive model is proposed which is conditioned on the article’s topics and based entirely on convolutional neural networks, outperforming an oracle extractive system and state-of-the-art abstractive approaches when evaluated automatically and by humans.
Proceedings ArticleDOI

Ranking Sentences for Extractive Summarization with Reinforcement Learning

TL;DR: The authors conceptualized extractive summarization as a sentence ranking task and proposed a novel training algorithm which globally optimizes the ROUGE evaluation metric through a reinforcement learning objective, which outperforms state-of-the-art extractive and abstractive systems when evaluated automatically and by humans.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Proceedings Article

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: It is conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and it is proposed to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Proceedings Article

The PageRank Citation Ranking : Bringing Order to the Web

TL;DR: This paper describes PageRank, a mathod for rating Web pages objectively and mechanically, effectively measuring the human interest and attention devoted to them, and shows how to efficiently compute PageRank for large numbers of pages.
Related Papers (5)