scispace - formally typeset
Open AccessProceedings ArticleDOI

Learning to Ask: Neural Question Generation for Reading Comprehension

Xinya Du, +2 more
- Vol. 1, pp 1342-1352
TLDR
This paper proposed an attention-based sequence learning model for question generation from text passages in reading comprehension, which is trainable end-to-end via sequence-tosequence learning and significantly outperforms the state-of-the-art rule-based system.
Abstract
We study automatic question generation for sentences from text passages in reading comprehension. We introduce an attention-based sequence learning model for the task and investigate the effect of encoding sentence- vs. paragraph-level information. In contrast to all previous work, our model does not rely on hand-crafted rules or a sophisticated NLP pipeline; it is instead trainable end-to-end via sequence-to-sequence learning. Automatic evaluation results show that our system significantly outperforms the state-of-the-art rule-based system. In human evaluations, questions generated by our system are also rated as being more natural (i.e.,, grammaticality, fluency) and as more difficult to answer (in terms of syntactic and lexical divergence from the original text and reasoning needed to answer).

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings Article

Contrastive Domain Adaptation for Question Answering using Limited Text Corpora

TL;DR: In contrastive domain adaptation for QA (CAQA) as discussed by the authors, the authors combine question generation and domain-invariant learning to answer out-of-domain questions in settings with limited text corpora.
Book ChapterDOI

Review of Parameters, Approaches and Challenges in Reading Comprehension Systems

TL;DR: This paper presented an exhaustive study of the various approaches used by different authors for feature extraction in the domain of reading comprehension, and discussed some of the open challenges in modeling the comprehension systems.
Proceedings ArticleDOI

Graph and Question Interaction Aware Graph2Seq Model for Knowledge Base Question Generation

TL;DR: Zhang et al. as discussed by the authors proposed a graph and question interaction enhanced Graph2Seq model, in which they designed an encoder-decoder parallel enhancement mechanism and applied the knowledge distillation for both intermediate representation and prediction distribution to employ the knowledge of the target question into the graph representation.
Journal ArticleDOI

FFCI: A Framework for Interpretable Automatic Evaluation of Summarization

TL;DR: This article propose a framework for fine-grained summarization evaluation that comprises four elements: faithfulness (degree of factual consistency with the source), focus (precision of summary content relative to the reference), coverage, and inter-sentential coherence (document fluency between adjacent sentences).
Posted Content

Improving Controllability of Educational Question Generation by Keyword Provision.

TL;DR: In this paper, the authors report a state-of-the-art exam-like QG model by advancing the current best model from 11.96 to 20.19 (in terms of BLEU 4 score).
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Proceedings ArticleDOI

Bleu: a Method for Automatic Evaluation of Machine Translation

TL;DR: This paper proposed a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run.
Proceedings Article

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: It is conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and it is proposed to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Proceedings ArticleDOI

Learning Phrase Representations using RNN Encoder--Decoder for Statistical Machine Translation

TL;DR: In this paper, the encoder and decoder of the RNN Encoder-Decoder model are jointly trained to maximize the conditional probability of a target sequence given a source sequence.