scispace - formally typeset
Open AccessProceedings ArticleDOI

Learning to Ask: Neural Question Generation for Reading Comprehension

Xinya Du, +2 more
- Vol. 1, pp 1342-1352
TLDR
This paper proposed an attention-based sequence learning model for question generation from text passages in reading comprehension, which is trainable end-to-end via sequence-tosequence learning and significantly outperforms the state-of-the-art rule-based system.
Abstract
We study automatic question generation for sentences from text passages in reading comprehension. We introduce an attention-based sequence learning model for the task and investigate the effect of encoding sentence- vs. paragraph-level information. In contrast to all previous work, our model does not rely on hand-crafted rules or a sophisticated NLP pipeline; it is instead trainable end-to-end via sequence-to-sequence learning. Automatic evaluation results show that our system significantly outperforms the state-of-the-art rule-based system. In human evaluations, questions generated by our system are also rated as being more natural (i.e.,, grammaticality, fluency) and as more difficult to answer (in terms of syntactic and lexical divergence from the original text and reasoning needed to answer).

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted Content

Unified Language Model Pre-training for Natural Language Understanding and Generation

TL;DR: A new Unified pre-trained Language Model (UniLM) that can be fine-tuned for both natural language understanding and generation tasks that compares favorably with BERT on the GLUE benchmark, and the SQuAD 2.0 and CoQA question answering tasks.
Proceedings ArticleDOI

Paragraph-level Neural Question Generation with Maxout Pointer and Gated Self-attention Networks

TL;DR: A maxout pointer mechanism with gated self-attention encoder to address the challenges of processing long text inputs for question generation, which outperforms previous approaches with either sentence-level or paragraph-level inputs.
Proceedings ArticleDOI

ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training.

TL;DR: A new sequence-to-sequence pre-training model called ProphetNet is presented, which introduces a novel self-supervised objective named future n-gram prediction and the proposed n-stream self-attention mechanism that predicts the next n tokens simultaneously based on previous context tokens at each time step.
Proceedings ArticleDOI

Question Generation for Question Answering

TL;DR: Experimental results show that, by using generated questions as an extra signal, significant QA improvement can be achieved.
References
More filters
Proceedings Article

Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks

TL;DR: This paper proposed a set of proxy tasks that evaluate reading comprehension via question answering, such as chaining facts, simple induction, deduction and many more, which are designed to be prerequisites for any system that aims to be capable of conversing with a human.
Proceedings Article

MCTest: A Challenge Dataset for the Open-Domain Machine Comprehension of Text

TL;DR: MCTest is presented, a freely available set of stories and associated questions intended for research on the machine comprehension of text that requires machines to answer multiple-choice reading comprehension questions about fictional stories, directly tackling the high-level goal of open-domain machine comprehension.
Proceedings ArticleDOI

Addressing the Rare Word Problem in Neural Machine Translation

TL;DR: This paper proposed and implemented an effective technique to address the problem of out-of-vocabulary (OOV) word translation in NMT, which trains an NMT system on data that is augmented by the output of a word alignment algorithm, and then uses this information in a post-processing step that translates every OOV word using a dictionary.
Proceedings ArticleDOI

Summarizing Source Code using a Neural Attention Model

TL;DR: This paper presents the first completely datadriven approach for generating high level summaries of source code, which uses Long Short Term Memory (LSTM) networks with attention to produce sentences that describe C# code snippets and SQL queries.
Proceedings Article

Scalable Modified Kneser-Ney Language Model Estimation

TL;DR: An efficient algorithm to estimate large modified Kneser-Ney models including interpolation using Streaming and sorting enables the algorithm to scale to much larger models by using a fixed amount of RAM and variable amount of disk.