scispace - formally typeset
Open AccessJournal ArticleDOI

A text abstraction summary model based on BERT word embedding and reinforcement learning

Reads0
Chats0
TLDR
A novel hybrid model of extractive-abstractive to combine BERT (Bidirectional Encoder Representations from Transformers) word embedding with reinforcement learning is proposed, which converts the human-written abstractive summaries to the ground truth labels.
Abstract
As a core task of natural language processing and information retrieval, automatic text summarization is widely applied in many fields. There are two existing methods for text summarization task at present: abstractive and extractive. On this basis we propose a novel hybrid model of extractive-abstractive to combine BERT (Bidirectional Encoder Representations from Transformers) word embedding with reinforcement learning. Firstly, we convert the human-written abstractive summaries to the ground truth labels. Secondly, we use BERT word embedding as text representation and pre-train two sub-models respectively. Finally, the extraction network and the abstraction network are bridged by reinforcement learning. To verify the performance of the model, we compare it with the current popular automatic text summary model on the CNN/Daily Mail dataset, and use the ROUGE (Recall-Oriented Understudy for Gisting Evaluation) metrics as the evaluation method. Extensive experimental results show that the accuracy of the model is improved obviously.

read more

Citations
More filters
Journal ArticleDOI

Deep Learning Based Abstractive Text Summarization: Approaches, Datasets, Evaluation Measures, and Challenges

TL;DR: It was determined that most abstractive text summarisation models faced challenges such as the unavailability of a golden token at testing time, out-of-vocabulary words, summary sentence repetition, inaccurate sentences, and fake facts.
Journal ArticleDOI

A survey on the techniques, applications, and performance of short text semantic similarity

TL;DR: This survey conducts a comprehensive and systematic analysis of semantic similarity, proposing three categories of semantic similarities: corpus‐based, knowledge-based, and deep learning (DL)‐based and evaluating state‐of‐the‐art DL methods on four common datasets which proved that DL‐based can better solve the challenges of the short text similarity, such as sparsity and complexity.
Journal ArticleDOI

Deep reinforcement and transfer learning for abstractive text summarization: A review

TL;DR: Automatic Text Summarization (ATS) is an important area in NLP as mentioned in this paper with the goal of shortening a long text into a more compact version by conveying the most important points in a readable form.
Journal ArticleDOI

Deep reinforcement and transfer learning for abstractive text summarization: A review

TL;DR: Automatic Text Summarization (ATS) is an important area in Natural Language Processing (NLP) with the goal of shortening a long text into a more compact version by conveying the most important points in a readable form as mentioned in this paper.
Journal ArticleDOI

T-BERTSum: Topic-Aware Text Summarization Based on BERT

TL;DR: Zhang et al. as mentioned in this paper proposed a topic-aware extractive and abstractive summarization model named T-BERTSum, based on Bidirectional Encoder Representations from Transformers (BERTs), which can simultaneously infer topics and generate summarization from social texts.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article

Attention is All you Need

TL;DR: This paper proposed a simple network architecture based solely on an attention mechanism, dispensing with recurrence and convolutions entirely and achieved state-of-the-art performance on English-to-French translation.
Proceedings ArticleDOI

Learning Phrase Representations using RNN Encoder--Decoder for Statistical Machine Translation

TL;DR: In this paper, the encoder and decoder of the RNN Encoder-Decoder model are jointly trained to maximize the conditional probability of a target sequence given a source sequence.
Posted Content

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: In this paper, the authors propose to use a soft-searching model to find the parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Posted Content

Sequence to Sequence Learning with Neural Networks

TL;DR: This paper presents a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure, and finds that reversing the order of the words in all source sentences improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
Related Papers (5)