scispace - formally typeset
Open AccessPosted Content

Automated News Summarization Using Transformers

TLDR
This article presented a comprehensive comparison of a few transformer architecture based pre-trained models for text summarization and human generated summaries for evaluating and comparing the summaries generated by machine learning models.
Abstract
The amount of text data available online is increasing at a very fast pace hence text summarization has become essential. Most of the modern recommender and text classification systems require going through a huge amount of data. Manually generating precise and fluent summaries of lengthy articles is a very tiresome and time-consuming task. Hence generating automated summaries for the data and using it to train machine learning models will make these models space and time-efficient. Extractive summarization and abstractive summarization are two separate methods of generating summaries. The extractive technique identifies the relevant sentences from the original document and extracts only those from the text. Whereas in abstractive summarization techniques, the summary is generated after interpreting the original text, hence making it more complicated. In this paper, we will be presenting a comprehensive comparison of a few transformer architecture based pre-trained models for text summarization. For analysis and comparison, we have used the BBC news dataset that contains text data that can be used for summarization and human generated summaries for evaluating and comparing the summaries generated by machine learning models.

read more

Citations
More filters
Journal ArticleDOI

Abstractive Text Summarization of Hindi Corpus Using Transformer Encoder-Decoder Model

TL;DR: In this paper , the Transformer encoder-decoder architecture is employed to extract contextual dependencies and yield better semantic representations for Hindi, engendering an abstractive summary. But, this method is limited to Hindi news.
Proceedings ArticleDOI

CNA: A Dataset for Parsing Discourse Structure on Chinese News Articles

TL;DR: Li et al. as mentioned in this paper presented CNA, a Chinese news corpus containing 1155 news articles annotated by human experts, which covers four domains and four news media sources, and proposed a document-level neural network model with multiple sentence features.
Proceedings ArticleDOI

Text Summarization using Transformer Model

TL;DR: This paper proposed a text summarization method based on the Text-to-Text Transfer Transformer (T5) model, which achieved an average of ROUGE1, RougeGE2, and RougeGEL scores of 45.62, 25.58, and 36.53, respectively.
Proceedings ArticleDOI

The application of artificial intelligence on different types of literature reviews - A comparative study

TL;DR: In this paper , a comparative study of systematic and semi-systematic literature reviews is performed to determine the potential of AI applications in both types of literature review processes, and a new tool integrating various AI applications along the research process that improve the speed, quality, and cost-efficiency of the overall research process.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Posted Content

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

TL;DR: A new language representation model, BERT, designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Posted Content

Attention Is All You Need

TL;DR: A new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely is proposed, which generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.
Posted Content

Generating Sequences With Recurrent Neural Networks

TL;DR: This paper shows how Long Short-term Memory recurrent neural networks can be used to generate complex sequences with long-range structure, simply by predicting one data point at a time.
Journal ArticleDOI

A Comprehensive Survey on Transfer Learning

TL;DR: Transfer learning aims to improve the performance of target learners on target domains by transferring the knowledge contained in different but related source domains as discussed by the authors, in which the dependence on a large number of target-domain data can be reduced for constructing target learners.