scispace - formally typeset
Open AccessProceedings ArticleDOI

A Hierarchical Network for Abstractive Meeting Summarization with Cross-Domain Pretraining

Reads0
Chats0
TLDR
A novel abstractive summary network that adapts to the meeting scenario is proposed with a hierarchical structure to accommodate long meeting transcripts and a role vector to depict the difference among speakers.
Abstract
With the abundance of automatic meeting transcripts, meeting summarization is of great interest to both participants and other parties. Traditional methods of summarizing meetings depend on complex multi-step pipelines that make joint optimization intractable. Meanwhile, there are a handful of deep neural models for text summarization and dialogue systems. However, the semantic structure and styles of meeting transcripts are quite different from articles and conversations. In this paper, we propose a novel abstractive summary network that adapts to the meeting scenario. We design a hierarchical structure to accommodate long meeting transcripts and a role vector to depict the difference among speakers. Furthermore, due to the inadequacy of meeting summary data, we pretrain the model on large-scale news summary data. Empirical results show that our model outperforms previous approaches in both automatic metrics and human evaluation. For example, on ICSI dataset, the ROUGE-1 score increases from 34.66% to 46.28%.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

QMSum: A New Benchmark for Query-based Multi-domain Meeting Summarization

TL;DR: This work defines a new query-based multi-domain meeting summarization task, where models have to select and summarize relevant spans of meetings in response to a query, and introduces QMSum, a new benchmark for this task.
Proceedings ArticleDOI

MediaSum: A Large-scale Media Interview Dataset for Dialogue Summarization

TL;DR: This paper introduces MediaSum, a large-scale media interview dataset consisting of 463.6K transcripts with abstractive summaries that can be used in transfer learning to improve a model’s performance on other dialogue summarization tasks.
Proceedings ArticleDOI

Structure-Aware Abstractive Conversation Summarization via Discourse and Action Graphs

TL;DR: This work proposes to explicitly model the rich structures in conversations for more precise and accurate conversation summarization, by first incorporating discourse relations between utterances and action triples through structured graphs to better encode conversations, and then designing a multi-granularity decoder to generate summaries by combining all levels of information.
Posted Content

Generating SOAP Notes from Doctor-Patient Conversations

TL;DR: This paper describes a unique dataset of patient visit records, consisting of transcripts, paired SOAP notes, and annotations marking noteworthy utterances that support each summary sentence, and presents the first study to evaluate complete pipelines for leveraging these transcripts to train machine learning model to generate these notes.
Proceedings ArticleDOI

How Domain Terminology Affects Meeting Summarization Performance

TL;DR: This paper creates gold-standard annotations for domain terminology on a sizable meeting corpus; they are known as jargon terms and reveal that domain terminology can have a substantial impact on summarization performance.
References
More filters
Proceedings Article

Attention is All you Need

TL;DR: This paper proposed a simple network architecture based solely on an attention mechanism, dispensing with recurrence and convolutions entirely and achieved state-of-the-art performance on English-to-French translation.
Proceedings ArticleDOI

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

TL;DR: BERT as mentioned in this paper pre-trains deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Proceedings Article

Sequence to Sequence Learning with Neural Networks

TL;DR: The authors used a multilayered Long Short-Term Memory (LSTM) to map the input sequence to a vector of a fixed dimensionality, and then another deep LSTM to decode the target sequence from the vector.
Proceedings Article

ROUGE: A Package for Automatic Evaluation of Summaries

TL;DR: Four different RouGE measures are introduced: ROUGE-N, ROUge-L, R OUGE-W, and ROUAGE-S included in the Rouge summarization evaluation package and their evaluations.
Proceedings Article

TextRank: Bringing Order into Text

Rada Mihalcea, +1 more
TL;DR: TextRank, a graph-based ranking model for text processing, is introduced and it is shown how this model can be successfully used in natural language applications.