scispace - formally typeset
Open AccessProceedings ArticleDOI

MovieChats: Chat like Humans in a Closed Domain

Reads0
Chats0
TLDR
This work takes a close look at the movie domain and presents a large-scale high-quality corpus with fine-grained annotations in hope of pushing the limit of movie-domain chatbots.
Abstract: 
Being able to perform in-depth chat with humans in a closed domain is a precondition before an open-domain chatbot can be ever claimed. In this work, we take a close look at the movie domain and present a large-scale high-quality corpus with fine-grained annotations in hope of pushing the limit of movie-domain chatbots. We propose a unified, readily scalable neural approach which reconciles all subtasks like intent prediction and knowledge retrieval. The model is first pretrained on the huge general-domain data, then finetuned on our corpus. We show this simple neural approach trained on high-quality data is able to outperform commercial systems replying on complex rules. On both the static and interactive tests, we find responses generated by our system exhibits remarkably good engagement and sensibleness close to human-written ones. We further analyze the limits of our work and point out potential directions for future work

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

AST-Trans: Code Summarization with Efficient Tree-Structured Attention

TL;DR: AST-Trans is proposed in this paper which exploits two types of node relationships in the AST: ancestor-descendant and sibling relationships and applies the tree-structured attention to dynamically allocate weights for relevant nodes and exclude irrelevant nodes based on these two relationships.
Journal ArticleDOI

Recent Advances in Neural Text Generation: A Task-Agnostic Survey

TL;DR: A task-agnostic survey of recent advances in neural text generation is presented, which group under the following four headings: data construction, neural frameworks, training and inference strategies, and evaluation metrics.
Journal ArticleDOI

Towards information-rich, logical dialogue systems with knowledge-enhanced neural models

TL;DR: A comprehensive review of knowledge-enhanced dialogue systems is given, summarizes research progress to solve challenges, and proposes some open issues and research directions.
Journal ArticleDOI

MDIA: A Benchmark for Multilingual Dialogue Generation in 46 Languages

TL;DR: M DIA is presented, the first large-scale multilingual benchmark for dialogue generation across low- to high-resource languages and it covers real-life conversations in 46 languages across 19 language families.
Journal ArticleDOI

A Survey on Legal Judgment Prediction: Datasets, Metrics, Models and Challenges

TL;DR: Up-to-date andhensive review of existing LJP tasks, data sets, models andevaluations are provided to help researchers and legal professionals understand the status of LJP.
References
More filters
Proceedings ArticleDOI

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

TL;DR: BERT as mentioned in this paper pre-trains deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Posted Content

RoBERTa: A Robustly Optimized BERT Pretraining Approach

TL;DR: It is found that BERT was significantly undertrained, and can match or exceed the performance of every model published after it, and the best model achieves state-of-the-art results on GLUE, RACE and SQuAD.
Proceedings ArticleDOI

Deep contextualized word representations

TL;DR: This paper introduced a new type of deep contextualized word representation that models both complex characteristics of word use (e.g., syntax and semantics), and how these uses vary across linguistic contexts (i.e., to model polysemy).
Related Papers (5)