scispace - formally typeset
Proceedings ArticleDOI

Out-of-Domain Slot Value Detection for Spoken Dialogue Systems with Context Information

TLDR
A Recurrent Neural Network encoder-decoder model is used and a method that uses only in-domain data is proposed that is robust against over-fitting problems because it is independent of the slot values of the training data.
Abstract
This paper proposes an approach to detecting-of-domain slot values from user utterances in spoken dialogue systems based on contexts. The approach detects keywords of slot values from utterances and consults domain knowledge (i.e., an ontology) to check whether the keywords are-of-domain. This can prevent the systems from responding improperly to user requests. We use a Recurrent Neural Network (RNN) encoder-decoder model and propose a method that uses only in-domain data. The method replaces word embedding vectors of the keywords corresponding to slot values with random vectors during training of the model. This allows using context information. The model is robust against over-fitting problems because it is independent of the slot values of the training data. Experiments show that the proposed method achieves a 65% gain in F1 score relative to a baseline model and a further 13 percentage points by combining with other methods.

read more

Citations
More filters
Proceedings ArticleDOI

Flexibly-Structured Model for Task-Oriented Dialogues.

TL;DR: This architecture is scalable to real-world scenarios and is shown through an empirical evaluation to achieve state-of-the-art performance on both the Cambridge Restaurant dataset and the Stanford in-car assistant dataset.
Journal ArticleDOI

NLP-Based Query-Answering System for Information Extraction from Building Information Models

TL;DR: In this paper , the authors developed a QA system for BIM information extraction (IE) by using natural language processing (NLP) methods to build a virtual assistant for construction project team members.
Proceedings ArticleDOI

Slot Filling with Weighted Multi-Encoders for Out-of-Domain Values.

TL;DR: A new method for slot filling of out-ofdomain (OOD) slot values, which are not included in the training data, in spoken dialogue systems, using two encoders, which distinctly encode contexts and keywords, respectively.
Posted Content

Interactive teaching for conversational AI

TL;DR: A new Teachable AI system that is capable of learning new language nuggets called concepts, directly from end users using live interactive teaching sessions, and it is demonstrated that this method is very promising in leading way to build more adaptive and personalized language understanding models.
References
More filters
Journal Article

Dropout: a simple way to prevent neural networks from overfitting

TL;DR: It is shown that dropout improves the performance of neural networks on supervised learning tasks in vision, speech recognition, document classification and computational biology, obtaining state-of-the-art results on many benchmark data sets.
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Posted Content

Efficient Estimation of Word Representations in Vector Space

TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Proceedings Article

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: It is conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and it is proposed to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Proceedings ArticleDOI

Learning Phrase Representations using RNN Encoder--Decoder for Statistical Machine Translation

TL;DR: In this paper, the encoder and decoder of the RNN Encoder-Decoder model are jointly trained to maximize the conditional probability of a target sequence given a source sequence.
Related Papers (5)