scispace - formally typeset
Search or ask a question

Showing papers on "Natural language understanding published in 2016"


BookDOI
18 Nov 2016
TL;DR: The view is that dealing with anaphoric language can be decomposed into two complementary tasks: identifying what a text potentially makes available for anaphor reference and how it does so and constraining the candidate set of a given anaphic expression down to one possible choice.
Abstract: : Extended natural language communication between a person engaged in solving a problem or seeking information and a machine providing assistance requires the machine to be able to deal with anaphoric language in a perspicuous, transportable non-ad hoc way. This report takes the view that dealing with anaphoric language can be decomposed into two complementary tasks: (1) identifying what a text potentially makes available for anaphoric reference; and (2) constraining the candidate set of a given anaphoric expression down to one possible choice. The second task has been called the 'Anaphor resolution' problem and, to date, has stimulated much research in psychology and artificial intelligence natural language understanding. The focus of this report is the first task - that of identifying what a text makes available for anaphoric reference and how it does so.

335 citations


Proceedings Article
01 May 2016
TL;DR: This paper revisits and extends these dependency graph representations in light of the recent Universal Dependencies (UD) initiative and provides a detailed account of an enhanced and an enhanced++ English UD representation.
Abstract: Many shallow natural language understanding tasks use dependency trees to extract relations between content words. However, strict surface-structure dependency trees tend to follow the linguistic structure of sentences too closely and frequently fail to provide direct relations between content words. To mitigate this problem, the original Stanford Dependencies representation also defines two dependency graph representations which contain additional and augmented relations that explicitly capture otherwise implicit relations between content words. In this paper, we revisit and extend these dependency graph representations in light of the recent Universal Dependencies (UD) initiative and provide a detailed account of an enhanced and an enhanced++ English UD representation. We further present a converter from constituency to basic, i.e., strict surface structure, UD trees, and a converter from basic UD trees to enhanced and enhanced++ English UD graphs. We release both converters as part of Stanford CoreNLP and the Stanford Parser.

251 citations


Journal ArticleDOI
TL;DR: This paper presents a model that takes into account the variations in natural language and ambiguities in grounding them to robotic instructions with appropriate environment context and task constraints, based on an energy function that encodes such properties in a form isomorphic to a conditional random field.
Abstract: It is important for a robot to be able to interpret natural language commands given by a human. In this paper, we consider performing a sequence of mobile manipulation tasks with instructions described in natural language. Given a new environment, even a simple task such as boiling water would be performed quite differently depending on the presence, location and state of the objects. We start by collecting a dataset of task descriptions in free-form natural language and the corresponding grounded task-logs of the tasks performed in an online robot simulator. We then build a library of verb-environment instructions that represents the possible instructions for each verb in that environment, these may or may not be valid for a different environment and task context. We present a model that takes into account the variations in natural language and ambiguities in grounding them to robotic instructions with appropriate environment context and task constraints. Our model also handles incomplete or noisy natural language instructions. It is based on an energy function that encodes such properties in a form isomorphic to a conditional random field. We evaluate our model on tasks given in a robotic simulator and show that it successfully outperforms the state of the art with 61.8% accuracy. We also demonstrate a grounded robotic instruction sequence on a PR2 robot using the Learning from Demonstration approach.

225 citations


Proceedings ArticleDOI
01 Aug 2016
TL;DR: This work presents WIKIREADING, a large-scale natural language understanding task and publicly-available dataset with 18 million instances, and compares various state-of-the-art DNNbased architectures for document classification, information extraction, and question answering.
Abstract: We present WIKIREADING, a large-scale natural language understanding task and publicly-available dataset with 18 million instances. The task is to predict textual values from the structured knowledge base Wikidata by reading the text of the corresponding Wikipedia articles. The task contains a rich variety of challenging classification and extraction sub-tasks, making it well-suited for end-to-end models such as deep neural networks (DNNs). We compare various state-of-the-art DNNbased architectures for document classification, information extraction, and question answering. We find that models supporting a rich answer space, such as word or character sequences, perform best. Our best-performing model, a word-level sequence to sequence model with a mechanism to copy out-of-vocabulary words, obtains an accuracy of 71.8%.

138 citations


Posted Content
TL;DR: This paper presents a new state-of-the-art result, achieving the accuracy of 88.3% on the standard benchmark, the Stanford Natural Language Inference dataset, through an enhanced sequential encoding model, which outperforms the previous best model that employs more complicated network architectures.
Abstract: Reasoning and inference are central to human and artificial intelligence. Modeling inference in human language is notoriously challenging but is fundamental to natural language understanding and many applications. With the availability of large annotated data, neural network models have recently advanced the field significantly. In this paper, we present a new state-of-the-art result, achieving the accuracy of 88.3% on the standard benchmark, the Stanford Natural Language Inference dataset. This result is achieved first through our enhanced sequential encoding model, which outperforms the previous best model that employs more complicated network architectures, suggesting that the potential of sequential LSTM-based models have not been fully explored yet in previous work. We further show that by explicitly considering recursive architectures, we achieve additional improvement. Particularly, incorporating syntactic parse information contributes to our best result; it improves the performance even when the parse information is added to an already very strong system.

130 citations


Journal ArticleDOI
Percy Liang1
TL;DR: Semantic parsing is a rich fusion of the logical and the statistical worlds and can be applied to practically any type of data type.
Abstract: For building question answering systems and natural language interfaces, semantic parsing has emerged as an important and powerful paradigm. Semantic parsers map natural language into logical forms, the classic representation for many important linguistic phenomena. The modern twist is that we are interested in learning semantic parsers from data, which introduces a new layer of statistical and computational issues. This article lays out the components of a statistical semantic parser, highlighting the key challenges. We will see that semantic parsing is a rich fusion of the logical and the statistical world, and that this fusion will play an integral role in the future of natural language understanding systems.

122 citations


Proceedings Article
12 Feb 2016
TL;DR: This work presents a system that excels at all the tasks except one and demonstrates that the introduction of a reasoning module significantly improves the performance of an intelligent agent.
Abstract: A group of researchers from Facebook has recently proposed a set of 20 question-answering tasks (Facebook's bAbl dataset) as a challenge for the natural language understanding ability of an intelligent agent. These tasks are designed to measure various skills of an agent, such as: fact based question-answering, simple induction, the ability to find paths, co-reference resolution and many more. Their goal is to aid in the development of systems that can learn to solve such tasks and to allow a proper evaluation of such systems. They show existing systems cannot fully solve many of those toy tasks. In this work, we present a system that excels at all the tasks except one. The proposed model of the agent uses the Answer Set Programming (ASP) language as the primary knowledge representation and reasoning language along with the standard statistical Natural Language Processing (NLP) models. Given a training dataset containing a set of narrations, questions and their answers, the agent jointly uses a translation system, an Inductive Logic Programming algorithm and Statistical NLP methods to learn the knowledge needed to answer similar questions. Our results demonstrate that the introduction of a reasoning module significantly improves the performance of an intelligent agent.

98 citations


Journal ArticleDOI
TL;DR: An original experiment comparing five successful artificial dialogue systems with an online version of Eliza to find if current dialogue systems use the same, psychotherapist questioning technique as Joseph Weizenbaum's 1960 natural language understanding programme, Eliza, shows statistical significance shows these dialogue systems are an improvement on their predecessor.

88 citations


Posted Content
TL;DR: This article presented WikiReading, a large-scale natural language understanding task and publicly-available dataset with 18 million instances, where the task is to predict textual values from the structured knowledge base Wikidata by reading the text of the corresponding Wikipedia articles.
Abstract: We present WikiReading, a large-scale natural language understanding task and publicly-available dataset with 18 million instances. The task is to predict textual values from the structured knowledge base Wikidata by reading the text of the corresponding Wikipedia articles. The task contains a rich variety of challenging classification and extraction sub-tasks, making it well-suited for end-to-end models such as deep neural networks (DNNs). We compare various state-of-the-art DNN-based architectures for document classification, information extraction, and question answering. We find that models supporting a rich answer space, such as word or character sequences, perform best. Our best-performing model, a word-level sequence to sequence model with a mechanism to copy out-of-vocabulary words, obtains an accuracy of 71.8%.

84 citations


Proceedings Article
Jason Weston1
20 Apr 2016
TL;DR: This paper proposed a model incorporating predictive look-ahead to learn from a teacher's response to answer questions without any reward-based supervision at all, and showed that the model can learn to answer question correctly without reward at all.
Abstract: A long-term goal of machine learning research is to build an intelligent dialog agent. Most research in natural language understanding has focused on learning from fixed training sets of labeled data, with supervision either at the word level (tagging, parsing tasks) or sentence level (question answering, machine translation). This kind of supervision is not realistic of how humans learn, where language is both learned by, and used for, communication. In this work, we study dialog-based language learning, where supervision is given naturally and implicitly in the response of the dialog partner during the conversation. We study this setup in two domains: the bAbI dataset of (Weston et al., 2015) and large-scale question answering from (Dodge et al., 2015). We evaluate a set of baseline learning strategies on these tasks, and show that a novel model incorporating predictive lookahead is a promising approach for learning from a teacher's response. In particular, a surprising result is that it can learn to answer questions correctly without any reward-based supervision at all.

80 citations


Patent
29 Jun 2016
TL;DR: The authors use a hierarchical organization of intents/commands and entity types, and trained models associated with those hierarchies, so that commands and entities may be determined for incoming text queries without necessarily determining a domain for the incoming text.
Abstract: A system capable of performing natural language understanding (NLU) without the concept of a domain that influences NLU results. The present system uses a hierarchical organizations of intents/commands and entity types, and trained models associated with those hierarchies, so that commands and entity types may be determined for incoming text queries without necessarily determining a domain for the incoming text. The system thus operates in a domain agnostic manner, in a departure from multi-domain architecture NLU processing where a system determines NLU results for multiple domains simultaneously and then ranks them to determine which to select as the result.

Posted Content
Jason Weston1
TL;DR: The authors proposed a model incorporating predictive look-ahead to learn from a teacher's response to answer questions without any reward-based supervision at all, and showed that the model can learn to answer question correctly without reward at all.
Abstract: A long-term goal of machine learning research is to build an intelligent dialog agent. Most research in natural language understanding has focused on learning from fixed training sets of labeled data, with supervision either at the word level (tagging, parsing tasks) or sentence level (question answering, machine translation). This kind of supervision is not realistic of how humans learn, where language is both learned by, and used for, communication. In this work, we study dialog-based language learning, where supervision is given naturally and implicitly in the response of the dialog partner during the conversation. We study this setup in two domains: the bAbI dataset of (Weston et al., 2015) and large-scale question answering from (Dodge et al., 2015). We evaluate a set of baseline learning strategies on these tasks, and show that a novel model incorporating predictive lookahead is a promising approach for learning from a teacher's response. In particular, a surprising result is that it can learn to answer questions correctly without any reward-based supervision at all.

Proceedings ArticleDOI
08 Sep 2016
TL;DR: The proposed multi-task model delivers better performance with less data by leveraging patterns that it learns from the other tasks, and supports an open vocabulary, which allows the models to generalize to unseen words.
Abstract: The goal of this paper is to use multi-task learning to efficiently scale slot filling models for natural language understanding to handle multiple target tasks or domains. The key to scalability is reducing the amount of training data needed to learn a model for a new task. The proposed multi-task model delivers better performance with less data by leveraging patterns that it learns from the other tasks. The approach supports an open vocabulary, which allows the models to generalize to unseen words, which is particularly important when very little training data is used. A newly collected crowd-sourced data set, covering four different domains, is used to demonstrate the effectiveness of the domain adaptation and open vocabulary techniques.

Journal ArticleDOI
TL;DR: This work is the first, to the knowledge, to integrate a computational model of general language understanding and humor theory to quantitatively predict humor at a fine‐grained level and is presented as an example of a framework for applying models of language processing to understand higher level linguistic and cognitive phenomena.

Book ChapterDOI
01 Jan 2016
TL;DR: The authors proposed a probabilistic framework that enables robots to follow commands given in natural language, without any prior knowledge of the environment, exploiting environment information implicit in the instruction, thereby treating language as a type of sensor that is used to formulate a prior distribution over the unknown parts of the environments.
Abstract: Natural language provides a flexible, intuitive way for people to command robots, which is becoming increasingly important as robots transition to working alongside people in our homes and workplaces. To follow instructions in unknown environments, robots will be expected to reason about parts of the environments that were described in the instruction, but that the robot has no direct knowledge about. However, most existing approaches to natural language understanding require that the robot’s environment be known a priori. This paper proposes a probabilistic framework that enables robots to follow commands given in natural language, without any prior knowledge of the environment. The novelty lies in exploiting environment information implicit in the instruction, thereby treating language as a type of sensor that is used to formulate a prior distribution over the unknown parts of the environment. The algorithm then uses this learned distribution to infer a sequence of actions that are most consistent with the command, updating our belief as we gather more metric information. We evaluate our approach through simulation as well as experiments on two mobile robots; our results demonstrate the algorithm’s ability to follow navigation commands with performance comparable to that of a fully-known environment.

Proceedings ArticleDOI
01 Jun 2016
TL;DR: Two distinct models that capture semantic frame chains and discourse information while abstracting over the specific mentions of predicates and entities are developed.
Abstract: Natural language understanding often requires deep semantic knowledge. Expanding on previous proposals, we suggest that some important aspects of semantic knowledge can be modeled as a language model if done at an appropriate level of abstraction. We develop two distinct models that capture semantic frame chains and discourse information while abstracting over the specific mentions of predicates and entities. For each model, we investigate four implementations: a "standard" N-gram language model and three discriminatively trained "neural" language models that generate embeddings for semantic frames. The quality of the semantic language models (SemLM) is evaluated both intrinsically, using perplexity and a narrative cloze test and extrinsically - we show that our SemLM helps improve performance on semantic natural language processing tasks such as co-reference resolution and discourse parsing.

Journal ArticleDOI
TL;DR: In this paper, a survey discusses requirements and challenges of developing such systems and reports 26 graphical systems that exploit natural language interfaces and addresses both artificial intelligence and visualization aspects, which serve as a frame of reference to researchers and to enable further advances in the field.
Abstract: A natural language interface exploits the conceptual simplicity and naturalness of the language to create a high-level user-friendly communication channel between humans and machines. One of the promising applications of such interfaces is generating visual interpretations of semantic content of a given natural language that can be then visualized either as a static scene or a dynamic animation. This survey discusses requirements and challenges of developing such systems and reports 26 graphical systems that exploit natural language interfaces and addresses both artificial intelligence and visualization aspects. This work serves as a frame of reference to researchers and to enable further advances in the field.

Posted Content
TL;DR: The experiments on the benchmark Air Travel Information System data show that the proposed K-SAN architecture can effectively extract salient knowledge from substructures with an attention mechanism, and outperform the performance of the state-of-the-art neural network based frameworks.
Abstract: Natural language understanding (NLU) is a core component of a spoken dialogue system. Recently recurrent neural networks (RNN) obtained strong results on NLU due to their superior ability of preserving sequential information over time. Traditionally, the NLU module tags semantic slots for utterances considering their flat structures, as the underlying RNN structure is a linear chain. However, natural language exhibits linguistic properties that provide rich, structured information for better understanding. This paper introduces a novel model, knowledge-guided structural attention networks (K-SAN), a generalization of RNN to additionally incorporate non-flat network topologies guided by prior knowledge. There are two characteristics: 1) important substructures can be captured from small training data, allowing the model to generalize to previously unseen test data; 2) the model automatically figures out the salient substructures that are essential to predict the semantic tags of the given sentences, so that the understanding performance can be improved. The experiments on the benchmark Air Travel Information System (ATIS) data show that the proposed K-SAN architecture can effectively extract salient knowledge from substructures with an attention mechanism, and outperform the performance of the state-of-the-art neural network based frameworks.

Posted Content
TL;DR: SemLM as mentioned in this paper proposes to model important aspects of semantic knowledge as a language model if done at an appropriate level of abstraction, which helps improve performance on semantic natural language processing tasks such as co-reference resolution and discourse parsing.
Abstract: Natural language understanding often requires deep semantic knowledge. Expanding on previous proposals, we suggest that some important aspects of semantic knowledge can be modeled as a language model if done at an appropriate level of abstraction. We develop two distinct models that capture semantic frame chains and discourse information while abstracting over the specific mentions of predicates and entities. For each model, we investigate four implementations: a "standard" N-gram language model and three discriminatively trained "neural" language models that generate embeddings for semantic frames. The quality of the semantic language models (SemLM) is evaluated both intrinsically, using perplexity and a narrative cloze test and extrinsically - we show that our SemLM helps improve performance on semantic natural language processing tasks such as co-reference resolution and discourse parsing.

Journal ArticleDOI
TL;DR: It is considered that the joint models, or end-to-end models, will be one important trend for developing Human-Computer dialogue systems.

Proceedings ArticleDOI
07 Jul 2016
TL;DR: This talk will emphasize the two topics of how NLP can contribute to understanding textual relationships and how deep learning approaches substantially aid in this goal, as well as looking at some of the recent work in these areas.
Abstract: There is a lot of overlap between the core problems of information retrieval (IR) and natural language processing (NLP). An IR system gains from understanding a user need and from understanding documents, and hence being able to determine whether a document has information that satisfies the user need. Much of NLP is about the same thing: Natural language understanding aims to understand the meaning of questions and documents and meaning relationships. The exciting recent application of deep learning approaches in NLP has brought new tools for effectively understanding language semantics. In principle, there should be a lot of synergy, though in practice the concerns of IR on large systems and macro-scale understanding have tended to contrast with the emphasis in NLP on language structure and micro-scale understanding. My talk will emphasize the two topics of how NLP can contribute to understanding textual relationships and how deep learning approaches substantially aid in this goal. One basic -- and very successful tool -- has been the new generation of distributed word representations: neural word embeddings. However, beyond just word meanings, we need to understand how to compose the meanings of larger pieces of text. Two requirements for that are good ways to understand the structure of human language utterances and ways to compose their meanings. Deep learning methods can help for both tasks. Finally, we need to understand relationships between pieces of text, to be able to do tasks such as Natural Language Inference (or Recognizing Textual Entailment) and Question Answering, and I will look at some of our recent work in these areas, both with and without the help of neural networks

01 Jan 2016
TL;DR: A theory of "syntactic semantics" is advocated as a way of understanding how computers can think (and how the Chinese-RoomArgument objection to the Turing Test can be overcome).
Abstract: I advocate a theory of "syntactic semantics" as a way of understanding how computers can think (and how the Chinese-RoomArgument objection to the Turing Test can be overcome): (1) Semantics, considered as the study of relations between symbols and meanings, can be turned into syntax a study of relations among symbols (including meanings) and hence syntax (ie, symbol manipulation) can suffice for the semantical enterprise (contra Searle) (2) Semantics, considered as the process of understanding one domain (by modeling it) in terms of another, can be viewed recursively: The base case of semantic understanding understanding a domain in terms of itself is "syntactic understanding" (3) An internal (or "narrow"), first-person point of view makes an external (or "wide"), third-person point of view otiose for purposes of understanding cognition

Patent
23 Nov 2016
TL;DR: In this paper, a natural language understanding model is trained using respective natural language example inputs corresponding to a plurality of applications, and a determination is made as to whether a value of a first parameter of the first application is to be obtained using natural language interaction.
Abstract: A natural language understanding model is trained using respective natural language example inputs corresponding to a plurality of applications. A determination is made as to whether a value of a first parameter of a first application is to be obtained using a natural language interaction. Using the natural language understanding model, at least a portion of the first application is generated.

Patent
16 Aug 2016
TL;DR: In this paper, the authors present a system for transforming formal and informal natural language user inputs into a more formal, machine-readable, structured representation of a search query, which is used for automated searches for the most relevant items available for purchase in an electronic marketplace.
Abstract: Systems and methods for transforming formal and informal natural language user inputs into a more formal, machine-readable, structured representation of a search query. In one scenario, a processed sequence of user inputs and machine-generated prompts for further data from a user in a multi-turn interactive dialog improves the efficiency and accuracy of automated searches for the most relevant items available for purchase in an electronic marketplace. Analysis of user inputs may discern user intent, user input type, a dominant object of user interest, item categories, item attributes, attribute values, and item recipients. Other inputs considered may include dialog context, item inventory-related information, and external knowledge to improve inference of user intent from user input. Different types of analyses of the inputs each yield results that are interpreted in aggregate and coordinated via a knowledge graph based on past users' interactions with the electronic marketplace and/or inventory-related data.

Proceedings Article
19 Sep 2016
TL;DR: This paper presents an unusual middle-out procedure that targets mid-level symbols and traverses the grammar by both forward chaining and backward chaining, expanding symbols conditionally by testing against the current game state.
Abstract: The Expressive Intelligence Studio is developing a new approach to freeform conversational interaction in playable media that combines dialogue management, natural language generation (NLG), and natural language understanding. In this paper, we present our method for dialogue generation, which has been fully implemented in a game we are developing called Talk of the Town . Eschewing a traditional NLG pipeline, we take up a novel approach that combines human language expertise with computer generativity. Specifically, this method utilizes a tool that we have developed for authoring context-free grammars (CFGs) whose productions come packaged with explicit metadata. Instead of terminally expanding top-level symbols — the conventional way of generating from a CFG — we employ an unusual middle-out procedure that targets mid-level symbols and traverses the grammar by both forward chaining and backward chaining, expanding symbols conditionally by testing against the current game state. In this paper, we present our method, discuss a series of associated authoring patterns, and situate our approach against the few earlier projects in this area.

Proceedings ArticleDOI
01 Nov 2016
TL;DR: A practical technique that addresses the issue of reuse of slots across different domains and tasks in a web-scale language understanding system: Microsoft’s personal digital assistant Cortana.
Abstract: Natural language understanding is the core of the human computer interactions. However, building new domains and tasks that need a separate set of models is a bottleneck for scaling to a large number of domains and experiences. In this paper, we propose a practical technique that addresses this issue in a web-scale language understanding system: Microsoft’s personal digital assistant Cortana. The proposed technique uses a constrained decoding method with a universal slot tagging model sharing the same schema as the collection of slot taggers built for each domain. The proposed approach allows reusing of slots across different domains and tasks while achieving virtually the same performance as those slot taggers trained per domain fashion.

BookDOI
02 Dec 2016
TL;DR: A novel word sense disambiguation (WSD) discriminative model is proposed which integrates structural context with the local context to handle long distance sense dependency and multireference lexicon dependency within the sentence.
Abstract: This book constitutes the joint refereed proceedings of the5th CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2016, and the 24th International Conference on Computer Processing of Oriental Languages, ICCPOL 2016, held in Kunming, China, in December 2016. The 48 revised full papers presented together with 41 short papers were carefully reviewed and selected from 216 submissions. The papers cover fundamental research in language computing, multi-lingual access, webmining/text mining, machine learning for NLP, knowledge graph, NLP for social network, as well as applications in language computing

Proceedings Article
12 Feb 2016
TL;DR: The authors propose various unstructured and structured models that capture fulfillment cues such as the subject's emotional state and actions, and demonstrate the importance of understanding the narrative and discourse structure to address this task.
Abstract: The ability to comprehend wishes or desires and their fulfillment is important to Natural Language Understanding. This paper introduces the task of identifying if a desire expressed by a subject in a given short piece of text was fulfilled. We propose various unstructured and structured models that capture fulfillment cues such as the subject's emotional state and actions. Our experiments with two different datasets demonstrate the importance of understanding the narrative and discourse structure to address this task.

Posted Content
TL;DR: This article used multi-task learning to efficiently scale slot-filling models for NLP to handle multiple target tasks or domains, which can reduce the amount of training data needed to learn a model for a new task.
Abstract: The goal of this paper is to use multi-task learning to efficiently scale slot filling models for natural language understanding to handle multiple target tasks or domains. The key to scalability is reducing the amount of training data needed to learn a model for a new task. The proposed multi-task model delivers better performance with less data by leveraging patterns that it learns from the other tasks. The approach supports an open vocabulary, which allows the models to generalize to unseen words, which is particularly important when very little training data is used. A newly collected crowd-sourced data set, covering four different domains, is used to demonstrate the effectiveness of the domain adaptation and open vocabulary techniques.

Posted Content
Percy Liang1
TL;DR: In this paper, the authors lay out the components of a statistical semantic parser, highlighting the key challenges of learning semantic parsers from data, and show that semantic parsing is a rich fusion of the logical and statistical world, and that this fusion will play an integral role in the future of natural language understanding systems.
Abstract: For building question answering systems and natural language interfaces, semantic parsing has emerged as an important and powerful paradigm. Semantic parsers map natural language into logical forms, the classic representation for many important linguistic phenomena. The modern twist is that we are interested in learning semantic parsers from data, which introduces a new layer of statistical and computational issues. This article lays out the components of a statistical semantic parser, highlighting the key challenges. We will see that semantic parsing is a rich fusion of the logical and the statistical world, and that this fusion will play an integral role in the future of natural language understanding systems.