scispace - formally typeset
Open AccessPosted Content

Neural Production Systems

TLDR
In this paper, a set of rule templates are applied by binding placeholder variables in the rules to specific entities, and the best fitting rules are applied to update entity properties, which achieves a flexible, dynamic flow of control and serves to factorize entity specific and rule-based information.
Abstract
Visual environments are structured, consisting of distinct objects or entities. These entities have properties -- both visible and latent -- that determine the manner in which they interact with one another. To partition images into entities, deep-learning researchers have proposed structural inductive biases such as slot-based architectures. To model interactions among entities, equivariant graph neural nets (GNNs) are used, but these are not particularly well suited to the task for two reasons. First, GNNs do not predispose interactions to be sparse, as relationships among independent entities are likely to be. Second, GNNs do not factorize knowledge about interactions in an entity-conditional manner. As an alternative, we take inspiration from cognitive science and resurrect a classic approach, production systems, which consist of a set of rule templates that are applied by binding placeholder variables in the rules to specific entities. Rules are scored on their match to entities, and the best fitting rules are applied to update entity properties. In a series of experiments, we demonstrate that this architecture achieves a flexible, dynamic flow of control and serves to factorize entity-specific and rule-based information. This disentangling of knowledge achieves robust future-state prediction in rich visual environments, outperforming state-of-the-art methods using GNNs, and allows for the extrapolation from simple (few object) environments to more complex environments.

read more

Citations
More filters
Posted Content

Discrete-Valued Neural Communication

TL;DR: In this paper, the authors extend the quantization mechanism from the vector-quantized variational autoencoder to multi-headed discretization with shared codebooks and use it for discrete-valued neural communication (DVNC).
Posted Content

Projection: A Mechanism for Human-like Reasoning in Artificial Intelligence.

TL;DR: In this article, the relevance of projection to commonsense knowledge problem is discussed, and the importance of using knowledge projection to solve the commonsense problem of applying knowledge to varied or challenging situations, across a range of AI domains.
Posted Content

Knowledge accumulating: The general pattern of learning.

TL;DR: In this paper, the authors proposed a pattern that explains how to accumulate knowledge to solve sparse feedback problems, and analyzed how sparse feedback affects algorithm perfomance, and then suggested a pattern for accumulating knowledge.
Posted Content

Dynamic Inference with Neural Interpreters

TL;DR: Neural interpreters as mentioned in this paper factorizes inference in a self-attention network as a system of modules, which are called ''functions'' and are routed through a sequence of functions in a way that is end-to-end learned.
References
More filters
Proceedings Article

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: It is conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and it is proposed to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Posted Content

Attention Is All You Need

TL;DR: A new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely is proposed, which generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.
Journal ArticleDOI

Objective Criteria for the Evaluation of Clustering Methods

TL;DR: This article proposes several criteria which isolate specific aspects of the performance of a method, such as its retrieval of inherent structure, its sensitivity to resampling and the stability of its results in the light of new data.
Journal ArticleDOI

The Graph Neural Network Model

TL;DR: A new neural network model, called graph neural network (GNN) model, that extends existing neural network methods for processing the data represented in graph domains, and implements a function tau(G,n) isin IRm that maps a graph G and one of its nodes n into an m-dimensional Euclidean space.
Journal ArticleDOI

Adaptive mixtures of local experts

TL;DR: A new supervised learning procedure for systems composed of many separate networks, each of which learns to handle a subset of the complete set of training cases, which is demonstrated to be able to be solved by a very simple expert network.
Related Papers (5)