scispace - formally typeset
Search or ask a question

Showing papers by "Michael I. Jordan published in 1991"


Journal ArticleDOI
TL;DR: A new supervised learning procedure for systems composed of many separate networks, each of which learns to handle a subset of the complete set of training cases, which is demonstrated to be able to be solved by a very simple expert network.
Abstract: We present a new supervised learning procedure for systems composed of many separate networks, each of which learns to handle a subset of the complete set of training cases. The new procedure can be viewed either as a modular version of a multilayer supervised network, or as an associative version of competitive learning. It therefore provides a new link between these two apparently different approaches. We demonstrate that the learning procedure divides up a vowel discrimination task into appropriate subtasks, each of which can be solved by a very simple expert network.

4,338 citations


Journal ArticleDOI
TL;DR: A novel modular connectionist architecture is presented in which the networks composing the architecture compete to learn the training patterns, and an outcome of the competition is that different networks learn different training patterns and, thus, learn to compute different functions.

496 citations


Journal ArticleDOI
TL;DR: A more biologically plausible learning rule is described, using reinforcement learning, which is applied to the problem of how area 7a in the posterior parietal cortex of monkeys might represent visual space in head-centered coordinates and shows that a neural network does not require back propagation to acquire biologically interesting properties.
Abstract: Many recent studies have used artificial neural network algorithms to model how the brain might process information. However, back-propagation learning, the method that is generally used to train these networks, is distinctly "unbiological." We describe here a more biologically plausible learning rule, using reinforcement learning, which we have applied to the problem of how area 7a in the posterior parietal cortex of monkeys might represent visual space in head-centered coordinates. The network behaves similarly to networks trained by using back-propagation and to neurons recorded in area 7a. These results show that a neural network does not require back propagation to acquire biologically interesting properties.

228 citations


Proceedings Article
02 Dec 1991
TL;DR: A neural network architecture that discovers a recursive decomposition of its input space that uses competition among networks to recursively split the input space into nested regions and to learn separate associative mappings within each region.
Abstract: In this paper we present a neural network architecture that discovers a recursive decomposition of its input space. Based on a generalization of the modular architecture of Jacobs, Jordan, Nowlan, and Hinton (1991), the architecture uses competition among networks to recursively split the input space into nested regions and to learn separate associative mappings within each region. The learning algorithm is shown to perform gradient ascent in a log likelihood function that captures the architecture's hierarchical structure.

172 citations


Journal ArticleDOI
TL;DR: Two neural networks are developed with architecture similar to Zipser and Andersen's model and trained to perform the same task using a more biologically plausible learning procedure than backpropagation, which corroborates the validity of this neural network's computational algorithm as a plausible model of how area 7a may perform coordinate transformations.
Abstract: Area 7a of the posterior parietal cortex of the primate brain is concerned with representing head-centered space by combining information about the retinal location of a visual stimulus and the position of the eyes in the orbits. An artificial neural network was previously trained to perform this coordinate transformation task using the backpropagation learning procedure, and units in its middle layer (the hidden units) developed properties very similar to those of area 7a neurons presumed to code for spatial location (Andersen and Zipser, 1988; Zipser and Andersen, 1988). We developed two neural networks with architecture similar to Zipser and Andersen's model and trained them to perform the same task using a more biologically plausible learning procedure than backpropagation. This procedure is a modification of the Associative Reward-Penalty (AR-P) algorithm (Barto and Anandan, 1985), which adjusts connection strengths using a global reinforcement signal and local synaptic information. Our networks learn to perform the task successfully to any degree of accuracy and almost as quickly as with backpropagation, and the hidden units develop response properties very similar to those of area 7a neurons. In particular, the probability of firing of the hidden units in our networks varies with eye position in a roughly planar fashion, and their visual receptive fields are large and have complex surfaces. The synaptic strengths computed by the AR-P algorithm are equivalent to and interchangeable with those computed by backpropagation. Our networks also perform the correct transformation on pairs of eye and retinal positions never encountered before. All of these findings are unaffected by the interposition of an extra layer of units between the hidden and output layers. These results show that the response properties of the hidden units of a layered network trained to perform coordinate transformations, and their similarity with those of area 7a neurons, are not a specific result of backpropagation training. The fact that they can be obtained by a more biologically plausible learning rule corroborates the validity of this neural network's computational algorithm as a plausible model of how area 7a may perform coordinate transformations.

80 citations


Book ChapterDOI
01 Jun 1991
TL;DR: It is demonstrated that certain classical problems associated with the notion of the “teacher― in supervised learning can be solved by judicious use of learned internal models as components of the adaptive system.
Abstract: Internal models of the environment have an important role to play in adaptive systems in general and are of particular importance for the supervised learning paradigm. In this paper we demonstrate that certain classical problems associated with the notion of the “teacher― in supervised learning can be solved by judicious use of learned internal models as components of the adaptive system. In particular, we show how supervised learning algorithms can be utilized in cases in which an unknown dynamical system intervenes between actions and desired outcomes.

33 citations


Proceedings ArticleDOI
26 Jun 1991
TL;DR: This work describes a multi-network, or modular, connectionist architecture that learns to perform control tasks using a piecewise control strategy, where the architecture's networks compete to learn the training patterns.
Abstract: Methodologies for designing piecewise control laws, such as gain scheduling, are useful because they circumvent the problem of determining a fixed global model of the plant dynamics. Instead, the dynamics are approximated using local models that vary with the plant's operating point. We describe a multi-network, or modular, connectionist architecture that learns to perform control tasks using a piecewise control strategy. The architecture's networks compete to learn the training patterns. As a result, a plant's parameter space is partitioned into a number of regions, and a different network learns a control law in each region.

24 citations


Proceedings Article
02 Dec 1991
TL;DR: A paradigm for modeling speech production based on neural networks based on characteristics of the musculoskeletal system is proposed and a cascade neural network is used to generate continuous motor commands from a sequence of discrete articulatory targets.
Abstract: We propose a paradigm for modeling speech production based on neural networks. We focus on characteristics of the musculoskeletal system. Using real physiological data - articulator movements and EMG from muscle activity - a neural network learns the forward dynamics relating motor commands to muscles and the ensuing articulator behavior. After learning, simulated perturbations, were used to asses properties of the acquired model, such as natural frequency, damping, and interarticulator couplings. Finally, a cascade neural network is used to generate continuous motor commands from a sequence of discrete articulatory targets.

14 citations