scispace - formally typeset
Open Access

Infinite RAAM: A Principled Connectionist Basis for Grammatical Competence

Simon D. Levy, +2 more
- Vol. 22, Iss: 22
Reads0
Chats0
TLDR
This paper presents Infinite RAAM (IRAAM), a new fusion of recurrent neural networks with fractal geometry, allowing us to understand the behavior of these networks as dynamical sys- tems.
Abstract
Infinite RAAM: A Principled Connectionist Basis for Grammatical Competence Simon Levy, Ofer Melnik and Jordan Pollack levy, melnik, pollack@cs.brandeis.edu Dynamical and Evolutionary Machine Organization Volen Center for Complex Systems, Brandeis University, Waltham, MA 02454, USA February 6, 2000 Abstract This paper presents Infinite RAAM (IRAAM), a new fusion of recurrent neural networks with fractal geometry, allowing us to understand the behavior of these networks as dynamical sys- tems. Our recent work with IRAAMs has shown that they are capable of generating the context-free (non-regular) language for arbitrary values of . This paper expands upon that work, showing that IRAAMs are capable of generating syntac- tically ambiguous languages but seem less capable of gener- ating certain context-free constructions that are absent or dis- favored in natural languages. Together, these demonstrations support our belief that IRAAMs can provide an explanatorily adequate connectionist model of grammatical competence in natural language. Natural Language Issues In an early and extremely influential paper, Noam Chomsky (1956) showed that natural languages (NL’s) cannot be mod- eled by a finite-state automaton, because of the existence of center-embedded constructions. A second and equally im- portant observation from this work was that a minimally ade- quate NL grammar must be ambiguous, assigning more than one structure (interpretation) to some sentences, for example, They are flying planes. The first observation led to the development of Chomsky’s formal hierarchy of languages, based on the computational resources of the machines needed to recognize them. In this hierarchy, Chomsky’s observation about center-embedding is expressed by saying that NL’s are non-regular; i.e., they can- not be generated by having only rules of the form a grammar , where and are non-terminal symbols and is a terminal symbol. Whether NL’s are merely non-regular, belonging in the next, context-free (CF) level of the Chomsky hierarchy, or are more powerful, belonging further up in the hierarchy, became the subject of heated debate (Higginbotham 1984; Postal and Langendoen 1984; Shieber 1985). Non-CF phenomena such as reduplication/copying (Culy 1985) and crossed serial de- pendencies (Bresnan, Kaplan, Peters, and Zaenen 1982) sug- gested that a more powerful approach, using syntactic trans- formations (Chomsky 1957) was called for, but some re- searchers criticized transformations as having arbitrary power and thus failing to constrain the types of languages that could be expressed (Gazdar 1982). Further criticism of the entire formal approach came from observing that even CF gram- mars (CFGs) had the power to generate structures, such as a sequence followed by its mirror image, that did not seem to occur in NL (Manaster-Ramer 1986), or which placed an extraordinary burden on the human parsing mechanism when they did occur (Bach, Brown, and Marslen-Wilson 1986). Connectionism and Natural Language While debates about the complexity of NL were raging, connectionism was beginning to awaken from a fifteen-year sleep. In connectionist models many researchers found a way of embodying flexibility, graceful degradation, and other non-rigid properties that seem to characterize real cognitive systems like NL. This research culminated the publication of a highly controversial paper by Rumelhart and McClel- land (1986) which provided a connectionist account of part of the grammar of English using a feed-forward neural net- work. The paper was soon criticized by more traditional cog- nitive scientists (Fodor and Pylyshyn 1988; Pinker and Prince 1988), who cited the non-generative nature of such connec- tionist models as a fundamental shortcoming of the entire field. Partly in response to these criticisms, many connection- ists have spent the past decade investigating network models which support generativity through recurrent (feedback) con- nections (Lawrence, Giles, and Fong 1998; Rodriguez, Wiles, and Elman 1999; Williams and Zipser 1989). The research we present here is an attempt to contribute to this effort while focusing as strongly as possible on the natural language is- sues described above. Such an attempt faces a number of challenges. First, despite analysis of how a network’s dynamics con- tribute to its generativity, it is often uncertain whether the dynamics can support generation of well-formed strings be- yond a certain length. That is, it is unknown whether the net- work has a true “competence” for the language of which it has learned a few exemplars, or is merely capable of generating a finite, and hence regular, subset of the language. 1 Second, it is often easier to model weak, rather than strong genera- tive capacity, by building networks that generate or recognize strings having certain properties, without assigning any syn- tactic structure to the strings. Third, this lack of syntactic structure inhibits the formulation of an account of syntactic ambiguity in such networks, making them less plausible as models of NL. To be fair, not all connectionists, or cognitive scientists, take seriously the notion that human language has infinite generative ca- pacity. Though we obviously do not have the resources to argue the issue here, we are certain that a model with a provably infinite competence would be more persuasive to the cognitive science com- munity as a whole than would a model without one.

read more

Citations
More filters
Journal ArticleDOI

Creating hierarchical categories using cell assemblies

TL;DR: These simulations show that networks containing CAs can be used to learn hierarchical categories and how novel subcategories gain default information from their super-category.
Journal ArticleDOI

Mindless Intelligence

TL;DR: Artificial intelligence has stalled because of its preoccupation with simulating the human mind, but studying intelligence in natural systems, outside the mind, can reinvigorate the field.
Journal ArticleDOI

An instance of coincidence detection architecture relying on temporal coding

TL;DR: guided propagation networks (GPNs) are aimed at a generic real-time machine, based on time-space coincidence testing, and developed to address both cognitive and engineering applications.
Proceedings ArticleDOI

Analyzing state dynamics in a recurrent neural network

TL;DR: An analysis of state in simple recurrent networks is presented when they are trained to predict the outputs in a finite state grammar, and when they is required to generalize to a similar grammar.
Proceedings Article

On the Relationship between Symbolic and Neural Computation.

TL;DR: It is suggested that traditional context-free grammars are best understood as a special case of neural network computation; the special case derives its power from the presence of certain kinds of symmetries in the weight values.
References
More filters
Book

Fractals Everywhere

TL;DR: Focusing on how fractal geometry can be used to model real objects in the physical world, this up-to-date edition featurestwo 16-page full-color inserts, problems and tools emphasizing fractal applications, and an answers section.
Journal ArticleDOI

A learning algorithm for continually running fully recurrent neural networks

TL;DR: The exact form of a gradient-following learning algorithm for completely recurrent networks running in continually sampled time is derived and used as the basis for practical algorithms for temporal supervised learning tasks.
Journal ArticleDOI

Connectionism and cognitive architecture: a critical analysis

TL;DR: Differences between Connectionist proposals for cognitive architecture and the sorts of models that have traditionally been assumed in cognitive science are explored and the possibility that Connectionism may provide an account of the neural structures in which Classical cognitive architecture is implemented is considered.
Journal ArticleDOI

Three models for the description of language

TL;DR: It is found that no finite-state Markov process that produces symbols with transition from state to state can serve as an English grammar, and the particular subclass of such processes that produce n -order statistical approximations to English do not come closer, with increasing n, to matching the output of anEnglish grammar.
Journal ArticleDOI

On language and connectionism: analysis of a parallel distributed processing model of language acquisition

TL;DR: It is concluded that connectionists' claims about the dispensability of rules in explanations in the psychology of language must be rejected, and that, on the contrary, the linguistic and developmental facts provide good evidence for such rules.