TL;DR: The objective of the current research is to outline a functional description of grammatical construction processing based on principles of psycholinguistics, develop a model of how these functions can be implemented in human neurophysiology, and demonstrate the feasibility of the resulting model in processing languages of typologically diverse natures, that is, English, French, and Japanese.
Abstract: One of the functions of everyday human language is to communicate meaning. Thus, when one hears or reads the sentence, “John gave a book to Mary,” some aspect of an event concerning the transfer of possession of a book from John to Mary is (hopefully) transmitted. One theoretical approach to language referred to as construction grammar emphasizes this link between sentence structure and meaning in the form of grammatical constructions. The objective of the current research is to (1) outline a functional description of grammatical construction processing based on principles of psycholinguistics, (2) develop a model of how these functions can be implemented in human neurophysiology, and then (3) demonstrate the feasibility of the resulting model in processing languages of typologically diverse natures, that is, English, French, and Japanese. In this context, particular interest will be directed toward the processing of novel compositional structure of relative phrases. The simulation results are discussed in the context of recent neurophysiological studies of language processing.
One of the long-term quests of cognitive neuroscience has been to link functional aspects of language processing to its underlying neurophysiology, that is, to understand how neural mechanisms allow the mapping of the surface structure of a sentence onto a conceptual representation of its meaning.
The current research takes this multidisciplinary approach within the theoretical framework associated with grammatical constructions.
These mappings vary along a continuum of complexity.
The universal grammar is there from the outset, and the challenge is to understand how it got there.
This issue is partially addressed in the current research.
Theoretical Framework for Grammatical Constructions
Figure 1C illustrates a more complex construction that contains an embedded relative phrase.
Part of the response to this lies in the specific ways in which function and content words are structurally organized in distinct sentence types.
The general idea here is that these constructions are templates into which a variety of open class elements (nouns, verbs, etc.) can be inserted in order to express novel meanings.
Several major issues can be raised with respect to this characterization.
Implementing Grammatical Constructions in a Neurocomputational Model
Determining the construction type for a given sentence consists in analyzing or recoding the sentence as a sequence of open class and closed class elements, and then performing sequence recognition on this recoded sequence.
The reordering in the form–meaning mapping in Figure 1B can be characterized as the abstract structure ABCD- Dominey, Hoen, and Inui 2089 BACD where A-D represent variable slots.
Given the initial validation of this model in the neurophysiological and neuropsychological studies cited above, the authors can now proceed with a more detailed study of how this can contribute to an understanding of the possible implementation of grammatical constructions, as illustrated in Figure 2.
As the sentence is processed word by word, a process of lexical categorization identifies open and closed class words.
In particular, the authors propose that the formation of the ConstructionIndex as a neural pattern of activity will rely on sequence processing in recurrent cortical networks, and that the retrieval of the FormToMeaning component will rely on a corticostriatal associative memory.
Goals of the Current Study
The authors have previously demonstrated that the model in Figure 2 can learn a variety of constructions that can then be used in different language interaction contexts (Dominey & Boucher, 2005a, b; Dominey, 2000).
The proposed model of grammatical construction processing should accommodate different types of languages.
That is, rather than looking for patterns at the level of the entire sentence or ConstructionIndex, the model could apply exactly the same mechanisms to identify subpatterns within the sentence.
This will decompose the sentence into the principal and the relative phrases, with the outcome that both will more likely correspond to existing constructions .
This type of segmentation approach has been explored by Miikkulainen (1996), discussed below.
METHODS
In the first set of experiments (Experiment 1A–C) for each of the three languages English, Japanese, and French, respectively, the authors will expose the model to a set of <sentence, meaning> pairs in a training phase.
The authors then validate that the model has learned these sentences by presenting the sentences alone and comparing the 2092 Journal of Cognitive Neuroscience Volume 18, Number 12 generated meanings with the actual meanings.
The authors compare the meaning generated for each sentence with its actual meaning.
In the second set of experiments (Experiment 2A–D) the authors attempt to determine whether pattern matching within a sentence can be used to extract relative phrases.
The net result is the capability to handle structurally novel sentences by decomposing them into their phrasal components in a construction grammar setting.
Model Overview
The model is trained on <sentence, meaning> pairs and then is tested with new sentences to generate the appropriate meanings.
Open class words are directed to the open class array (OCA), and are represented by an empty slot in the ConstructionIndex.
Given the FormToMeaning mapping for the current sentence, the authors can now store it in the ConstructionInventory, associated with the corresponding ConstructionIndex for that sentence.
Now let us consider how the system could accommodate complex sentences as in Figure 1C (and the relative and conjoined sentences in Table 4).
Thus, the authors now present the generalized compositional solution.
Processing Compositional Structure of Relative Phrases
As described above, the model will account for sentences with relative phrases by treating the entire sentence as a distinct construction, mapping the open class elements onto two distinct instances of the SceneEventArray as indicated in Figure 1C.
The pseudocode in Table 2 describes the process for extracting relative phrases.
Likewise, the authors demonstrated a hybrid neural network’s ability to fill working memory ‘‘slots’’ with specific sequence elements and then modulate the retrieval of the slot contents to transform these variable sequences for abstract sequence processing (Dominey & Ramus, 2000; Dominey et al., 1998) and syntactic comprehension (Dominey et al., 2003; Dominey, 2002).
For each of these test sentences, the model successfully identified the corresponding construction, applied the corresponding sentence-to-meaning mapping, and generated the correct meaning.
An example of this word order flexibility of Japanese with respect to English is illustrated in Table 5 with the English passive ditransitive forms that can be.
Language English French Japanese
Dominey, Hoen, and Inui 2095 expressed in four different common manners in Japanese (Constructions 9–12).
Indeed, the model successfully discriminates between all of the 26 construction types based on the ConstructionIndex unique to each construction type, and associates the correct FormToMeaning mapping with each of them.
This demonstration with Japanese is an important validation that at least for this subset of constructions, the construction-based model is applicable both to fixed word order languages such as English, as well as free word order languages such as Japanese.
This also provides further validation for the proposal of Bates et al. (1982) and MacWhinney (1982) that thematic roles are indicated by a constellation of cues including grammatical markers and word order.
As for the English and Japanese studies, the model was trained on a set of French sentences (illustrated in Table 6) that were paired with their meanings.
Example Sentences and Meanings Grammatical Constructions
Push(block, cylinder) Verb(agent, object) 2. The cylinder was pushed by the block.
The block that pushed the cylinder touched the moon.
Agent1 that verb1ed object2 verb2ed object3.
The cylinder that was pushed by the block gave the cat to the dog.
Dual-event Conjoined Constructions
The block pushed the cylinder and the moon.
The moon and the block were given to the cylinder by the cat.
In the previous experiments, for each of the three languages, a limited set of sentences including those with relative phrases was learned, based on a noncompositional mechanism.
For these sentences, the verbs in the main and relative phrases in the OpenClassArray became associated with the event components of two distinct copies of the SceneEventArray, corresponding to constructions like that in Figure 1C.
The chief pushed the boy and the block.
For the English experiments, the model was first trained on five sentence types corresponding to the active and passive transitive and dative constructions (1–4 in Table 4), and the relative ‘‘The boy the dog hit,’’ along with their corresponding meaning representations.
Similarly for Sentence 2, application of the phrase structure extraction yielded three phrases: ‘‘The dog bit the man,’’ ‘‘The cat hit the dog,’’ and ‘‘The boy saw the cat,’’ for each of which the corresponding meanings were extracted based on the previously learned constructions.
Cat wo relverb boy ga dog ni-yotte osareta Girl-ga girl-wo bit dog-ga bit cat-wo chased boywo saw.
Psychological Reality of the Model
The first principle inherent in the model is that instead of representing <sentence, meaning> mappings in terms of a generative grammar, they are represented directly in a structured inventory of grammatical constructions that are nothing more than these mappings (Tomasello, 2003; Goldberg, 1995).
The essential innate capabilities the authors postulate are mechanisms for (1) abstract (variable based) structure mapping and for 2100 Journal of Cognitive Neuroscience Volume 18, Number 12 (2) pattern-based embedded structure processing.
What has been lacking, however, is a specification of how grammatical constructions can be combined in order to generate new constructions on the fly, using these same sequence processing mechanisms.
This approach should extend to accommodate more generalized relative phrase structure, as well as prepositional and other categories of embedded phrases.
Thus, when variations in word order are used in order to place the pragmatic focus on a particular item, as in the Japanese examples 9–12 in Table 5, distinct constructions will be employed to capture these differences.
Division of Labor in the Construction Inventory
Construction grammar assumes no strict division between the lexicon and syntax in that both lexical and syntactic constructions pair form with meaning (see Goldberg, 1995).
Subsequent developmental categorization of an open class category of concrete nouns would allow verb-based constructions such as ‘‘Gimme X.’’.
A pertinent account of such distributional analysis is presented in Cartwright and Brent (1997).
The current model represents what the authors consider a necessary refinement that is wholly consistent with construction grammar, and indeed provides insight into a consistent explanation for compositionality.
Temporal Dynamics and Resolution of Detail
The theoretical foundation of the model is that the sentence-to-meaning mapping is performed based on the configuration of closed class elements (the ConstructionIndex) in the sentence (consistent with Bates et al., 1982; MacWhinney, 1982).
In the current implementation, to reduce processing complexity, the authors wait until the entire sentence has been seen before extracting the corresponding mapping.
Miyamoto (2002) observed that in some cases, marking on the noun phrases indicates phrase boundaries prior to the arrival of the verb, so that the detection of the relative phrase structure can actually precede the nonfinal verb (Miyamoto, 2002).
In general, if delay is decreased, then the required information must come from an alternative source (such as the double o constraint).
Thus, in building up the ConstructionIndex incrementally, when such information is not available then the incremental processing will induce temporary failures including well-documented garden path effects.
Underlying Neurophysiology
Part of the importance of the current approach is that the underlying primitive functions that provide this construction processing capability can be linked in a plausible manner to known neurophysiological systems.
The resulting pattern of cortical activity projects to the caudate nucleus of the striatum, and via these modifiable corticostriatal synapses, retrieves the appropriate FormToMeaning mapping that is then implemented in the frontal transformation processing network that includes BA 44.
The authors have simulated how a recurrent cortical network can encode the sequential structure of closed and open class words in a sentence, corresponding to the ConstructionIndex, in order to retrieve the appropriate FormToMeaning mapping based on corticostriatal associative memory (Dominey et al., 2003; Dominey, 2002, 2005a, 2005b).
This result is consistent with previous fMRI studies that showed activation of these areas in sentence comprehension tasks (e.g., Kuperberg et al., 2003; Baumgaertner, Weiller, & Büchel, 2002), particularly for sequential or structural processing aspects in sentences (Newman, Just, Keller, Roth, & Carpenter, 2003; Dapretto & Bookheimer, 1999; Kang, Constable, Gore, & Avrutin, 1999; Inui et al., 1998).
This yielded a profile of sentence comprehension performance highly correlated (R2 = .89) with that of the nine agrammatic patients the authors studied (Dominey, 2002).
Related Neurocomputational Parsing Models
The current proposal is consistent with and complimentary to several existing neurocomputational models of language processing.
Grodzinsky (2000) argues that Broca’s area is involved in a more specialized form of syntactic transformation processing.
These views correspond rather well with their characterization of BA 44 for processing structural transformations of linguistic and nonlinguistic sequences, and BA 45 for inserting semantic content into this mapping process for thematic role assignment.
The authors further suggest that this transformation processing is guided by templates that are extracted from a corticostriatal memory system, based on the configuration of closed class elements coded in a recurrent corticocortical network (corresponding to the ConstructionIndex).
It is noteworthy that syntactic processing in language production activates cortical regions adjacent to and overlapping BA 44 (Indefrey et al., 2001).
Extraction of Relative Phrase Structure
Part of the novelty of their approach is that the extraction of relative phrase structure relies on these same capabilities.
Each extracted phrase is processed by the construction model, and then the remainder of the original sequence is processed.
At the extreme, this is related to lexically driven approaches including that of Vosse and Kempen (2000), lexical functional grammar (LFG; Bresnan, 2001), and head-driven phrase structure grammar (HPSG; Pollard & Sag, 1994), in which lexical entries include type descriptions that form a hierarchy that, combined with grammatical rules, allows for unificationbased parsing.
One source of variability is the structure of the noun phrase.
Acknowledgments
This study was supported in part by the HFSP Organization, the French ACI NIC and TTT Projects, LAFMI, and the ESF.
Reprint requests should be sent to Peter Ford Dominey, Institut des Sciences Cognitives, CNRS UMR 5015, 67, Boulevard Pinel, 69675 Bron Cedex, France, or via e-mail: dominey@isc.cnrs.fr and www.isc.cnrs.fr/.
TL;DR: This review systematically surveys both current ways of generating/adapting the reservoirs and training different types of readouts, and offers a natural conceptual classification of the techniques, which transcends boundaries of the current ''brand-names'' of reservoir methods.
Abstract: Echo State Networks and Liquid State Machines introduced a new paradigm in artificial recurrent neural network (RNN) training, where an RNN (the reservoir) is generated randomly and only a readout is trained. The paradigm, becoming known as reservoir computing, greatly facilitated the practical application of RNNs and outperformed classical fully trained RNNs in many tasks. It has lately become a vivid research field with numerous extensions of the basic idea, including reservoir adaptation, thus broadening the initial paradigm to using different methods for training the reservoir and the readout. This review systematically surveys both current ways of generating/adapting the reservoirs and training different types of readouts. It offers a natural conceptual classification of the techniques, which transcends boundaries of the current ''brand-names'' of reservoir methods, and thus aims to help in unifying the field and providing the reader with a detailed ''map'' of it.
TL;DR: Networks involving the temporal cortex and the inferior frontal cortex with a clear left lateralization were shown to support syntactic processes, whereas less lateralized temporo-frontal networks subserve semantic processes.
Abstract: Language processing is a trait of human species. The knowledge about its neurobiological basis has been increased considerably over the past decades. Different brain regions in the left and right hemisphere have been identified to support particular language functions. Networks involving the temporal cortex and the inferior frontal cortex with a clear left lateralization were shown to support syntactic processes, whereas less lateralized temporo-frontal networks subserve semantic processes. These networks have been substantiated both by functional as well as by structural connectivity data. Electrophysiological measures indicate that within these networks syntactic processes of local structure building precede the assignment of grammatical and semantic relations in a sentence. Suprasegmental prosodic information overtly available in the acoustic language input is processed predominantly in a temporo-frontal network in the right hemisphere associated with a clear electrophysiological marker. Studies with patients suffering from lesions in the corpus callosum reveal that the posterior portion of this structure plays a crucial role in the interaction of syntactic and prosodic information during language processing.
1,306 citations
Cites methods from "A Neurolinguistic Model of Grammati..."
...Based on their data, they proposed the following functional subdivision of Broca’s area and its adjacent regions in the prefrontal cortex (54, 121): superior and posterior regions (BA 6/9/46/44) are engaged in sequential and structural aspects of processing, whereas anterior and inferior regions (BA 11/47/45) are implicated in context information insertion into structural matrixes selected by the upper regions....
TL;DR: The results showed that listening to action-related sentences activates a left fronto-parieto-temporal network that includes the pars opercularis of the inferior frontal gyrus (Broca's area), as well as the inferior parietal lobule, the intraparietal sulcus, and the posterior middle temporal gyrus.
Abstract: Observing actions made by others activates the cortical circuits responsible for the planning and execution of those same actions This observation–execution matching system (mirror-neuron system) is thought to play an important role in the understanding of actions made by others In an fMRI experiment, we tested whether this system also becomes active during the processing of action-related sentences Participants listened to sentences describing actions performed with the mouth, the hand, or the leg Abstract sentences of comparable syntactic structure were used as control stimuli The results showed that listening to action-related sentences activates a left fronto-parieto-temporal network that includes the pars opercularis of the inferior frontal gyrus (Broca's area), those sectors of the premotor cortex where the actions described are motorically coded, as well as the inferior parietal lobule, the intraparietal sulcus, and the posterior middle temporal gyrus These data provide the first direct evidence that listening to sentences that describe actions engages the visuomotor circuits which subserve action execution and observation
TL;DR: Cognitive developmental robotics aims to provide new understanding of how human's higher cognitive functions develop by means of a synthetic approach that developmentally constructs cognitive functions through interactions with the environment, including other agents.
Abstract: Cognitive developmental robotics (CDR) aims to provide new understanding of how human's higher cognitive functions develop by means of a synthetic approach that developmentally constructs cognitive functions. The core idea of CDR is ldquophysical embodimentrdquo that enables information structuring through interactions with the environment, including other agents. The idea is shaped based on the hypothesized development model of human cognitive functions from body representation to social behavior. Along with the model, studies of CDR and related works are introduced, and discussion on the model and future issues are argued.
TL;DR: No studies have satisfied all the criteria necessary to support the more specific neurobiological claims made by the two hypotheses-namely, that each level of verb meaning is associated with mirror neurons in the pertinent brain regions, which would require demonstrating that within those regions the same neuronal populations are engaged during (a) the linguistic processing of particular motor features of verb mean, (b) the execution of actions with the corresponding motor features, and (c) the observation ofactions with the equivalent motor features.
Abstract: Verbs have two separate levels of meaning. One level reflects the uniqueness of every verb and is called the "root". The other level consists of a more austere representation that is shared by all the verbs in a given class and is called the "event structure template". We explore the following hypotheses about how, with specific reference to the motor features of action verbs, these two distinct levels of semantic representation might correspond to two distinct levels of the mirror neuron system. Hypothesis 1: Root-level motor features of verb meaning are partially subserved by somatotopically mapped mirror neurons in the left primary motor and/or premotor cortices. Hypothesis 2: Template-level motor features of verb meaning are partially subserved by representationally more schematic mirror neurons in Brodmann area 44 of the left inferior frontal gyrus. Evidence has been accumulating in support of the general neuroanatomical claims made by these two hypotheses-namely, that each level of verb meaning is associated with the designated cortical areas. However, as yet no studies have satisfied all the criteria necessary to support the more specific neurobiological claims made by the two hypotheses-namely, that each level of verb meaning is associated with mirror neurons in the pertinent brain regions. This would require demonstrating that within those regions the same neuronal populations are engaged during (a) the linguistic processing of particular motor features of verb meaning, (b) the execution of actions with the corresponding motor features, and (c) the observation of actions with the corresponding motor features.
190 citations
Cites methods from "A Neurolinguistic Model of Grammati..."
...This is analogous to Dominey et al.’s (2006) structure mapping process, since both...
TL;DR: This book presents the most complete exposition of the theory of head-driven phrase structure grammar, introduced in the authors' "Information-Based Syntax and Semantics," and demonstrates the applicability of the HPSG approach to a wide range of empirical problems.
Abstract: This book presents the most complete exposition of the theory of head-driven phrase structure grammar (HPSG), introduced in the authors' "Information-Based Syntax and Semantics." HPSG provides an integration of key ideas from the various disciplines of cognitive science, drawing on results from diverse approaches to syntactic theory, situation semantics, data type theory, and knowledge representation. The result is a conception of grammar as a set of declarative and order-independent constraints, a conception well suited to modelling human language processing. This self-contained volume demonstrates the applicability of the HPSG approach to a wide range of empirical problems, including a number which have occupied center-stage within syntactic theory for well over twenty years: the control of "understood" subjects, long-distance dependencies conventionally treated in terms of "wh"-movement, and syntactic constraints on the relationship between various kinds of pronouns and their antecedents. The authors make clear how their approach compares with and improves upon approaches undertaken in other frameworks, including in particular the government-binding theory of Noam Chomsky.
TL;DR: In this paper, the interaction between verbs and constructions is discussed, and relations among constructions are investigated in the context of English Ditransitive construction and English Caused-Motion construction.
Abstract: Acknowledgments 1: Introduction 2: The Interaction between Verbs and Constructions 3: Relations among Constructions 4: On Linking 5: Partial Productivity 6: The English Ditransitive Construction 7: The English Caused-Motion Construction 8: The English Resultative Construction 9: The Way Construction 10: Conclusion Notes Bibliography Index
TL;DR: The authors argue that the essence of language is its symbolic dimension, which rests on the uniquely human ability to comprehend intention, and that children pick up these patterns in the buzz of words they hear around them.
Abstract: Drawing together a vast body of empirical research in cognitive science, linguistics, and developmental psychology, Michael Tomasello demonstrates that we don't need a self-contained "language instinct" to explain how children learn language. Their linguistic ability is interwoven with other cognitive abilities. Tomasello argues that the essence of language is its symbolic dimension, which rests on the uniquely human ability to comprehend intention. Grammar emerges as the speakers of a language create linguistic constructions out of recurring sequences of symbols, children pick up these patterns in the buzz of words they hear around them. Constructing a Language offers a compellingly argued, psychologically sound new vision for the study of language acquisition.
TL;DR: This review argues that sentence processing is supported by a temporo-frontal network, within this network, temporal regions subserve aspects of identification and frontal regions the building of syntactic and semantic relations.
Abstract: Functional dissociations within the neural basis of auditory sentence processing are difficult to specify because phonological, syntactic and semantic information are all involved when sentences are perceived. In this review I argue that sentence processing is supported by a temporo-frontal network. Within this network, temporal regions subserve aspects of identification and frontal regions the building of syntactic and semantic relations. Temporal analyses of brain activation within this network support syntax-first models because they reveal that building of syntactic structure precedes semantic processes and that these interact only during a later stage.
TL;DR: It is proposed that "language" disorders, such as specific language impairment and non-fluent and fluent aphasia, may be profitably viewed as impairments primarily affecting one or the other brain system, and suggested a new neurocognitive framework for the study of lexicon and grammar.
Abstract: The structure of the brain and the nature of evolution suggest that, despite its uniqueness, language likely depends on brain systems that also subserve other functions. The declarative/procedural (DP) model claims that the mental lexicon of memorized word-specific knowledge depends on the largely temporal-lobe substrates of declarative memory, which underlies the storage and use of knowledge of facts and events. The mental grammar, which subserves the rule-governed combination of lexical items into complex representations, depends on a distinct neural system. This system, which is composed of a network of specific frontal, basal-ganglia, parietal and cerebellar structures, underlies procedural memory, which supports the learning and execution of motor and cognitive skills, especially those involving sequences. The functions of the two brain systems, together with their anatomical, physiological and biochemical substrates, lead to specific claims and predictions regarding their roles in language. These predictions are compared with those of other neurocognitive models of language. Empirical evidence is presented from neuroimaging studies of normal language processing, and from developmental and adult-onset disorders. It is argued that this evidence supports the DP model. It is additionally proposed that "language" disorders, such as specific language impairment and non-fluent and fluent aphasia, may be profitably viewed as impairments primarily affecting one or the other brain system. Overall, the data suggest a new neurocognitive framework for the study of lexicon and grammar.