scispace - formally typeset
Open AccessJournal ArticleDOI

Modelling the N400 brain potential as change in a probabilistic representation of meaning.

TLDR
The authors provide a unified explanation of the N400 in a neural network model that avoids the commitments of traditional approaches to meaning in language and connects human language comprehension with recent deep learning approaches to language processing.
Abstract
The N400 component of the event-related brain potential has aroused much interest because it is thought to provide an online measure of meaning processing in the brain. However, the underlying process remains incompletely understood and actively debated. Here we present a computationally explicit account of this process and the emerging representation of sentence meaning. We simulate N400 amplitudes as the change induced by an incoming stimulus in an implicit and probabilistic representation of meaning captured by the hidden unit activation pattern in a neural network model of sentence comprehension, and we propose that the process underlying the N400 also drives implicit learning in the network. The model provides a unified account of 16 distinct findings from the N400 literature and connects human language comprehension with recent deep learning approaches to language processing.

read more

Content maybe subject to copyright    Report

1
I like coffee with cream and dog? Change in an implicit
probabilistic representation captures meaning processing in the
brain
Milena Rabovsky*, Steven S. Hansen, & James L. McClelland*
Department of Psychology, Stanford University
Word count: 10275
*Corresponding authors:
Milena Rabovsky (milena.rabovsky@gmail.com)
James L. McClelland (
mcclelland@stanford.edu
)
.CC-BY-NC-ND 4.0 International licensea
certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under
The copyright holder for this preprint (which was notthis version posted June 14, 2018. ; https://doi.org/10.1101/138149doi: bioRxiv preprint

2
Abstract
The N400 component of the event-related brain potential has aroused much interest
because it is thought to provide an online measure of meaning processing in the brain. Yet,
the underlying process remains incompletely understood and actively debated. Here, we
present a computationally explicit account of this process and the emerging representation of
sentence meaning. We simulate N400 amplitudes as the change induced by an incoming
stimulus in an implicit and probabilistic representation of meaning captured by the hidden
unit activation pattern in a neural network model of sentence comprehension, and we propose
that the process underlying the N400 also drives implicit learning in the network. The model
provides a unified account of 16 distinct findings from the N400 literature and connects
human language processing with successful deep learning approaches to language processing.
.CC-BY-NC-ND 4.0 International licensea
certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under
The copyright holder for this preprint (which was notthis version posted June 14, 2018. ; https://doi.org/10.1101/138149doi: bioRxiv preprint

3
I like coffee with cream and dog? Change in an implicit probabilistic representation
captures meaning processing in the brain
The N400 component of the event-related brain potential (ERP) has received a great
deal of attention, as it promises to shed light on the brain basis of meaning processing. The
N400 is a negative deflection recorded over centro-parietal areas peaking around 400 ms after
the presentation of a potentially meaningful stimulus. The first report of the N400 showed that
it occurred on presentation of a word violating expectations established by context: given “I
take my coffee with cream and the anomalous word dog produces a larger N400 than the
congruent word sugar
1
. Since this study, the N400 has been used as a dependent variable in
over 1000 studies and has been shown to be modulated by a wide range of variables including
sentence context, category membership, repetition, and lexical frequency, amongst others
2
.
However, despite the large amount of data on the N400, its functional basis continues to be
debated: various competing verbal descriptive theories have been proposed
3–8
, but their
capacity to capture all the relevant data is difficult to determine unambiguously due to the
lack of implementation, and none has yet offered a generally accepted account
2
.
Here, we provide both support for and formalization of the view that the N400 reflects
the input-driven update of a representation of sentence meaning – one that implicitly and
probabilistically represents all aspects of meaning as it evolves in real time during
comprehension
2
. We do so by presenting an explicit computational model of this process.
The model is trained and tested using materials generated by a simplified artificial
microworld (see below) in which we can manipulate variables that have been shown to affect
the N400, allowing us to explore how these factors affect processing. The use of these
synthetic materials prevents us from simulating N400 responses to the specific sentences used
in empirical experiments. Nevertheless, using these artificial materials, we are able to show
that the model can capture the effects of a broad range of factors on N400 amplitudes.
.CC-BY-NC-ND 4.0 International licensea
certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under
The copyright holder for this preprint (which was notthis version posted June 14, 2018. ; https://doi.org/10.1101/138149doi: bioRxiv preprint

4
The model does not exactly correspond to any existing account of the N400, as it
implements a distinct perspective on language comprehension. Existing accounts are often
grounded, at least in part, in modes of theorizing based on constructs originating in the
1950’s
9
, in which symbolic representations (e.g., of the meanings of words) are retrieved from
memory and subsequently integrated into a compositional representation – an annotated
structural description thought to serve as the representation of the meaning of a sentence
10–12
.
Even though perspectives on language processing have evolved in a variety of ways, many
researchers maintain the notion that word meanings are first retrieved from memory and
subsequently assigned to roles in a compositional representation. The account we offer here
does not employ these constructs and thus may contribute to the effort to rethink aspects of
several foundational issues: What does it mean to understand language? What are the
component parts of the process? Do we construct a structural description of a spoken
utterance in our mind, or do we more directly construct a representation of the speaker’s
meaning? Our work suggests different answers than those often given to these questions.
Our model, called the Sentence Gestalt (SG) model, was initially developed 30 years
ago
13,14
with the goal of illustrating how language understanding might occur without relying
on the traditional mode of theorizing described above. The model sought to offer a
functional-level characterization of language understanding in which each word in a sentence
someone hears or reads provides clues that constrain the formation of an implicit
representation of the event or situation described by the sentence. The initial work with the
model
14
established that it could capture several core aspects of language, including the ability
to resolve ambiguities of several kinds; to use word order and semantic constraints
appropriately; and to represent events described by sentences never seen during the network’s
training. A subsequent model using a similar approach successfully mastered a considerably
more complex linguistic environment
15
. The current work extending the SG model to address
N400 amplitudes complements efforts to model neurophysiological details underlying the
.CC-BY-NC-ND 4.0 International licensea
certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under
The copyright holder for this preprint (which was notthis version posted June 14, 2018. ; https://doi.org/10.1101/138149doi: bioRxiv preprint

5
N400
16–18
.
The design of the model (Fig. 1) reflects the principle that listeners continually update
their representation of the situation or event being described as each incoming word of a
sentence is presented. The representation is an internal representation (putatively
corresponding to a pattern of neural activity, modeled in an artificial neural network) called
the sentence gestalt (SG) that depends on connection-based knowledge in the update part of
the network. The SG pattern can be used to guide responses to potential queries about the
event or situation being described by the sentence (see Implicit probabilistic theory of
meaning section in online methods). The model is trained with sentences and queries about
the events the sentences describe, so that it can, if probed, provide responses to such queries.
Although we focus on a very simple microworld of events and sentences that can describe
them, the model exemplifies a wider conception of a neural activation state that represents a
persons subjective understanding of a broad range of situations and of the kinds of inputs that
can be used to update this understanding. The input could be in the form of language
expressing states of affairs (e.g., “Her hair is red.”) or even non-declarative language. For
example, the question “What time is it?” communicates that the speaker would like to know
the time. Though we focus only on linguistic input here, the input guiding the formation of
these representations could also come from witnessing events directly; from pictures, sounds,
or movies; or from any combination of linguistic or other forms of input.
The magnitude of the update produced by each successive word of a sentence
corresponds to the change in the model’s implicit representation that is produced by the word,
and it is this change, we propose, that is reflected in N400 amplitudes. Specifically, the
semantic update (SU) induced by the current word n is defined as the sum across the units in
the SG layer of the absolute value of the change in each unit’s activation produced by the
current word n. For a given unit (indexed below by the subscript i), the change is simply the
difference between the unit’s activation after word n and after word n-1:
.CC-BY-NC-ND 4.0 International licensea
certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under
The copyright holder for this preprint (which was notthis version posted June 14, 2018. ; https://doi.org/10.1101/138149doi: bioRxiv preprint

Figures
Citations
More filters

The ERP response to the amount of information conveyed by words in sentences (vol 140, pg 1, 2015)

TL;DR: The authors investigated whether event-related potentials (ERPs) too are predicted by information measures and found that different information measures quantify cognitively different processes and that readers do not make use of a sentence's hierarchical structure for generating expectations about the upcoming word.
Posted ContentDOI

The neural architecture of language: Integrative reverse-engineering converges on a model for predictive processing

TL;DR: It is found that the most powerful ‘transformer’ networks predict neural responses at nearly 100% and generalize across different datasets and data types (fMRI, ECoG), suggesting that inherent structure – and not just experience with language – crucially contributes to a model’s match to the brain.
Journal ArticleDOI

Multimodal Language Processing in Human Communication.

TL;DR: Cognitive mechanisms that may explain the binding of multiple, temporally offset signals under tight time constraints posed by a turn-taking system are proposed and called for a multimodal, situated psycholinguistic framework to unravel the full complexities of human language processing.
Journal ArticleDOI

Toward a Neurobiologically Plausible Model of Language-Related, Negative Event-Related Potentials.

TL;DR: A theoretical framework based on a predictive coding architecture is proposed, within which negative language-related ERP components such as the N400 can be accounted for in a neurobiologically plausible manner and suggests that latency and topography differences between these components reflect the locus of prediction errors and model updating within a hierarchically organized cortical predictive coding Architecture.
References
More filters
Book

Reinforcement Learning: An Introduction

TL;DR: This book provides a clear and simple account of the key ideas and algorithms of reinforcement learning, which ranges from the history of the field's intellectual foundations to the most recent developments and applications.
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Journal ArticleDOI

Finding Structure in Time

TL;DR: A proposal along these lines first described by Jordan (1986) which involves the use of recurrent links in order to provide networks with a dynamic memory and suggests a method for representing lexical categories and the type/token distinction is developed.
Journal ArticleDOI

A Neural Substrate of Prediction and Reward

TL;DR: Findings in this work indicate that dopaminergic neurons in the primate whose fluctuating output apparently signals changes or errors in the predictions of future salient and rewarding events can be understood through quantitative theories of adaptive optimizing control.
Book

Modularity of mind

Related Papers (5)
Frequently Asked Questions (8)
Q1. What is the only knowledge that is apparent in the model’s performance at the output layer?

The only knowledge that is apparent in the model’s performance at the output layer concerns the possible filler concepts for the agent role and their relative frequency, as well as a beginning tendency to activate the correct agent slightly more than the others. 

The query-answer form is used instead of directly providing the complete event description at the output layer to keep the set of probes and fillers more open-ended and to suggest the broader framework that the task of sentencecomprehension consists in building internal representations that can be used as a basis to respond to probes13. 

For type (2), changing position of agent and action, the conditional probability of the semantic features associated with the critical word (again, crucially, not at this position in the sentence but in general within the described event) is 1.0 in the condition with the changed word order and .4 in the condition with the normal word order. 

The authors assume that in reality, the adjustment of the semantic activation occurs continuously in time as auditory or visual language input is processed, so that the earliest arriving information about a word (whether auditory or visual) immediately influences the evolving SG representation64. 

a study found SV word order to be a valid cue to the agent role in 95/100 of sentences in English but only 35/100 sentences in Dutch41. 

0. Furthermore, the expected value of the summed divergence measure is 0 if the estimates match the probabilities for all C.Because the real learning environment is rich and probabilistic, the number of possiblesentences that may occur in the environment is indefinite, and it would not in general be possible to represent the estimates of the conditional probabilities explicitly (e.g. by listing them in a table). 

The authors used two-sided paired t-tests to analyze differences between conditions; when a simulation experiment involved more than one comparison, significance levels were Bonferroni-corrected within the simulation experiment. 

The authors also examined the model’s capacity to assign roles correctly when the reversalanomaly context (e.g., ‘the fox on the poacher’) was followed by a verb that it had experienced in such contexts during training (e.g. ‘watched’; see Supplementary Fig. 13 for details on the training environment).