scispace - formally typeset
A

Andrew J. Anderson

Researcher at University of Rochester

Publications -  42
Citations -  1168

Andrew J. Anderson is an academic researcher from University of Rochester. The author has contributed to research in topics: Cognition & Semantic memory. The author has an hindex of 16, co-authored 41 publications receiving 821 citations. Previous affiliations of Andrew J. Anderson include University of Trento & Queen Mary University of London.

Papers
More filters
Journal ArticleDOI

Semantic Context Enhances the Early Auditory Encoding of Natural Speech

TL;DR: A novel approach is addressed using a recently introduced method for quantifying the semantic context of speech and relating it to a commonly used method for indexing low-level auditory encoding of speech to suggest a mechanism that links top-down prior information with bottom-up sensory processing in the context of natural, narrative speech listening.
Journal ArticleDOI

Predicting Neural Activity Patterns Associated with Sentences Using a Neurobiologically Motivated Model of Semantic Representation

TL;DR: The results show how a neurobiologically motivated semantic model can decompose sentence-level fMRI data into activation features for component words, which can be recombined to predict activation patterns for new sentences.
Journal ArticleDOI

Visually grounded and textual semantic models differentially decode brain activity associated with concrete and abstract nouns

TL;DR: This work applies state-of-the-art computational models to decode functional Magnetic Resonance Imaging activity patterns, elicited by participants reading and imagining a diverse set of both concrete and abstract nouns, and confirms that current computational models are sufficiently advanced to assist in investigating the representational structure of abstract concepts in the brain.
Journal ArticleDOI

Reading visually embodied meaning from the brain: Visually grounded computational models decode visual-object mental imagery induced by written text

TL;DR: By capturing latent visual-semantic structure their models provide a route into analyzing neural representations derived from past perceptual experience rather than stimulus-driven brain activity, and verify the benefit of combining multimodal data to model human-like semantic representations.