scispace - formally typeset
Open AccessPosted Content

A mathematical theory of semantic development in deep neural networks

Reads0
Chats0
TLDR
Notably, this simple neural model qualitatively recapitulates many diverse regularities underlying semantic development, while providing analytic insight into how the statistical structure of an environment can interact with nonlinear deep-learning dynamics to give rise to these regularities.
Abstract
An extensive body of empirical research has revealed remarkable regularities in the acquisition, organization, deployment, and neural representation of human semantic knowledge, thereby raising a fundamental conceptual question: what are the theoretical principles governing the ability of neural networks to acquire, organize, and deploy abstract knowledge by integrating across many individual experiences? We address this question by mathematically analyzing the nonlinear dynamics of learning in deep linear networks. We find exact solutions to this learning dynamics that yield a conceptual explanation for the prevalence of many disparate phenomena in semantic cognition, including the hierarchical differentiation of concepts through rapid developmental transitions, the ubiquity of semantic illusions between such transitions, the emergence of item typicality and category coherence as factors controlling the speed of semantic processing, changing patterns of inductive projection over development, and the conservation of semantic similarity in neural representations across species. Thus, surprisingly, our simple neural model qualitatively recapitulates many diverse regularities underlying semantic development, while providing analytic insight into how the statistical structure of an environment can interact with nonlinear deep learning dynamics to give rise to these regularities.

read more

Citations
More filters
Journal ArticleDOI

If deep learning is the answer, what is the question?

TL;DR: A road map of how neuroscientists can use deep networks to model and understand biological brains is offered to offer a road map for systems neuroscience research in the age of deep learning.
Journal ArticleDOI

Artificial Neural Networks for Neuroscientists: A Primer.

TL;DR: This pedagogical Primer introduces artificial neural networks and demonstrates how they have been fruitfully deployed to study neuroscientific questions, and details how to customize the analysis, structure, and learning of ANNs to better address a wide range of challenges in brain research.
Posted Content

Gradient Starvation: A Learning Proclivity in Neural Networks

TL;DR: This work provides a theoretical explanation for the emergence of feature imbalance in neural networks and develops guarantees for a novel regularization method aimed at decoupling feature learning dynamics, improving accuracy and robustness in cases hindered by gradient starvation.
Posted Content

Implicit Regularization of Discrete Gradient Dynamics in Linear Neural Networks.

TL;DR: This work studies the discrete gradient dynamics of the training of a two-layer linear network with the least-squares loss using a time rescaling to show that this dynamics sequentially learns the solutions of a reduced-rank regression with a gradually increasing rank.
Journal ArticleDOI

Modeling the Influence of Data Structure on Learning in Neural Networks: The Hidden Manifold Model

TL;DR: A generative model for structured data sets that is to construct high-dimensional inputs that lie on a lower-dimensional manifold, with labels that depend only on their position within this manifold, akin to a single layer decoder or generator in a generative adversarial network is introduced.
References
More filters
Journal ArticleDOI

Reducing the Dimensionality of Data with Neural Networks

TL;DR: In this article, an effective way of initializing the weights that allows deep autoencoder networks to learn low-dimensional codes that work much better than principal components analysis as a tool to reduce the dimensionality of data is described.
Journal ArticleDOI

Basic objects in natural categories

TL;DR: In this paper, the authors define basic objects as those categories which carry the most information, possess the highest category cue validity, and are the most differentiated from one another, and thus the most distinctive from each other.
Journal ArticleDOI

Family Resemblances: Studies in the Internal Structure of Categories

TL;DR: In this paper, the authors explored the hypothesis that the members of categories which are considered most prototypical are those with most attributes in common with other members of the category and least attributes with other categories and found that family resemblance offers an alternative to criterial features in defining categories.
Journal ArticleDOI

Distributed and Overlapping Representations of Faces and Objects in Ventral Temporal Cortex

TL;DR: The functional architecture of the object vision pathway in the human brain was investigated using functional magnetic resonance imaging to measure patterns of response in ventral temporal cortex while subjects viewed faces, cats, five categories of man-made objects, and nonsense pictures, and a distinct pattern of response was found for each stimulus category.
Related Papers (5)