scispace - formally typeset
Journal ArticleDOI

Agnostic Physics-Driven Deep Learning

Benjamin Scellier, +3 more
- 30 May 2022 - 
- Vol. abs/2205.15021
Reads0
Chats0
TLDR
This work establishes that a physical system can perform statistical learning without gradient computations, via an Agnostic Equilibrium Propagation procedure that combines energy minimization, homeostatic control, and nudging towards the correct response.
Abstract
This work establishes that a physical system can perform statistical learning without gradient computations, via an Agnostic Equilibrium Propagation (Æqprop) procedure that combines energy minimization, homeostatic control, and nudging towards the correct response. In Æqprop, the specifics of the system do not have to be known: the procedure is based only on external manipulations, and produces a stochastic gradient descent without explicit gradient computations. Thanks to nudging, the system performs a true, order-one gradient step for each training sample, in contrast with order-zero methods like reinforcement or evolutionary strategies, which rely on trial and error. This procedure considerably widens the range of potential hardware for statistical learning to any system with enough con-trollable parameters, even if the details of the system are poorly known. Æqprop also establishes that in natural (bio)physical systems, genuine gradient-based statistical learning may result from generic, relatively simple mechanisms, without backpropagation and its requirement for analytic knowledge of partial derivatives.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Beyond Backpropagation: Bilevel Optimization Through Implicit Differentiation and Equilibrium Propagation

TL;DR: This review examines gradient-based techniques to solve bilevel optimization problems, leveraging the toolbox of implicit differentiation and, for the first time applied to this setting, the equilibrium propagation theorem.
Journal ArticleDOI

Frequency propagation: Multi-mechanism learning in nonlinear physical networks

TL;DR: In this article , the authors introduce frequency propagation, a learning algorithm for nonlinear physical networks, in which an activation current is applied at a set of input nodes at one frequency, and an error current was applied at output nodes at another frequency, each conductance is updated proportionally to the product of the two coefficients.
References
More filters

Automatic differentiation in PyTorch

TL;DR: An automatic differentiation module of PyTorch is described — a library designed to enable rapid research on machine learning models that focuses on differentiation of purely imperative programs, with a focus on extensibility and low overhead.

Gradient-based learning applied to document recognition

TL;DR: This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task, and Convolutional neural networks are shown to outperform all other techniques.
Journal ArticleDOI

Memristor-The missing circuit element

TL;DR: In this article, the memristor is introduced as the fourth basic circuit element and an electromagnetic field interpretation of this relationship in terms of a quasi-static expansion of Maxwell's equations is presented.
Posted Content

Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms

TL;DR: Fashion-MNIST is intended to serve as a direct drop-in replacement for the original MNIST dataset for benchmarking machine learning algorithms, as it shares the same image size, data format and the structure of training and testing splits.
Book ChapterDOI

Large-Scale Machine Learning with Stochastic Gradient Descent

Léon Bottou
TL;DR: A more precise analysis uncovers qualitatively different tradeoffs for the case of small-scale and large-scale learning problems.