2020
DOI: 10.48550/arxiv.2006.01981
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Training End-to-End Analog Neural Networks with Equilibrium Propagation

Jack Kendall,
Ross Pantone,
Kalpana Manickavasagam
et al.

Abstract: We introduce a principled method to train end-to-end analog neural networks by stochastic gradient descent. In these analog neural networks, the weights to be adjusted are implemented by the conductances of programmable resistive devices such as memristors [Chua, 1971], and the nonlinear transfer functions (or 'activation functions') are implemented by nonlinear components such as diodes. We show mathematically that a class of analog neural networks (called nonlinear resistive networks) are energy-based models… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
39
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 23 publications
(39 citation statements)
references
References 32 publications
(55 reference statements)
0
39
0
Order By: Relevance
“…We will now separate the Lagrangian into sets of terms which will aid in our analysis. We assume the network takes the form of a deep neural network, similar to our previous work [37], with layers of nonlinear, dynamic subcircuits (neurons) separated by fully connected layers of linear resistors (synapses).…”
Section: Derivation Of the Gradientmentioning
confidence: 99%
“…We will now separate the Lagrangian into sets of terms which will aid in our analysis. We assume the network takes the form of a deep neural network, similar to our previous work [37], with layers of nonlinear, dynamic subcircuits (neurons) separated by fully connected layers of linear resistors (synapses).…”
Section: Derivation Of the Gradientmentioning
confidence: 99%
“…The fundamental status of the memristor, however, has remained contentious [45]. Since the 2008 discovery, interest in the memristor has surged, primarily due to its potential in neuromorphic computers and analog neural networks [27], [29]- [32]. We note that the invention of the memristor by Chua [44] follows the first paper of monotone circuits [16] by a decade, but that, to the best of the authors' knowledge, all the system-theoretic investigations of the memristor have focused on passivity rather than monotonicity, although Theorem 1 of Chua [44] can be interpreted as showing that monotonicity is equivalent to passivity for a memristor.…”
Section: B 1-port Elementsmentioning
confidence: 99%
“…This line of research exploits the algorithmic significance of maximal monotonicity, but has progressively become detached from its physical significance. On the other hand, recent years have witnessed a surge of interest in the physical significance of memristive nonlinear circuits [26]- [32], but with little emphasis on algorithms and system analysis.…”
Section: Introductionmentioning
confidence: 99%
“…In order for such physical networks to learn on their own, they cannot minimize an arbitrary cost function by gradient descent since that is a global process that requires knowing all the microscopic details at once, carrying out the global computation of gradient descent, and then manipulating networks at the microscopic (node or edge) level. Rather, approaches such as contrastive learning (1)(2)(3)(4), equilibrium propagation (5,6), directed aging (7)(8)(9)(10)(11)(12) and coupled learning (13,14) use local rules, in which learning degrees of freedom (e.g. the conductances of edges in electrical networks of variable resistors) respond to physical degrees of freedom (e.g.…”
Section: Introductionmentioning
confidence: 99%