2019
DOI: 10.48550/arxiv.1906.08862
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Stored-program Memory

Hung Le,
Truyen Tran,
Svetha Venkatesh

Abstract: Neural networks powered with external memory simulate computer behaviors. These models, which use the memory to store data for a neural controller, can learn algorithms and other complex tasks. In this paper, we introduce a new memory to store weights for the controller, analogous to the stored-program memory in modern computer architectures. The proposed model, dubbed Neural Storedprogram Memory, augments current memory-augmented neural networks, creating differentiable machines that can switch programs throu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 14 publications
0
5
0
Order By: Relevance
“…We note that this problem of finding techniques for efficient storage and retrieval of memory elements is not new in the machine learning research community and there's been multiple proposals for achieving efficient storage and retrieval for different domains (Santoro et al, 2016;Gulcehre et al, 2017;Le et al, 2019). One of the simplest of these proposals is the Neural Turing Machine (NTM) and the memory addressing mechanism it proposes.…”
Section: Matrix Representation In Neural Memorymentioning
confidence: 99%
“…We note that this problem of finding techniques for efficient storage and retrieval of memory elements is not new in the machine learning research community and there's been multiple proposals for achieving efficient storage and retrieval for different domains (Santoro et al, 2016;Gulcehre et al, 2017;Le et al, 2019). One of the simplest of these proposals is the Neural Turing Machine (NTM) and the memory addressing mechanism it proposes.…”
Section: Matrix Representation In Neural Memorymentioning
confidence: 99%
“…Many neural network models take the form of a first order (in weights) recurrent neural network (RNN) and have been taught to learn context free and context-sensitive counter languages [17,9,5,64,70,56,48,66,8,36,8,67]. However, from a theoretical perspective, RNNs augmented with an external memory have historically been shown to be more capable of recognizing context free languages (CFLs), such as with a discrete stack [10,55,61], or, more recently, with various differentiable memory structures [33,26,24,39,73,28,72,25,40,41,3,42]. Despite positive results, prior work on CFLs was unable to achieve perfect generalization on data beyond the training dataset, highlighting a troubling difficulty in preserving long term memory.…”
Section: Related Workmentioning
confidence: 99%
“…In a more recent parallel thread in machine learning, various memory networks have been devised to augment traditional neural networks [Graves et al, 2014, Sukhbaatar et al, 2015, Munkhdalai et al, 2019, Le et al, 2019, Bartunov et al, 2019. Memory-augmented neural networks utilize a more stable external memory system analogous to computer memory, in contrast to more volatile storage mechanisms such as recurrent neural networks [Rodriguez et al, 2019].…”
Section: Introductionmentioning
confidence: 99%
“…Key-value networks date back to at least the 1980s with Sparse Distributed Memory (SDM) as a model of human long-term memory [Kanerva, 1988[Kanerva, , 1992. Inspired by random-access memory in computers, it is at the core of many memory networks recently developed in machine learning [Graves et al, 2014, Sukhbaatar et al, 2015, Banino et al, 2020, Le et al, 2019. A basic key-value network contains a key matrix K and a value matrix V .…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation