2016
DOI: 10.1038/nature20101
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid computing using a neural network with dynamic external memory

Abstract: Artificial neural networks are remarkably adept at sensory processing, sequence learning and reinforcement learning, but are limited in their ability to represent variables and data structures and to store data over long timescales, owing to the lack of an external memory. Here we introduce a machine learning model called a differentiable neural computer (DNC), which consists of a neural network that can read from and write to an external memory matrix, analogous to the random-access memory in a conventional c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
1,025
0
8

Year Published

2017
2017
2023
2023

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 1,230 publications
(1,034 citation statements)
references
References 26 publications
1
1,025
0
8
Order By: Relevance
“…However, this could be reversed, giving a device that learns to construct context-free programs (e.g., expression trees) given only observed outputs; one application would be unsupervised parsing. Such an extension of the work would make it an alternative to architectures that have an explicit external memory such as neural Turing machines (Graves et al, 2014) and memory networks (Weston et al, 2015). However, as with those models, without supervision of the stack operations, formidable computational challenges must be solved (e.g., marginalizing over all latent stack operations), but sampling techniques and techniques from reinforcement learning have promise here (Zaremba and Sutskever, 2015), making this an intriguing avenue for future work.…”
Section: Resultsmentioning
confidence: 99%
“…However, this could be reversed, giving a device that learns to construct context-free programs (e.g., expression trees) given only observed outputs; one application would be unsupervised parsing. Such an extension of the work would make it an alternative to architectures that have an explicit external memory such as neural Turing machines (Graves et al, 2014) and memory networks (Weston et al, 2015). However, as with those models, without supervision of the stack operations, formidable computational challenges must be solved (e.g., marginalizing over all latent stack operations), but sampling techniques and techniques from reinforcement learning have promise here (Zaremba and Sutskever, 2015), making this an intriguing avenue for future work.…”
Section: Resultsmentioning
confidence: 99%
“…Nevertheless, TD-LSTM might not work well when the opinion word is far from the target, because the captured feature is likely to be lost ( reported similar problems of LSTM-based models in machine translation). (Graves et al, 2014) introduced the concept of memory for NNs and proposed a differentiable process to read and write memory, which is called Neural Turing Machine (NTM). Attention mechanism, which has been used successfully in many areas Rush et al, 2015), can be treated as a simplified version of NTM because the size of memory is unlimited and we only need to read from it.…”
Section: Related Workmentioning
confidence: 99%
“…This distinguishes the keyvariable memory from other memory-augmented neural networks that use continuous differentiable embeddings as the values of memory entries (Weston et al, 2014;Graves et al, 2016a).…”
Section: Entity Resolvermentioning
confidence: 99%