2019 International Joint Conference on Neural Networks (IJCNN) 2019
DOI: 10.1109/ijcnn.2019.8851909
|View full text |Cite
|
Sign up to set email alerts
|

Language Modeling through Long-Term Memory Network

Abstract: Recurrent Neural Networks (RNN), Long Short-Term Memory Networks (LSTM), and Memory Networks which contain memory are popularly used to learn patterns in sequential data. Sequential data has long sequences that hold relationships. RNN can handle long sequences but suffers from the vanishing and exploding gradient problems. While LSTM and other memory networks address this problem, they are not capable of handling long sequences (50 or more data points long sequence patterns). Language modelling requiring learn… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 12 publications
(5 citation statements)
references
References 64 publications
0
4
0
Order By: Relevance
“…Such a model could also improve preference recalling as the previous user history will be contained in a separate memory. A forgetting mechanism (e.g., similar to the work of Kirkpatrick et al, 2017 ; Nugaliyadde et al, 2019 ; Wang et al, 2020a ; Rae et al, 2020 ) can be introduced to remove or compress old memories for increasing the efficiency of memory retrieval and reducing catastrophic forgetting.…”
Section: Discussionmentioning
confidence: 99%
“…Such a model could also improve preference recalling as the previous user history will be contained in a separate memory. A forgetting mechanism (e.g., similar to the work of Kirkpatrick et al, 2017 ; Nugaliyadde et al, 2019 ; Wang et al, 2020a ; Rae et al, 2020 ) can be introduced to remove or compress old memories for increasing the efficiency of memory retrieval and reducing catastrophic forgetting.…”
Section: Discussionmentioning
confidence: 99%
“…However, they face challenges such as gradient vanishing or exploding when dealing with long sequences. These challenges result in a limited memory span and the loss of early input information [45]. In order to address these challenges, long short-term memory (LSTM) is specifically designed to handle long-term dependencies in time series data by three gates and a cell state [46].…”
Section: Bidirectional Long Short-term Memorymentioning
confidence: 99%
“…Deep neural network (DNN) is a popular ANN [33] and used for both mediumterm and long-term predictions. Recurrent neural network (RNN) and long shortterm memory (LSTM) network are the most used deep neural network (DNN), especially networks that adapt feedback loop from past inputs [49]. RNN and LSTM have surpassed other DNN models that do not employ feedback loops [50].…”
Section: Artificial Neural Networkmentioning
confidence: 99%