2018
DOI: 10.1007/s00521-017-3322-z
|View full text |Cite
|
Sign up to set email alerts
|

Character-level recurrent neural networks in practice: comparing training and sampling schemes

Abstract: Recurrent neural networks are nowadays successfully used in an abundance of applications, going from text, speech and image processing to recommender systems. Backpropagation through time is the algorithm that is commonly used to train these networks on specific tasks. Many deep learning frameworks have their own implementation of training and sampling procedures for recurrent neural networks, while there are in fact multiple other possibilities to choose from and other parameters to tune. In existing literatu… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 13 publications
0
3
0
Order By: Relevance
“…Another recurrencebased approach to ERC is called COSMIC (Ghosal et al, 2020), a knowledge-based model that is related to DialogueRNN in its network structure and that adds information about, e.g., causal relations, mental states, and actions to improve performance. Even though these recurrence-based models can in theory handle infinitely long sequences, in practice long-term contextual information is not always propagated due to vanishing gradients and practical limitations in recurrent depth when applying backpropagation-through-time (Pascanu et al, 2013;De Boom et al, 2019).…”
Section: Machine Learning Approaches To Ercmentioning
confidence: 99%
“…Another recurrencebased approach to ERC is called COSMIC (Ghosal et al, 2020), a knowledge-based model that is related to DialogueRNN in its network structure and that adds information about, e.g., causal relations, mental states, and actions to improve performance. Even though these recurrence-based models can in theory handle infinitely long sequences, in practice long-term contextual information is not always propagated due to vanishing gradients and practical limitations in recurrent depth when applying backpropagation-through-time (Pascanu et al, 2013;De Boom et al, 2019).…”
Section: Machine Learning Approaches To Ercmentioning
confidence: 99%
“…analysis, text analysis, speech recognition, and other fields, providing solutions to a variety of real-world problems [Sun et al 2021]. The basic knowledge of transfer learning, numerous types of methodologies utilised to accomplish transfer learning, as well as how transfer learning was applied in many subfields of medical image analysis were all evaluated by the authors [Boom et al 2019]. The review demonstrates that recent developments in DL, particularly advances in transfer learning, have enabled the identification, classification, and quantification of specific patterns from a large number of medical pictures [Hwang and Sung, 2017].…”
Section: Theoretical Backgroundmentioning
confidence: 99%
“…An ensemble of deep convolutional neural networks was also used by Reference [ 9 ] in a similar task; however, it was a single-label classification problem. A single-label approach was also presented in Reference [ 10 ], where the authors explored the possibility of generating hashtags for an input image and leveraged it to generate meaningful anecdotes connected to the essence of the image by applying a character-level Recurrent (RNN) Neural Network [ 11 , 12 ].…”
Section: Introductionmentioning
confidence: 99%