2016
DOI: 10.1007/978-3-319-51469-7_8
|View full text |Cite
|
Sign up to set email alerts
|

Comparing Hidden Markov Models and Long Short Term Memory Neural Networks for Learning Action Representations

Abstract: Abstract. In this paper we are concerned with learning models of actions and compare a purely generative model based on Hidden Markov Models to a discriminatively trained recurrent LSTM network in terms of their properties and their suitability to learn and represent models of actions. Specifically we compare the performance of the two models regarding the overall classification accuracy, the amount of training sequences required and how early in the progression of a sequence they are able to correctly classif… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(16 citation statements)
references
References 13 publications
0
16
0
Order By: Relevance
“…We use vanilla RNN units, as we experimentally observe that gated recurrent layers quickly lead to overfitting problems. We speculate that this is due to the small size of the dataset used here compared to datasets usually employed to train deep LSTM and GRU recurrent networks [20,13]. We use a one-dimensional global average pooling layer to summarise temporal patterns extracted by the recurrent layers.…”
Section: Convolutional Recurrent Neural Networkmentioning
confidence: 99%
“…We use vanilla RNN units, as we experimentally observe that gated recurrent layers quickly lead to overfitting problems. We speculate that this is due to the small size of the dataset used here compared to datasets usually employed to train deep LSTM and GRU recurrent networks [20,13]. We use a one-dimensional global average pooling layer to summarise temporal patterns extracted by the recurrent layers.…”
Section: Convolutional Recurrent Neural Networkmentioning
confidence: 99%
“…This benchmark can be described as either a fast Markov model or a history search, and in essense is a conditional probability maximization method. When there are constraints on data or computational power, Hidden Markov Models can match the performance of LSTMs (Panzner & Cimiano, 2016) so we believe (Vaswani et al, 2017): The Transformer consists of an encoder and decoder each made up of N blocks. Input is a sequence of events, output is a sequence of predicted events.…”
Section: Benchmarkmentioning
confidence: 99%
“…This benchmark can be described as either a fast Markov model or a history search, and in essense is a conditional probability maximization method. When there are constraints on data or computational power, Hidden Markov Models can match the performance of LSTMs [Panzner and Cimiano, 2016] so we believe this is a good benchmark that both the LSTM from Sucholutsky et al…”
Section: Benchmarkmentioning
confidence: 99%