2020
DOI: 10.1016/j.patrec.2020.01.023
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical Attention Network for Action Segmentation

Abstract: Temporal segmentation of events is an essential task and a precursor for the automatic recognition of human actions in the video. Several attempts have been made to capture frame-level salient aspects through attention but they lack the capacity to effectively map the temporal relationships in between the frames as they only capture a limited span of temporal dependencies. To this end we propose a complete end-to-end supervised learning approach that can better learn relationships between actions over time, th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(1 citation statement)
references
References 39 publications
0
1
0
Order By: Relevance
“…These ResNet models are less complex (in terms of parameters) compared to previous pre-trained networks such as VGG networks, where even the deeper ResNet network (i.e. ResNet152) is less complex (11.3 Due to the aforementioned advantages, pre-trained ResNet networks have been widely used as a feature extraction backbone within both the action recognition domain [144], [145], [146], [46], [147] and related problem domains [148], [149], [150].…”
Section: Appendix a Feature Extractionmentioning
confidence: 99%
“…These ResNet models are less complex (in terms of parameters) compared to previous pre-trained networks such as VGG networks, where even the deeper ResNet network (i.e. ResNet152) is less complex (11.3 Due to the aforementioned advantages, pre-trained ResNet networks have been widely used as a feature extraction backbone within both the action recognition domain [144], [145], [146], [46], [147] and related problem domains [148], [149], [150].…”
Section: Appendix a Feature Extractionmentioning
confidence: 99%