2021
DOI: 10.1016/j.neucom.2021.04.121
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Two-Step Networks for Temporal Action Segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 28 publications
(8 citation statements)
references
References 14 publications
0
8
0
Order By: Relevance
“…Some TCNS works [17,18,8] mainly focus on improving receptive fields that model long-term dependencies with encoder structures, dilated convolutions, or deformable convolutions. [12,21] build architecture on the two-branch approach : One branch exploits wide long-term time receptive fields based on TCNs. The second exploits frame-boundary based on action boundary regression.…”
Section: Action Segmentationmentioning
confidence: 99%
See 1 more Smart Citation
“…Some TCNS works [17,18,8] mainly focus on improving receptive fields that model long-term dependencies with encoder structures, dilated convolutions, or deformable convolutions. [12,21] build architecture on the two-branch approach : One branch exploits wide long-term time receptive fields based on TCNs. The second exploits frame-boundary based on action boundary regression.…”
Section: Action Segmentationmentioning
confidence: 99%
“…It has been a hot topic in human action analysis, which is widely used in video surveillance [6], action teaching, and robotics [34]. Recently, some works [17,8,10,35,21] have studied the long range dependencies between correlated actions in action segmentation using temporal convolution networks (TCNs) for models. The TCNs enhance long-term receptive fields by increasing convolution layers.…”
Section: Introductionmentioning
confidence: 99%
“…If the model applies in the full video, our model could use postprocessing (Li et al 2021) to improve the performance of discrimination of action completeness. Evaluation Metrics To evaluate SVTAS task results, we adopt several metrics including frame-wise accuracy (Acc) (Farha and Gall 2019), mean average precision (mAP) metric with temporal IoU of 0.5 (denote by mAP@0.5) (Wang et al 2022a), the area under the AR (under specified temporal IoU thresholds for [0.5:0.05:1.0]) vs. AN (limiting the average number of proposals for each video and set to 100) curve (AUC) (Alwassel, Giancola, and Ghanem 2021) and the F1 score at temporal IoU threshold 0.1 (denote by F1@0.1) (Li et al 2022).…”
Section: Transegermentioning
confidence: 99%
“…Literature is rich of works on action recognition methodologies successfully applied to short videos analysis. In recent years, the focus has been on temporal segmentation of actions in long untrimmed videos 28 . In Industry 4.0 domain, where collaborative tasks are performed by humans and robots in highly varying conditions, it is imperative to recognize the exact beginning and ending of an action.…”
Section: Technical Validationmentioning
confidence: 99%