2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01347
|View full text |Cite
|
Sign up to set email alerts
|

CAG-QIL: Context-Aware Actionness Grouping via Q Imitation Learning for Online Temporal Action Localization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(21 citation statements)
references
References 30 publications
0
21
0
Order By: Relevance
“…With the proliferation of video platforms, video understanding tasks are drawing substantial attention in computer vision community. The prevailing convention for video processing [3,15,16,21,28] is still dividing the whole video into short non-overlapping snippets with a fixed duration, which neglects the semantic continuity of the video. On the other hand, cognitive scientists have observed that human senses the visual stream as a set of events [39], which alludes that there is room for research to find out a video parsing method * equal contribution, ordered by surname that preserves semantic validity and interpretability of video snippets.…”
Section: Introductionmentioning
confidence: 99%
“…With the proliferation of video platforms, video understanding tasks are drawing substantial attention in computer vision community. The prevailing convention for video processing [3,15,16,21,28] is still dividing the whole video into short non-overlapping snippets with a fixed duration, which neglects the semantic continuity of the video. On the other hand, cognitive scientists have observed that human senses the visual stream as a set of events [39], which alludes that there is room for research to find out a video parsing method * equal contribution, ordered by surname that preserves semantic validity and interpretability of video snippets.…”
Section: Introductionmentioning
confidence: 99%
“…The task has recently been received tremendous attention for from researchers [4,20,26,40,22,5,37,41,23,44,9,6,27,29,35,11,39,2]. Detailed approaches related to TAL can be found in the survey article [17]. One problem of TAL which prevents the model from the real-time application is that the model is allowed to exploit the future frames which is not suitable for real application.…”
Section: Temporal Action Localizationmentioning
confidence: 99%
“…In other words, the models are allowed to access the whole frames in a video so that they can take the relationships between all frames into account and apply the post-processing techniques such as Non-maximum suppression (NMS). However, they are inherently impractical for real-world applications such as live sports broadcasting where the frames are sequentially provided and the future Recently, online temporal action localization (On-TAL) has been introduced, incorporating TAL into streaming videos [17]. In the online setting, the model is not allowed to access future frames and can take the past and current frames only.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations