2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00109
|View full text |Cite
|
Sign up to set email alerts
|

Weakly-Supervised Action Localization by Generative Attention Modeling

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
79
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 148 publications
(80 citation statements)
references
References 42 publications
0
79
0
Order By: Relevance
“…Localization After obtaining the TCAM a for a query video and class , we threshold the a and group together consecutive snippets that are above a given threshold . Then, following the standard practice [30,34,38], we arrive at a set of action predictions ( , , ), where , and are the start, end, and prediction score of a certain prediction. We set the prediction score as the average of a of the individual snippets, that is, = 1 − +1 +1 = a ( ).…”
Section: Localization and Classificationmentioning
confidence: 99%
See 4 more Smart Citations
“…Localization After obtaining the TCAM a for a query video and class , we threshold the a and group together consecutive snippets that are above a given threshold . Then, following the standard practice [30,34,38], we arrive at a set of action predictions ( , , ), where , and are the start, end, and prediction score of a certain prediction. We set the prediction score as the average of a of the individual snippets, that is, = 1 − +1 +1 = a ( ).…”
Section: Localization and Classificationmentioning
confidence: 99%
“…Our TCAMs are similar in functionality with those in other weakly supervised works [31,32,38], however, a crucial difference is that, in our case, TCAMs are calculated based on similarities with reference videos as in [21], and not from class-based classifiers that are hardly trained from one/few examples. During training, we optimize a classification loss at video level, in order to ensure the interclass separability of learned features.…”
Section: Introductionmentioning
confidence: 97%
See 3 more Smart Citations