2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01040
|View full text |Cite
|
Sign up to set email alerts
|

Counting Out Time: Class Agnostic Video Repetition Counting in the Wild

Abstract: We introduce a dataset of annotations of temporal repetitions in videos. The dataset, OVR (pronounced as over), contains annotations for over 72K videos, with each annotation specifying the number of repetitions, the start and end time of the repetitions, and also a free-form description of what is repeating. The annotations are provided for videos sourced from Kinetics and Ego4D, and consequently cover both Exo and Ego viewing conditions, with a huge variety of actions and activities. Moreover, OVR is almost … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
125
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 84 publications
(125 citation statements)
references
References 50 publications
(46 reference statements)
0
125
0
Order By: Relevance
“…For the n-tuplet and synchronization loss to function correctly, it is required that each cine in our training set contains approximately one cardiac cycle. To enforce the one cycle constraint, we either used the provided average heart rate and frame rate data to approximate a single period, or in cases where this information was not available, we applied the class agnostic video repetition counting method (RepNet) proposed in [15]. In training studies with more than one cycle, all available cycles are considered as individual training samples.…”
Section: A Datasetmentioning
confidence: 99%
“…For the n-tuplet and synchronization loss to function correctly, it is required that each cine in our training set contains approximately one cardiac cycle. To enforce the one cycle constraint, we either used the provided average heart rate and frame rate data to approximate a single period, or in cases where this information was not available, we applied the class agnostic video repetition counting method (RepNet) proposed in [15]. In training studies with more than one cycle, all available cycles are considered as individual training samples.…”
Section: A Datasetmentioning
confidence: 99%
“…It is reliable in the unsupervised situation when the semantic content of the videos, the number of periods and the valid duration of the video are unknown. Debidatta et al [18] introduced a new symmetric matrix that helps achieve up-to-date accuracy. It acts as an intermediate layer to predict the cycle length and valid periodic length.…”
Section: Repetitive Countingmentioning
confidence: 99%
“…Therefore, we synthesize the Rep-Penn dataset by utilizing the single-cycle exercise video frames. Inspired by dataset synthesis methods from [18], we connect the forward video with the rewinding video instead of connecting the original video repeatedly. In this way, continuity of the body movement is guaranteed, and exercise of various periods can be generated for repetition counting.…”
Section: Rep-penn Benchmarkmentioning
confidence: 99%
See 1 more Smart Citation
“…Still, counting the repetition of exercise is a less studied field compared with the prosperous-growing field in human activity recognition. Dwibedi et al [10] pointed out that it is strenuous to search for suitable video from large-scale public datasets, as there is no specific keyword and annotations catering to this particular needs. The root cause of this phenomenon is that annotating video or signal across the temporal domain is labor-intensive and monotonous work.…”
Section: Introductionmentioning
confidence: 99%