2013 IEEE Conference on Computer Vision and Pattern Recognition 2013
DOI: 10.1109/cvpr.2013.80
|View full text |Cite
|
Sign up to set email alerts
|

Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots

Abstract: We

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
21
0

Year Published

2014
2014
2019
2019

Publication Types

Select...
6
4

Relationship

1
9

Authors

Journals

citations
Cited by 33 publications
(22 citation statements)
references
References 33 publications
0
21
0
Order By: Relevance
“…Our work also targets action recognition in static images, but, unlike any of the above, we equip static images with dynamics learned from videos. To our knowledge, the only prior static-image approach to explicitly leverage video dynamics is [4]. However, whereas [4] leverages video to augment training images for the low-shot learning scenario, our method leverages video as a motion prior that enhances test observations.…”
Section: Related Workmentioning
confidence: 99%
“…Our work also targets action recognition in static images, but, unlike any of the above, we equip static images with dynamics learned from videos. To our knowledge, the only prior static-image approach to explicitly leverage video dynamics is [4]. However, whereas [4] leverages video to augment training images for the low-shot learning scenario, our method leverages video as a motion prior that enhances test observations.…”
Section: Related Workmentioning
confidence: 99%
“…Our research is closely related to the recent work on visual data collection from web images [42,3,8,14] or weakly annotated videos [2]. Their goal is to collect training images from the Internet with minimum human supervision, but for predefined concepts.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, it is gaining popularity to address challenging computer vision problems by leveraging both images and videos. New powerful algorithms have been developed by pursuing synergic interplay between the two complementary domains of information, especially in the areas of adapting object detectors between images and videos [16,20], human activity recognition [3], and event detection [4]. However, the storyline reconstruction extracted from both images and videos still remains as a novel and largely under-addressed problem.…”
Section: Previous Workmentioning
confidence: 99%