2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00543
|View full text |Cite
|
Sign up to set email alerts
|

Exploit the Unknown Gradually: One-Shot Video-Based Person Re-identification by Stepwise Learning

Abstract: We focus on the one-shot learning for video-based person re-Identification (re-ID). Unlabeled tracklets for the person re-ID tasks can be easily obtained by preprocessing, such as pedestrian detection and tracking. In this paper, we propose an approach to exploiting unlabeled tracklets by gradually but steadily improving the discriminative capability of the Convolutional Neural Network (CNN) feature representation via stepwise learning. We first initialize a CNN model using one labeled tracklet for each identi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
285
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 360 publications
(294 citation statements)
references
References 33 publications
1
285
0
Order By: Relevance
“…Suh et al [39] propose a two-stream architecture to jointly learn the appearance feature and part feature, and fuse the image level features through a pooling strategy. Average pooling is also used in recent works [21,47], which apply unsupervised learning for video person ReID. Temporal pooling exhibits promising efficiency, but extracts frame features independently and ignores the temporal orders among adjacent frames.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Suh et al [39] propose a two-stream architecture to jointly learn the appearance feature and part feature, and fuse the image level features through a pooling strategy. Average pooling is also used in recent works [21,47], which apply unsupervised learning for video person ReID. Temporal pooling exhibits promising efficiency, but extracts frame features independently and ignores the temporal orders among adjacent frames.…”
Section: Related Workmentioning
confidence: 99%
“…As shown in our experiments and visualizations, GLTR presents strong discriminative power and robustness. We test our approach on a newly proposed Large-Scale Video dataset for person ReID (LS-VID) and four widely used video ReID datasets, including PRID [14], iLIDS-VID [43], MARS [56], and DukeMTMC-VideoReID [47,34], respectively. Experimental results show that GLTR achieves consistent performance superiority on those datasets.…”
Section: Introductionmentioning
confidence: 99%
“…Similar to our work, cross-camera tracklet association (labeling) [21], [28], [39] is more scalable for unsupervised Re-ID with no extra data or assumption on the similarity between source and target domains. Due to the limitation of existing datasets, most of them do not perform Re-ID learning in a pure unsupervised way.…”
Section: Related Workmentioning
confidence: 63%
“…DukeMTMC-VideoReID dataset is another large scale benchmark dataset for video-based person Re-ID, which is derived from the DukeMTMC dataset [56] and re-organized by Wu et al [57]. The DukeMTMC-VideoReID dataset contains totally 4,832 tracklets and 1,812 identities, it is separated into 702, 702 and 408 identities for training, testing and distraction.…”
Section: A Datasetsmentioning
confidence: 99%