2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00148
|View full text |Cite
|
Sign up to set email alerts
|

Towards Rich Feature Discovery With Class Activation Maps Augmentation for Person Re-Identification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
93
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 232 publications
(95 citation statements)
references
References 32 publications
1
93
0
Order By: Relevance
“…Nevertheless, none of the existing ReID models addresses omni-scale feature learning. In recent years, deep convolutional neural networks (CNNs) have been widely used in person ReID to learn discriminative features [2,27,31,42,47,63,65,77]. However, most of the CNNs adopted, such as ResNet [13], were originally designed for object category-level recognition tasks that are fundamentally different from the instance-level recognition task in ReID.…”
Section: Introductionmentioning
confidence: 99%
“…Nevertheless, none of the existing ReID models addresses omni-scale feature learning. In recent years, deep convolutional neural networks (CNNs) have been widely used in person ReID to learn discriminative features [2,27,31,42,47,63,65,77]. However, most of the CNNs adopted, such as ResNet [13], were originally designed for object category-level recognition tasks that are fundamentally different from the instance-level recognition task in ReID.…”
Section: Introductionmentioning
confidence: 99%
“…We compare the proposed method with 33 recent published works including (1) global feature based methods which aims to learn the global feature from the feature map directly, including PAN [74], DMML [7], DCDS [1], VCFL [30], MVPM [41], LRDNN [79], RB [35], LITM [63], IANet [23], Sphere [14], BNNeck [32], OSNet [78], AANet [46], DG-Net [72], BDB [12], Circle [42], SFT [31], (2) part based methods including PCB+RPP [43], Local [57], HPM [16], CASN [71], AutoReID [34], MGN [49], BHP [20] and Pyramidal [68] which utilize the semantic parts or horizontal stripes to extract part-level feature, and (3) attention based methods including MHAN [3], CAMA [58], SONA [53], CAR [80], SCAL [6], ABD-Net [8], DAAF [10] and RGA [65]. These methods are categorized into 3 types based on different backbones: the ones which employ ResNet-50 directly, the ones which modify ResNet-50 by introducing additional branches, attention subnets or dilated convolution, and the others which don't use ResNet-50.…”
Section: Comparison Resultsmentioning
confidence: 99%
“…Specifically, they aim to learn the features by improving loss functions [9,14,22,31,41,42,50,55,63], improving the training techniques [1,4,12,24,32,35,37,54], adding additional network modules [23,23,51,62], using extra semantic annotations [30,46,47,79] or generating more training samples [17,33,72,76,77]. Besides, more recent studies [3,6,8,10,27,28,38,46,48,53,58,61,64,65,67,80] integrate attention mechanisms into deep models to enhance the feature representation. To obtain the holistic features, most of these methods utilize global average pooling (GAP), global max pooling (GMP) or both of them on each channel of the feature map.…”
Section: Related Workmentioning
confidence: 99%
“…Because of the modularity of our work, this model could replace the extraction and classification steps without changing the acquisition and presentation results. Another direction for expansion is the usage of architectural changes, such as attention operations (Prabhakararao and Dandapat 2020), or post hoc interpretation tools, such as class activation maps (Yang et al 2019), which allow the visualisation of the features that have led the network to produce a specific diagnostic. Although current technologies are increasingly black-box given the advent of DL, the application of these post-hoc analysis tools can have a multiplicative effect in a man-plus-machine scenario, especially through the interactivity of AR, which would allow the visualisation of the decision maps and, thus, identify whether the classifier is confidently producing a prediction (Strodthoff et al 2020).…”
Section: Discussionmentioning
confidence: 99%