2019
DOI: 10.1007/978-3-030-20890-5_37
|View full text |Cite
|
Sign up to set email alerts
|

Coarse-to-Fine: A RNN-Based Hierarchical Attention Model for Vehicle Re-identification

Abstract: Xiu-Shen Wei 1[0000−0002−8200−1845] , Chen-Lin Zhang 2[0000−0002−3168−1852] , Lingqiao Liu 3[0000−0003−3584−795X] , Chunhua Shen 3[0000−0002−8648−8718] , and Jianxin Wu 2[0000−0002−2085−7568]Abstract Vehicle re-identification is an important problem and becomes desirable with the rapid expansion of applications in video surveillance and intelligent transportation. By recalling the identification process of human vision, we are aware that there exists a native hierarchical dependency when humans identify differ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 30 publications
(20 citation statements)
references
References 36 publications
0
19
0
Order By: Relevance
“…Recent works on vehicle Re-ID can be divided as viewpoint-independent and viewpoint-dependent. The viewpoint-independent methods such as PROVID [ 13 ], RHH-HA [ 16 ], RAM [ 3 ], Multi-Region Model (MRM) [ 31 ] mainly focus on learning of robust global and local features or learning of distance metric (Batch Sample (BS) [ 32 ] and Feature Distance Adversarial Network (FDA-Net) [ 5 ]). The viewpoint-dependent methods are dedicated to learning orientation-invariance features (Orientation Invariant Feature Embedding (OIFE) and OIFE + ST) [ 6 ] or multi-view features (VAMI and VAMI + ST) [ 8 ].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Recent works on vehicle Re-ID can be divided as viewpoint-independent and viewpoint-dependent. The viewpoint-independent methods such as PROVID [ 13 ], RHH-HA [ 16 ], RAM [ 3 ], Multi-Region Model (MRM) [ 31 ] mainly focus on learning of robust global and local features or learning of distance metric (Batch Sample (BS) [ 32 ] and Feature Distance Adversarial Network (FDA-Net) [ 5 ]). The viewpoint-dependent methods are dedicated to learning orientation-invariance features (Orientation Invariant Feature Embedding (OIFE) and OIFE + ST) [ 6 ] or multi-view features (VAMI and VAMI + ST) [ 8 ].…”
Section: Methodsmentioning
confidence: 99%
“…Cui et al [ 15 ] fused the classification features of color, vehicle model, and pasted marks on windshield as the final features to describe the vehicle. Some studies used a variety of attributes to identify vehicles from coarse to fine, such as Progress Vehicle Re-identification (PROVID) [ 13 ] and RNN-based Hierarchical Attention (RNN-HA) [ 16 ]. These coarse-to-fine approaches require multiple recognition processes and cannot be implemented end to end.…”
Section: Related Workmentioning
confidence: 99%
“…Wang et al [19] utilized 20 key-point locations of vehicles to extract orientation information and proposed an orientation invariant feature embedding module. De et al [23] proposed a two-stream Siamese classification model for vehicle re-ID, and Wei et al [24] proposed an recurrent neural network-based hierarchical attention (RNN-HA) network, which combines a large number of attributes for vehicle re-ID. Bai et al [14] proposed a group sensitive triplet embedding approach that can model the interclass differences.…”
Section: Vehicle Re-idmentioning
confidence: 99%
“…To address these issues, various approaches have been proposed including deep metric learning models [1,4,6,16,43,46,48] and informative region based approaches [24,40,52]. Deep metric learning approaches focus on learning an appropriate distance metric mainly based on global representation towards reducing the distances between identical vehicle images and enlarging those between different vehicle images.…”
Section: Introductionmentioning
confidence: 99%
“…As shown in the third and forth rows of Figure 1, the regions within blue boxes are crucial for identifying the two similar vehicles. To utilize such local visual cues, some works [24,40,52] exploited attention models to localize discriminative regions and learn detailed local representations. However, not all regions present discriminative content and are spatially aligned within the parts.…”
Section: Introductionmentioning
confidence: 99%