2019
DOI: 10.1007/978-3-030-29891-3_41
|View full text |Cite
|
Sign up to set email alerts
|

Master and Rookie Networks for Person Re-identification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2
1

Relationship

4
4

Authors

Journals

citations
Cited by 14 publications
(11 citation statements)
references
References 33 publications
0
11
0
Order By: Relevance
“…In relation to approaches considering model architectures [16,45,46,47,26,48,25,49,50], several proposals, that also inspired the proposed method, try to improve the optimization task by analyzing and modifying gradient-derived information. More accurately, the authors of [16] implement intermediate losses to reduce the vanishing gradient problem.…”
Section: Related Workmentioning
confidence: 99%
“…In relation to approaches considering model architectures [16,45,46,47,26,48,25,49,50], several proposals, that also inspired the proposed method, try to improve the optimization task by analyzing and modifying gradient-derived information. More accurately, the authors of [16] implement intermediate losses to reduce the vanishing gradient problem.…”
Section: Related Workmentioning
confidence: 99%
“…As can be seen, up to 25 joints are described for a skeleton, where a joint is defined by its (x, y) coordinates inside the RGB frame. The identified joints correspond to-nose (0), neck (1), right/left shoulder, elbow, wrist (2-7), hips middle point ( 8), right/left hip, knee, ankle, eye, ear, big toe, small toes, and heel (9)(10)(11)(12)(13)(14)(15)(16)(17)(18)(19)(20)(21)(22)(23)(24). While joint positions alone do not provide useful information, due to their strict correlation to the video they are extracted from, they can still be used to generate a detailed description of body movements, via the feature extraction module.…”
Section: Skeleton Joint Generationmentioning
confidence: 99%
“…where d(•, •) is the Euclidean distance; ∼ identifies adjacent joints of a given bone; while BE f l , BE f r , BE ll , BE lr , BE ct , BE al , BE ar , and BE h denote the left foot, right foot, left leg, right leg, chest, left arm, right arm, and head, via OpenPose joint sets (19,21), (22,24), (12,13,14), (9,10,11), (1,8), (5,6,7), (2,3,4), (0, 1), respectively. Summarizing, BE is global meta-feature (i.e., BE ∈ MF g ) defining bone length over feet, legs, chest, arms, and head, for a total of 8 distinct values over the entire recording S.…”
Section: Bones Extensionmentioning
confidence: 99%
See 1 more Smart Citation
“…Actually, in video surveillance systems, or more in general, in video monitoring systems [8,9], the whole input captured through cameras is provided by specific visual features [10][11][12] of moving subjects (e.g., skeleton joints or centre of gravity in case of people or vehicles, respectively) whose changes over time (e.g., position, speed, acceleration) determine both subjects themselves [13][14][15] and their interactive behaviour [16,17]. Other scenarios of complex interaction between humans and computers in which systems are driven by users' input are, for example, the re-identification systems [18,19], in which the matching between a probe person and a set of gallery people is computed on the basis of similarity measures coming from the subjects, including shapes, colours, textures, gaits, and others; the deception detection systems [20,21], in which, among other things, facial expressions, vocal tones, and chosen words are used to determine if a subject is lying or telling the truth; the sketch-based systems [22][23][24], in which freehand gestures forming graphical languages are used to express concepts and commands to drive general-purpose interfaces; and many others.…”
Section: Introductionmentioning
confidence: 99%