2017
DOI: 10.1016/j.cviu.2017.06.007
|View full text |Cite
|
Sign up to set email alerts
|

Compact descriptors for sketch-based image retrieval using a triplet loss convolutional neural network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
44
0
2

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 70 publications
(55 citation statements)
references
References 43 publications
(67 reference statements)
0
44
0
2
Order By: Relevance
“…The addition of fc layers to create cross-modal embedding (-R) slightly improves (importantly, does not degrade) the intermediate raster embedding R available via [6]. The method significantly outperforms recent triplet SBIR approaches [28,5]. Note that the S-I and S-S figures are non-comparable; they search different datasets.…”
Section: Evaluating Cross-modal Searchmentioning
confidence: 98%
See 2 more Smart Citations
“…The addition of fc layers to create cross-modal embedding (-R) slightly improves (importantly, does not degrade) the intermediate raster embedding R available via [6]. The method significantly outperforms recent triplet SBIR approaches [28,5]. Note that the S-I and S-S figures are non-comparable; they search different datasets.…”
Section: Evaluating Cross-modal Searchmentioning
confidence: 98%
“…7 (top). We perform two ablations to our proposed LiveSketch (LS) system: 1) querying with rasterized versions of the QD-345 queries (-R) using the proposed embedding S; 2) querying with rasterized queries in the intermediate embedding R (-R-I) which degenerates to [6]; we also baseline against two further recent SBIR techniques: the unshared triplet GoogleNet-V1 architecture proposed by Sangkloy et al [28], and the triplet edgemap approach of Bui et al [5]. We compute class-and instance-level precision for all queries resulting in 345 × 15 × 5 =∼ 26K MTurk annotations.…”
Section: Evaluating Cross-modal Searchmentioning
confidence: 99%
See 1 more Smart Citation
“…An early example was SketchANet [45], which performed sketch recognition using Alexnet [24]. More recently, triplet convolutional neural networks have gained interest as they have the capacity to deal with deep embedding spaces [9]. Improving the image similarity metric is a main challenge as these triplet architectures are used to measure similarities between images and sketches [10].…”
Section: D Sketching For Retrievalmentioning
confidence: 99%
“…This set of values can be seen as a feature vector with 2048 dimensions to be an input for another classifier, such as the SVM for example. If a more compact representation is needed, one can use dimensionality reduction methods or quantization based on PCA [51] or Product Quantization [52], [53].…”
Section: Beyond Classification: Fine-tuning Feature Extraction Anmentioning
confidence: 99%