2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.473
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Autoencoder for Zero-Shot Learning

Abstract: Existing zero-shot learning (ZSL) models typically learn a projection function from a feature space to a semantic embedding space (e.g. attribute space). However, such a projection function is only concerned with predicting the training seen class semantic representation (e.g. attribute prediction) or classification. When applied to test data, which in the context of ZSL contains different (unseen) classes without training data, a ZSL model typically suffers from the project domain shift problem. In this work,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

5
653
2
3

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
2
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 756 publications
(663 citation statements)
references
References 47 publications
5
653
2
3
Order By: Relevance
“…Conventional studies on transfer learning have used a method to streamline learning that involves transferring the weight of a DNN pre-trained on a large data set, such as Ima-geNet, to another task [17]. Zero-shot learning [9][10][11] [18] is an approach that can be used to classify unknown classes. Because these methods can only be applied to the same task as the pre-trained task, they cannot be used for two different tasks, as in the case of this study.…”
Section: B Transfer Learningmentioning
confidence: 99%
“…Conventional studies on transfer learning have used a method to streamline learning that involves transferring the weight of a DNN pre-trained on a large data set, such as Ima-geNet, to another task [17]. Zero-shot learning [9][10][11] [18] is an approach that can be used to classify unknown classes. Because these methods can only be applied to the same task as the pre-trained task, they cannot be used for two different tasks, as in the case of this study.…”
Section: B Transfer Learningmentioning
confidence: 99%
“…This idea is illustrated in Figure 2. Following the work in structured output prediction [45] and prior work in zero-shot 145 learning [38,39,[46][47][48][49][50], we use a compatibility function F : X × Y → R to model the mapping between the input and output embeddings. In this model, F takes a phosphosite -kinase pair (x i , y j ) as input and returns a scalar value which is proportional to the confidence of associating the site, x i , with kinase y i .…”
Section: Problem Formulationmentioning
confidence: 99%
“…In [24], Kodirov et al compared their proposed method, SAE, with more than 10 highly qualified methods using small datasets, which include AwA [16], CUB [27], aP &Y [38] and SUN [39]. The authors improved the accuracy of recognising unseen images at least %6 in comparison with SS−voc and at most %20 in comparison with basic attibutebased learning method, DAP.…”
Section: Pre-defined Parametersmentioning
confidence: 99%
“…Algorithm 1 Implementation of optimized zero-shot learning (SAE) [24] Require: A batch set of training input (X , Y ) and training size Ensure: The best mapping matrix (W ) for zero-shot learning 1: Tuning embedded parameters…”
Section: Introductionmentioning
confidence: 99%