The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2017
DOI: 10.1109/tip.2017.2666038
|View full text |Cite
|
Sign up to set email alerts
|

Nonlinear Deep Kernel Learning for Image Annotation

Abstract: Multiple kernel learning (MKL) is a widely used technique for kernel design. Its principle consists in learning, for a given support vector classifier, the most suitable convex (or sparse) linear combination of standard elementary kernels. However, these combinations are shallow and often powerless to capture the actual similarity between highly semantic data, especially for challenging classification tasks such as image annotation. In this paper, we redefine multiple kernels using deep multi-layer networks. I… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
62
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 70 publications
(62 citation statements)
references
References 39 publications
0
62
0
Order By: Relevance
“…Since the classification ability is quantified by the generalization error, we will attempt to develop a convergence bound of the generalization error of HSIC‐MKL based on the established theory of Rademacher complexities. Besides, the further validation of the use of the proposed algorithm on more real‐world applications, such as computer vision, speech and signal processing, and natural language processing, and expanding the proposed model to extreme learning machine and deep kernel learning are also important issues to be investigated.…”
Section: Resultsmentioning
confidence: 99%
“…Since the classification ability is quantified by the generalization error, we will attempt to develop a convergence bound of the generalization error of HSIC‐MKL based on the established theory of Rademacher complexities. Besides, the further validation of the use of the proposed algorithm on more real‐world applications, such as computer vision, speech and signal processing, and natural language processing, and expanding the proposed model to extreme learning machine and deep kernel learning are also important issues to be investigated.…”
Section: Resultsmentioning
confidence: 99%
“…Semantic-based image retrieval has become an active research topic in recent times. There were many techniques of image retrieval, which have been implemented aiming to reduce the "semantic gap" by modeling high-level semantics, such as techniques to build a model for mapping between low-level features and high-level semantics [2,21], query techniques based on ontology to accurately describe semantics for images [18,25], techniques for classification data [12,13,17], etc.…”
Section: Related Workmentioning
confidence: 99%
“…In our experiments (see Table 6), we use four elementary kernels (linear, polynomial, RBF and HI) combined with different features as inputs to the designed DKN and DMN networks: "handcrafted features" including GIST and SIFT and "learned features" taken from ResNet [50] (pretrained on the ImageNet) which is a very deep architecture consisting of 152 layers; the 2048 dimensional features of the last pooling layer are used in our annotation task. Using all these elementary kernels and features, we first train a DKN in a supervised way according to [49], then we design and finetune its associated DMNs with |S| = 700 and |S | = 3000 (as done in Table. 4).…”
Section: Initial and Fine-tuned Dmns Assuming The Weights {Wmentioning
confidence: 99%