2018
DOI: 10.3837/tiis.2018.01.019
|View full text |Cite
|
Sign up to set email alerts
|

Deep Image Annotation and Classification by Fusing Multi-Modal Semantic Topics

Abstract: Due to the semantic gap problem across different modalities, automatically retrieval from multimedia information still faces a main challenge. It is desirable to provide an effective joint model to bridge the gap and organize the relationships between them. In this work, we develop a deep image annotation and classification by fusing multi-modal semantic topics (DAC_mmst) model, which has the capacity for finding visual and non-visual topics by jointly modeling the image and loosely related text for deep image… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2019
2019
2019
2019

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 37 publications
0
2
0
Order By: Relevance
“…Based on the that maximizes this marginal likelihood, the distribution of the function value for the unobserved data is . That is, the posterior distribution has mean function and variance function , where and are calculated through the following equations: (9) where, denotes the covariance values between all observed data and the new data based on the hyperparameter value , and is the variance value at . Fig.…”
Section: Gaussian Process Regressionmentioning
confidence: 99%
“…Based on the that maximizes this marginal likelihood, the distribution of the function value for the unobserved data is . That is, the posterior distribution has mean function and variance function , where and are calculated through the following equations: (9) where, denotes the covariance values between all observed data and the new data based on the hyperparameter value , and is the variance value at . Fig.…”
Section: Gaussian Process Regressionmentioning
confidence: 99%
“…Previously methods build image annotation model based on three basic dependencies: relations between image and label (image/label), between images (image/image) and between labels (label/label). Assuming targets are independent to each other, they managed to establish the relations between images and labels [1][2][3][4][5][6][7][8]. Despite its powerful discriminative ability, it cannot detect those visually hard-to-detect targets, such as small and blured objects.…”
Section: Introductionmentioning
confidence: 99%