2013
DOI: 10.1109/tmm.2013.2280895
|View full text |Cite
|
Sign up to set email alerts
|

Web Multimedia Object Classification Using Cross-Domain Correlation Knowledge

Abstract: Abstract-Given a collection of web images with the corresponding textual descriptions, in this paper, we propose a novel cross-domain learning method to classify these web multimedia objects by transferring the correlation knowledge among different information sources. Here, the knowledge is extracted from unlabeled objects through unsupervised learning and applied to perform supervised classification tasks. To mine more meaningful correlation knowledge, instead of using commonly used visual words in the tradi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 28 publications
0
6
0
Order By: Relevance
“…A knowledge adaptation method for Ad Hoc multimedia event detection is proposed in [32]. In [33], cross-domain correlation knowledge is used for web multimedia object classification. In [34], a feature transformation method is proposed to indirectly transfer semantic knowledge between text and images.…”
Section: B Multi-domain Feature Learningmentioning
confidence: 99%
“…A knowledge adaptation method for Ad Hoc multimedia event detection is proposed in [32]. In [33], cross-domain correlation knowledge is used for web multimedia object classification. In [34], a feature transformation method is proposed to indirectly transfer semantic knowledge between text and images.…”
Section: B Multi-domain Feature Learningmentioning
confidence: 99%
“…Significant efforts have been taken to integrate visual and textual analyses [2][3][4]. For example, Wang et al [5] present an algorithm to learn the relations between scenes, objects, and texts with the help of image-level labels.…”
Section: Introductionmentioning
confidence: 99%
“…(1) Visual content and text are always separately learned, making the traditional methods hard to be trained end-to-end. (2) Learning tasks converted to classification problems, empowered by large-scale annotated data with end-to-end training using neural networks, which is not capable of describing concepts unseen in the training pairs. (3) e spatial relations defined by prepositions have to be learned with the bounding boxes of objects, which are so immoderately challenging to obtain.…”
Section: Introductionmentioning
confidence: 99%
“…The rapid growth of internet images also brings up many applications that make a good use of this big data, such as image retrieval [33], stitching [29], recommendation [62], image quality assessment [17], or benchmark dataset making [58], etc. It is also obvious that within these images the number and richness of contained objects are of great usefulness, especially in some data-driven applications related to object manipulation, such as object retrieval [51,57], classification [32], enhancement [59], etc. Traditionally, searching for images that contain desired objects is done through online search engines such as Google Images, which utilizes content-based image retrieval techniques to find images that contain similar objects, or simply acquires images by their tagged information.…”
Section: Introductionmentioning
confidence: 99%