2016
DOI: 10.1007/978-3-319-45886-1_20
|View full text |Cite
|
Sign up to set email alerts
|

From Traditional to Modern: Domain Adaptation for Action Classification in Short Social Video Clips

Abstract: Abstract. Short internet video clips like vines present a significantly wild distribution compared to traditional video datasets. In this paper, we focus on the problem of unsupervised action classification in wild vines using traditional labeled datasets. To this end, we use a data augmentation based simple domain adaptation strategy. We utilize semantic word2vec space as a common subspace to embed video features from both, labeled source domain and unlabled target domain. Our method incrementally augments th… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2016
2016
2016
2016

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 28 publications
0
2
0
Order By: Relevance
“…Similar to all these approaches we also used a neural network for learning embedding function but we focus only on hash tags. [16] maps the video representations to semantic space for improving action classification. Recent image captioning methods [23,6] have shown remarkable progress in generating rich descriptive sentences for natural images.…”
Section: Visual-semantic Joint Embeddingmentioning
confidence: 99%
See 1 more Smart Citation
“…Similar to all these approaches we also used a neural network for learning embedding function but we focus only on hash tags. [16] maps the video representations to semantic space for improving action classification. Recent image captioning methods [23,6] have shown remarkable progress in generating rich descriptive sentences for natural images.…”
Section: Visual-semantic Joint Embeddingmentioning
confidence: 99%
“…Vector representation of words is used by many recent approaches for joint visual semantic learning [17,16,23,27,6]. Socher et al [17] learns an embedding function which performs a mapping of image features to semantic word space.…”
Section: Visual-semantic Joint Embeddingmentioning
confidence: 99%