Proceedings of the 27th ACM International Conference on Multimedia 2019
DOI: 10.1145/3343031.3350889
|View full text |Cite
|
Sign up to set email alerts
|

Who, Where, and What to Wear?

Abstract: Fashion knowledge helps people to dress properly and addresses not only physiological needs of users, but also the demands of social activities and conventions. It usually involves three mutually related aspects of: occasion, person and clothing. However, there are few works focusing on extracting such knowledge, which will greatly benefit many downstream applications, such as fashion recommendation. In this paper, we propose a novel method to automatically harvest fashion knowledge from social media. We unify… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
1
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 25 publications
(3 citation statements)
references
References 34 publications
(53 reference statements)
0
1
0
Order By: Relevance
“…Then, we explore how to leverage the retrieved comments in multimodal classification and exploit a self-training framework to identify comments' hints which shape the cross-modal understanding (henceforth comment-aware self-training). This considers method feasibility in scenarios where large-scale labeled data is unavailable, which commonly appears in the realistic practice, because the annotation for multimodal data from social media is extremely expensive (Ma et al, 2019). Concretely, we adopt a teacher-student prototype (Meng et al, 2020;Shen et al, 2021) and tailor-make it to learn multimodal understanding with the help of user comments.…”
Section: Textmentioning
confidence: 99%
See 1 more Smart Citation
“…Then, we explore how to leverage the retrieved comments in multimodal classification and exploit a self-training framework to identify comments' hints which shape the cross-modal understanding (henceforth comment-aware self-training). This considers method feasibility in scenarios where large-scale labeled data is unavailable, which commonly appears in the realistic practice, because the annotation for multimodal data from social media is extremely expensive (Ma et al, 2019). Concretely, we adopt a teacher-student prototype (Meng et al, 2020;Shen et al, 2021) and tailor-make it to learn multimodal understanding with the help of user comments.…”
Section: Textmentioning
confidence: 99%
“…The labeled dataset L is usually limited in scales (Ma et al, 2019), posing the over-fitting concern. Meanwhile the retrieved posts, similar to the data in L, could form an unlabeled set (U = {x ′ i , c i } Kl i=1 ) to enrich training data.…”
Section: Self-training With Retrieved Postsmentioning
confidence: 99%
“…According to different studies, e-commerce retailers, such as Amazon, eBay, and Shopstyle, and social networking sites, such as Pinterest, Snapchat, Instagram, Facebook, Chictopia, and Lookbook, are now regarded as the most popular media for fashion advice and recommendations [15][16][17][18][19][20][21][22]. Research on textual content, such as posts and comments [23], emotion and information diffusion [24], and images has attracted the attention of modernday researchers, as it can help to predict fashion trends and facilitate the development of effective recommendation systems [5,[25][26][27]]. An effective recommendation system is a crucial tool for successfully conducting an e-commerce business.…”
Section: Introductionmentioning
confidence: 99%