Proceedings of the 29th ACM International Conference on Multimedia 2021
DOI: 10.1145/3474085.3475691
|View full text |Cite
|
Sign up to set email alerts
|

Collocation and Try-on Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 16 publications
(3 citation statements)
references
References 38 publications
0
3
0
Order By: Relevance
“…cross domain retrieval) (Tangseng, Yamaguchi, and Okatani 2017;Li et al 2017;Han et al 2017;Hsiao and Grauman 2018;Tangseng, Yamaguchi, and Okatani 2017;Shih et al 2018;Li et al 2020), set complementary item retrieval (Hu, Yi, and Davis 2015;Huang et al 2015;Liu et al 2012), personalized set complementary item prediction (requires user input) (Taraviya et al 2021;Chen et al 2019;Li et al 2020;Su et al 2021;Zheng et al 2021;Guan et al 2022b,a) and multi-modal complementary item prediction (Guan et al 2021). All these prior work focus on feature representation learning.…”
Section: Related Workmentioning
confidence: 99%
“…cross domain retrieval) (Tangseng, Yamaguchi, and Okatani 2017;Li et al 2017;Han et al 2017;Hsiao and Grauman 2018;Tangseng, Yamaguchi, and Okatani 2017;Shih et al 2018;Li et al 2020), set complementary item retrieval (Hu, Yi, and Davis 2015;Huang et al 2015;Liu et al 2012), personalized set complementary item prediction (requires user input) (Taraviya et al 2021;Chen et al 2019;Li et al 2020;Su et al 2021;Zheng et al 2021;Guan et al 2022b,a) and multi-modal complementary item prediction (Guan et al 2021). All these prior work focus on feature representation learning.…”
Section: Related Workmentioning
confidence: 99%
“…Suppose there are 𝑃 global attributes. We then deploy 𝑃 learnable condition masks [40] on f 𝑟 to derive the global attribute features of the reference image as follows,…”
Section: Attribute Feature Extractionmentioning
confidence: 99%
“…In addition, in the computer vision domain, Caramalau et al [2] presented a novel sequential GCN to learn node representations and distinguish sufficiently different unlabeled examples from labeled examples for active learning, and Zhang et al [53] devised a multimodal interaction GCN to jointly explore the complex intramodal relations and inter-modal interactions for temporal language localization in videos. Zheng et al [55] integrated disentangled item representations into a GCN to adaptively propagate the finegrained compatibility relationships among items for outfit compatibility modeling, Wang et al [46] developed a novel neural graph collaborative filtering method that integrated user-item interactions into a user embedding process for recommendation systems.…”
Section: Graph Convolutional Networkmentioning
confidence: 99%