The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
Proceedings of the 30th ACM International Conference on Information &Amp; Knowledge Management 2021
DOI: 10.1145/3459637.3482346
|View full text |Cite
|
Sign up to set email alerts
|

Generating Compositional Color Representations from Text

Abstract: We consider the cross-modal task of producing color representations for text phrases. Motivated by the fact that a significant fraction of user queries on an image search engine follow an (attribute, object) structure, we propose a generative adversarial network that generates color profiles for such bigrams. We design our pipeline to learn composition -the ability to combine seen attributes and objects to unseen pairs. We propose a novel dataset curation pipeline from existing public sources. We describe how … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 41 publications
0
2
0
Order By: Relevance
“…Kikuchi et al [10] proposed maximum likelihood estimation (MLE) and conditional variational autoencoder (CVAE) models using the Transformer-based network to recommend text and background colors for each element with text content in ecommerce mobile web pages. Moreover, Maheshwari et al [14] proposed a conditional generative adversarial networks (GAN) architecture for generating a color palette for image colorization with an attribute-object pair text input, such as 'warm sunshine' and 'cute dog'. Likewise, Bahng et al [3] proposed a conditional GAN based text-to-palette generation networks for image colorization that reflect the semantics of text input.…”
Section: Full Palette Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…Kikuchi et al [10] proposed maximum likelihood estimation (MLE) and conditional variational autoencoder (CVAE) models using the Transformer-based network to recommend text and background colors for each element with text content in ecommerce mobile web pages. Moreover, Maheshwari et al [14] proposed a conditional generative adversarial networks (GAN) architecture for generating a color palette for image colorization with an attribute-object pair text input, such as 'warm sunshine' and 'cute dog'. Likewise, Bahng et al [3] proposed a conditional GAN based text-to-palette generation networks for image colorization that reflect the semantics of text input.…”
Section: Full Palette Generationmentioning
confidence: 99%
“…However, this work only examined the relationships among colors in multiple palettes. Some studies based on multi-modality learning have aimed to generate a color palette based on textual information for image colorization [3,14]. The text in these works comprises a brief sequence, such as a single word (e.g., 'sunny'), an attribute-object pair (e.g., 'cute dog'), or a phrase (e.g., 'grape to strawberry').…”
Section: Introductionmentioning
confidence: 99%