Abstract:We consider the cross-modal task of producing color representations for text phrases. Motivated by the fact that a significant fraction of user queries on an image search engine follow an (attribute, object) structure, we propose a generative adversarial network that generates color profiles for such bigrams. We design our pipeline to learn composition -the ability to combine seen attributes and objects to unseen pairs. We propose a novel dataset curation pipeline from existing public sources. We describe how … Show more
“…Kikuchi et al [10] proposed maximum likelihood estimation (MLE) and conditional variational autoencoder (CVAE) models using the Transformer-based network to recommend text and background colors for each element with text content in ecommerce mobile web pages. Moreover, Maheshwari et al [14] proposed a conditional generative adversarial networks (GAN) architecture for generating a color palette for image colorization with an attribute-object pair text input, such as 'warm sunshine' and 'cute dog'. Likewise, Bahng et al [3] proposed a conditional GAN based text-to-palette generation networks for image colorization that reflect the semantics of text input.…”
Section: Full Palette Generationmentioning
confidence: 99%
“…However, this work only examined the relationships among colors in multiple palettes. Some studies based on multi-modality learning have aimed to generate a color palette based on textual information for image colorization [3,14]. The text in these works comprises a brief sequence, such as a single word (e.g., 'sunny'), an attribute-object pair (e.g., 'cute dog'), or a phrase (e.g., 'grape to strawberry').…”
Figure 1: Concept of multimodal color recommendation in vector graphic documents. The design samples on the left have textural information in the form of text contents and image labels. We extract color palettes for each visual element, such as image, graphic, and text, and reorder the colors based on color lightness. Our method replaces specified colors in the palettes with new colors based on the textual information and surrounding colors. By recoloring the original ones with the new colors, we obtain variants of the designs. The elements in the input samples are from the Crello-v2 dataset.
“…Kikuchi et al [10] proposed maximum likelihood estimation (MLE) and conditional variational autoencoder (CVAE) models using the Transformer-based network to recommend text and background colors for each element with text content in ecommerce mobile web pages. Moreover, Maheshwari et al [14] proposed a conditional generative adversarial networks (GAN) architecture for generating a color palette for image colorization with an attribute-object pair text input, such as 'warm sunshine' and 'cute dog'. Likewise, Bahng et al [3] proposed a conditional GAN based text-to-palette generation networks for image colorization that reflect the semantics of text input.…”
Section: Full Palette Generationmentioning
confidence: 99%
“…However, this work only examined the relationships among colors in multiple palettes. Some studies based on multi-modality learning have aimed to generate a color palette based on textual information for image colorization [3,14]. The text in these works comprises a brief sequence, such as a single word (e.g., 'sunny'), an attribute-object pair (e.g., 'cute dog'), or a phrase (e.g., 'grape to strawberry').…”
Figure 1: Concept of multimodal color recommendation in vector graphic documents. The design samples on the left have textural information in the form of text contents and image labels. We extract color palettes for each visual element, such as image, graphic, and text, and reorder the colors based on color lightness. Our method replaces specified colors in the palettes with new colors based on the textual information and surrounding colors. By recoloring the original ones with the new colors, we obtain variants of the designs. The elements in the input samples are from the Crello-v2 dataset.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.