2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2023
DOI: 10.1109/cvpr52729.2023.00978
|View full text |Cite
|
Sign up to set email alerts
|

Inversion-based Style Transfer with Diffusion Models

Yuxin Zhang,
Nisha Huang,
Fan Tang
et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 57 publications
(5 citation statements)
references
References 24 publications
0
4
0
Order By: Relevance
“…Generative models, including generative adversarial networks (GANs) [19], variational auto-encoders (VAEs) [28], are now integral to AI-supported design processes (e.g., [6,45]). Simultaneously, diffusion models, renowned for their ability to produce rich and diverse samples, have found applications in areas such as artistic style transfer [73]. With its fast-evolving capabilities, AI has been playing an increasing role in human-AI collaboration for creative design [13,44].…”
Section: Human-ai Collaboration For Creative Designmentioning
confidence: 99%
“…Generative models, including generative adversarial networks (GANs) [19], variational auto-encoders (VAEs) [28], are now integral to AI-supported design processes (e.g., [6,45]). Simultaneously, diffusion models, renowned for their ability to produce rich and diverse samples, have found applications in areas such as artistic style transfer [73]. With its fast-evolving capabilities, AI has been playing an increasing role in human-AI collaboration for creative design [13,44].…”
Section: Human-ai Collaboration For Creative Designmentioning
confidence: 99%
“…Following the previous works on music style transfer (Alinoori and Tzerpos 2022; Cífka et al 2021), we evaluate our method based on two criteria: (a) content preservation and (b) style fit. Taking inspiration from MUSIC-GEN (Copet et al 2023) and InST (Zhang et al 2023b), we compute the CLAP cosine similarity between the generated mel-spectrograms and the content mel-spectrograms to evaluate content preservation. Additionally, we calculate the CLAP cosine similarity between the generated melspectrograms and the corresponding textual description of the style to evaluate style fit.…”
Section: Quantitative Evaluationmentioning
confidence: 99%
“…The model requires retraining for different sets of images. Zhang et al [13] adopt a similar approach and apply the text inversion method to image style transfer. However, retraining the network for each style can be time-consuming.…”
Section: Inversion Of the Diffusion Modelmentioning
confidence: 99%
“…Recently, with the development of Diffusion Models (DMs) [7][8][9][10], there have been a few attempts to use DMs to render content images via textual style conditions. With their outstanding ability to produce rich stylizations, many DMs-based methods [11][12][13] produce high quality results. Furthermore, text-driven image stylization is feasible through image editing techniques [14][15][16][17].…”
Section: Introductionmentioning
confidence: 99%