2023
DOI: 10.48550/arxiv.2303.05125
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Cones: Concept Neurons in Diffusion Models for Customized Generation

Abstract: parameters, which reduces storage consumption by 90% compared with previous subject-driven generation methods. Extensive qualitative and quantitative studies on diverse scenarios show the superiority of our method in interpreting and manipulating diffusion models. Our code is available at https://github.com/Johanan528/Cones.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 22 publications
0
6
0
Order By: Relevance
“…Later, non-fine-tuning methods for customized generation emerged. Cones (Liu et al 2023) focuses on identifying the effective concept neurons related to the target concept, while ViCo (Hao et al 2023) proposes a plug-in image attention module to adjust the diffusion process. Other works (Wei et al 2023;Shi et al 2023;Li, Hou, and Loy 2023) explore achieving customized generation without finetuning.…”
Section: Related Workmentioning
confidence: 99%
“…Later, non-fine-tuning methods for customized generation emerged. Cones (Liu et al 2023) focuses on identifying the effective concept neurons related to the target concept, while ViCo (Hao et al 2023) proposes a plug-in image attention module to adjust the diffusion process. Other works (Wei et al 2023;Shi et al 2023;Li, Hou, and Loy 2023) explore achieving customized generation without finetuning.…”
Section: Related Workmentioning
confidence: 99%
“…Text-to-image diffusion models Diffusion models [10,19,21,41,[58][59][60][61][62] have proven to be highly effective in learning data distributions and have shown impressive results in image synthesis, leading to various applications [8,26,27,29,31,32,36,46,56,74]. Recent advancements have also explored transformer-based architectures [6,45,67].…”
Section: Related Workmentioning
confidence: 99%
“…Based on diffusion models, [7,17] have led to techniques like using placeholder words for object representation, enabling highfidelity customizations. Subsequent works [19,26,34,35] extend this by fine-tuning pretrained text-to-image models for new concept learning. These advancements have facilitated diverse applications, such as subject swapping [8], open-world generation [22], and non-rigid image editing [2].…”
Section: Subject-driven Image Generationmentioning
confidence: 99%