2021
DOI: 10.48550/arxiv.2105.14576
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

StyTr$^2$: Image Style Transfer with Transformers

Abstract: The goal of image style transfer is to render an image with artistic features guided by a style reference while maintaining the original content. Due to the locality and spatial invariance in CNNs, it is difficult to extract and maintain the global information of input images. Therefore, traditional neural style transfer methods are usually biased and content leak can be observed by running several times of the style transfer process with the same reference style image. To address this critical issue, we take … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(10 citation statements)
references
References 36 publications
0
6
0
Order By: Relevance
“…We demonstrated the effectiveness of our method on different types of conditional image generation tasks. SOTA pre-trained deterministic models, including LaMa (Suvorov et al 2022) for inpainting and StyTr 2 (Deng et al 2021) for style transfer, are used to qualitatively and quantitatively validate the feasibility and effectiveness of our proposed method. In addition, we also conducted experiments on tasks such as super-resolution, dehazing, and probabilistic generation to further discuss the generalizability and limitations of our proposed method.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We demonstrated the effectiveness of our method on different types of conditional image generation tasks. SOTA pre-trained deterministic models, including LaMa (Suvorov et al 2022) for inpainting and StyTr 2 (Deng et al 2021) for style transfer, are used to qualitatively and quantitatively validate the feasibility and effectiveness of our proposed method. In addition, we also conducted experiments on tasks such as super-resolution, dehazing, and probabilistic generation to further discuss the generalizability and limitations of our proposed method.…”
Section: Methodsmentioning
confidence: 99%
“…(Sanakoyeu et al 2018) introduced GAN structure for style transfer. Subsequent works improve the performance of neural style transfer in many aspects, including quality (Deng et al 2021) and generalization (Chiu 2019). (Dong et al 2015) took the lead in introducing learning-based method into well-posed vision tasks, e.g., super-resolution, denoising, and JPEG compression artifact reduction.…”
Section: Introductionmentioning
confidence: 99%
“…Token adoptation in vision tasks: At the moment, tokenbased models are widely applied in almost all domains in vision, including classification [22,46,65], object detection [6,16,90], segmentation [23,74], image generation [5,20,24,38,43], video understanding [1,2,4,9,25,28,41,45,47,49,56,85], dense prediction [54,75], point clouds processing [30,88], reinforcement learning [10,37] and tracking [60].…”
Section: Related Workmentioning
confidence: 99%
“…Multiple-Style-Per-Model NST methods have included Dumoulin et al [29], Li et al [30] and Zhang and Dana [31]. Finally GANs [32], CycleGANs [33] and image transformers [34] have been recently used for NST. Although there have been many advances in this field, the Gatys method is still considered to be the gold standard by most researchers in terms of the quality of its results [20].…”
Section: Feature Map Fusion Using Image Optimisationmentioning
confidence: 99%