2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) 2019
DOI: 10.1109/iccvw.2019.00397
|View full text |Cite
|
Sign up to set email alerts
|

Powering Virtual Try-On via Auxiliary Human Segmentation Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(9 citation statements)
references
References 4 publications
0
9
0
Order By: Relevance
“…They train a separate refinement network to combine the warp and the target image. VTNFP [53] extends the work by incorporatiing body segments prediction and later works follow similar procedure [37,24,42,22,2]. However, TPS transformation fails to produce reasonable warps, due to the noisiness of generated masks in our dataset, as shown in Figure 6 right.…”
Section: Related Workmentioning
confidence: 74%
See 1 more Smart Citation
“…They train a separate refinement network to combine the warp and the target image. VTNFP [53] extends the work by incorporatiing body segments prediction and later works follow similar procedure [37,24,42,22,2]. However, TPS transformation fails to produce reasonable warps, due to the noisiness of generated masks in our dataset, as shown in Figure 6 right.…”
Section: Related Workmentioning
confidence: 74%
“…The VITON dataset [17] contains pairs of product image (front-view, laying flat, white background) and studio images, 2D pose maps and pose key-points. It has been used by many works [45,11,15,53,24,22,2,37]. Some works [47,15,13,51] on multi-pose matching used DeepFashion [33] or MVC [32] and other self-collected datasets [12,21,47,55].…”
Section: Datasetsmentioning
confidence: 99%
“…The comparisons of VITON and CP-VTON are given in Figure 6(c). There were several improved works [4,76,210,220] based on CP-VTON. Different from the previous works that needed the in-shop clothing image for virtual try-on, FashionGAN [231] and M2E-TON [195] presented target try-on clothing image based on text description and model image respectively.…”
Section: State-of-the-art Methodsmentioning
confidence: 99%
“…The proposed PGN integrates two twinned subtasks that can be mutually refined under a unified network, i.e., semantic part segmentation, and instance-aware edge detection. Further, Ruan et al [152] proposed CE2P framework 4 containing three key modules, a high-resolution embedding module, a global context embedding module, and an edge perceiving module, for single human parsing. This work won first place within all three human parsing tracks in the seond Look Into Person (LIP) Challenge.…”
Section: 21mentioning
confidence: 99%
“…VITON [16] follows the idea of image generation and uses non-parametric geometric transform which makes all the procedure twostage, similar to SwapNet [48] with the difference in the task statement and training data. CP-VTON [56] further improves upon [16] by incorporating a full learnable thin-plate spline transformation, followed by CP-VTON+ [40], LA-VITON [22], Ayush et al [5] and ACGPN [60]. While the above-mentioned works rely on pre-trained human parsers and pose estimators, the recent work of Issenhuth er al.…”
Section: Modeling Clothing Appearancementioning
confidence: 99%