Virtual Try-On With Garment Self-Occlusion Conditions

被引:1
|
作者
Xing, Zhening [1 ]
Wu, Yuchen [1 ]
Liu, Si [2 ]
Di, Shangzhe [2 ]
Ma, Huimin [1 ]
机构
[1] Univ Sci & Technol Beijing, Sch Comp & Commuicat Engn, Beijing 100083, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Index Terms-Virtual try-on; generative adversarial network; self-occlusion;
D O I
10.1109/TMM.2022.3221346
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image-based virtual try-on focuses on changing the model's garment item to the target ones and preserving other visual features. To preserve the texture detail of the given in-shop garment, former methods use geometry-based methods (e.g., Thin-plate-spline interpolation) to realize garment warping. However, due to limited degree of freedom, geometry-based methods perform poorly when garment self-occlusion occurs, which is common in daily life. To address this challenge, we propose a novel occlusion-focused virtual try-on system. Compared to previous ones, our system contains three critical submodules, namely, Garment Part Modeling (GPM), a group of Garment Part Generators (GPGs), and Overlap Relation Estimator (ORE). GPM takes the pose landmarks as input, and progressively models the mask of body parts and garments. Based on these masks, GPGs are introduced to generate each garment part. Finally, ORE is proposed to model the overlap relationships between each garment part, and we bind the generated garments under the guidance of overlap relationships predicted by ORE. To make the most of extracted overlap relationships, we proposed an IoU-based hard example mining method for loss terms to handle the sparsity of the self-occlusion samples in the dataset. Furthermore, we introduce part affinity field as pose representation instead of landmark used widely by previous methods and achieve accuracy improvement on try-on layout estimation stage. We evaluate our model on the VITON dataset and found it can outperform previous approaches, especially on samples with garment self-occlusion.
引用
收藏
页码:7323 / 7336
页数:14
相关论文
共 21 条
  • [21] A SfM-based 3D face reconstruction method robust to self-occlusion by using a shape conversion matrix
    Lee, Sung Joo
    Park, Kang Ryoung
    Kim, Jaihie
    PATTERN RECOGNITION, 2011, 44 (07) : 1470 - 1486