Review of Image-Based Virtual Try-on: From Deep Learning to Diffusion Models

被引:0
作者
Yang, Haozhe [1 ]
Guo, Nan [1 ]
机构
[1] School of Computer Science and Engineering, Northeastern University, Shenyang
关键词
computer vision; diffusion model; image synthesis; virtual try-on; warping treatment;
D O I
10.3778/j.issn.1002-8331.2409-0340
中图分类号
学科分类号
摘要
Image-based virtual try-on, as an economical and convenient technology in the virtual try-on domain, aims to synthesize realistic fitting effects by combining model images with clothing images. It has significant applications in online shopping, fashion design, and animation. Recently, generative large models represented by diffusion models have driven new breakthroughs and transformations in the field due to their stronger generative capabilities compared to traditional deep learning methods. However, there is a lack of comprehensive analysis and overview of image-based virtual try-on in the era of large models. This paper summarizes the key techniques of image-based virtual try-on, categorizes mainstream methods into three baseline processes: data preprocessing, warping generation, and try-on result generation. It also analyzes the implementation methods used in representative literature, compares major process methods, and introduces commonly used datasets, evaluation metrics, and loss functions in image-based virtual try-on. Finally, the paper discusses the challenges and limitations of image-based virtual try-on in the context of large models, and outlines future development and improvement directions for relevant technologies. © 2025 Journal of Computer Engineering and Applications Beijing Co., Ltd.; Science Press. All rights reserved.
引用
收藏
页码:19 / 35
页数:16
相关论文
共 69 条
[1]  
SONG D, ZHANG X P, ZHOU J, Et al., Image-based virtual try-on: a survey, (2023)
[2]  
XU W W, UMENTANI N, CHAO Q W, Et al., Sensitivity-optimized rigging for example-based real-time clothing synthesis, ACM Transactions on Graphics, 33, 4, (2014)
[3]  
WANG L Y, LI H L, XIAO Q J, Et al., Automatic pose and wrinkle transfer for aesthetic garment display, Computer Aided Geometric Design, 89, (2021)
[4]  
WU N N, CHAO Q W, CHEN Y Z, Et al., AgentDress: real-time clothing synthesis for virtual agents using plausible deformations, IEEE Transactions on Visualization and Computer Graphics, 27, 11, pp. 4107-4118, (2021)
[5]  
PAN X Y, MAI J M, JIANG X W, Et al., Predicting loose-fitting garment deformations using bone- driven motion networks, Proceedings of the ACM SIGGRAPH 2022 Conference, (2022)
[6]  
JETCHEV N, BERGMANN U., The conditional analogy GAN: swapping fashion articles on people images, Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops, pp. 2287-2292, (2017)
[7]  
HAN X T, WU Z X, WU Z, Et al., VITON: an image-based virtual try- on network, Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7543-7552, (2018)
[8]  
WANG B C, ZHENG H B, LIANG X D, Et al., Toward characteristic-preserving image-based virtual try-on network, Proceedings of the 15th European Conference on Computer Vision, pp. 607-623, (2018)
[9]  
YU R Y, WANG X Q, XIE X H., VTNFP: an image-based virtual try-on network with body and clothing feature preservation, Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, pp. 10510-10519, (2019)
[10]  
ISSENHUTH T, MARY J, CALAUZENES C, Et al., End-to-end learning of geometric deformations of feature maps for virtual try-on, (2019)