Journal of Textile Research ›› 2023, Vol. 44 ›› Issue (05): 177-183.doi: 10.13475/j.fzxb.20220403101

• Apparel Engineering • Previous Articles     Next Articles

High-precision intelligent algorithm for virtual fitting based on texture feature learning

LIU Yuye, WANG Ping()   

  1. College of Information Science and Technology, Donghua University, Shanghai 201620, China
  • Received:2022-04-07 Revised:2022-11-30 Online:2023-05-15 Published:2023-06-09

Abstract:

Objective Virtual fitting provides users with a digital and interactive fashion fitting experience and meets the requirements for garment customization in the fashion industry by using machine vision, artificial intelligence and other technologies. It has attracted keen attention from international brands and researchers. However, due to the influence of various posture, occlusion and interruption in non-fitting area, the existing virtual fitting methods still have problems, such as distortion, blurring and low accuracy. In order to overcome these problems, this paper proposed a high-precision virtual fitting model named as C-CGAN based on texture feature learning.

Method A garment reconstruction network based on the idea of CGAN was proposed, which used the garment mask positioning and garment texture constraints to learn intelligently the garment reconstruction model under various postures. The encoder-decoder network was utilized to fuse the reconstructed garment and character features. In addition, a variety of comprehensive loss functions were employed to optimize the network performance. A rich texture dataset was eventually constructed based on the international virtual fitting dataset, followed by the development of a garment fitting system in PyTorch environment and its performance evaluation.

Results The results of C-CGAN showed more significant FID (Fréchet distance) and IS (initial score) optimization effect than that of the newly reported VITON and CP-VTON statistical metrics (Tab.2). However, the PSNR (peak signal to noise ratio) accuracy of CP-VTON was still low, which means it had a lot of distortion. Compared with CP-VTON, in the case of comparable IS, the FID of C-CGAN was reduced by about 11%, the SSIM (structural similarity) is increased by about 25%, and the PSNR was increased by about 78%. Therefore, the performance metrics of this network had significant advantages. In order to compare the visual fitting effect, CP-VTON and C-CGAN were both adopted to synthesize the texture of the model's original tops on the test dataset for comparison of the subjective visual similarity between the virtual fitting results and the real sample in dataset. The comparison results of the virtual fitting (Fig.7) in 9 difficult scenes (Tab.1) showed that CP-VTON was prone to large deformation distortion for some complex textures, such as stripes and wave points, and the model's arm was distorted when occluded. In contrast, C-CGAN was shown to be able to suppress effectively the interference of occlusion and garment texture, truly and exquisitely preserve the details of characters and texture, and had a higher similarity with real samples. Furthermore, in order to verify the applicability of this method in practical applications, a model in test dataset was selected whose original top's texture is light pinstripe. There were ups and downs and pleats at the model's front and waist, respectively, relating to her posture. The virtual garment replacement preview results of seven textures (Fig.8) showed that textured details and features varied on the model's chest and waist corresponding to the posture, such as the fold changes of pure color, the density changes of the wave point and the waveform variation of the stripe. In addition, C-CGAN was shown to preserve well the model characteristics of models and clothing characteristics of other areas.

Conclusion This paper presented extensive qualitative and quantitative evaluations on the C-CGAN method. The statistical metrics on the test dataset show that the similarity between the C-CGAN virtual fitting results and the real samples is higher, the accuracy is higher, and the distortion is smaller. The subjective visual comparison results of virtual fitting show that C-CGAN has better adaptability to difficult fitting scenes such as stripes, wave points and occlusion, and the reconstructed texture is more natural and delicate, with high matching sense of human posture and good adaptability. The virtual garment replacement preview test results show that C-CGAN can generate texture deformation adapted to human posture for color, stripe and wave point, and the generated image is clear. C-CGAN can provide a realistic virtual fitting experience, which can be widely used in digital fashion application scenarios such as interactive texture reloading and garment assisted design.

Key words: conditional generative adversarial network, encoder-decoder network, positioning and reconstruction, virtual fitting, garment customization

CLC Number: 

  • TS942.8

Fig.1

Fitting results in other documents. (a) Comparison result of VITON and CP-VTON; (b) Comparison result of TextureGAN"

Fig.2

Structure of GAN network"

Fig.3

Structure of CGAN network"

Fig.4

Garment reconstruction using CGAN network"

Fig.5

Structure of encoder-decoder network"

Fig.6

System flowchart of C-CGAN"

Fig.7

Comparison results of virtual fitting. (a) Actual images; (b)Textures of model's top;(c)Fitting results of C-CGAN;(d)Fitting results of CP-VTON"

Tab.1

Types of virtual fitting scene in Fig.7"

列数 有无遮挡 纹理类型 试穿区域
第1列 纯色 长袖
第2列 纯色 短袖
第3、5列 条纹 长袖
第4列 方格 短袖
第6列 波点 短袖
第7列 波点 长袖
第8、9列 条纹 短袖

Tab.2

Quality comparison results of fitting images"

方法 IS FID SSIM PSNR
VITON[13] 2.290 55.710 0.740 /
CP-VTON[14] 2.660 20.331 0.698 14.544
本文C-CGAN 2.535 18.080 0.871 25.907

Fig.8

Preview results of C-CGAN virtual garment replacement; (a)Source of texture;(b)Personalized model;(c) Fitting results"

[1] 李浩, 顾力文, 顾雯, 等. 基于消费者感知价值的线上线下服装定制模式[J]. 纺织学报, 2020, 41(9):128-135.
LI Hao, GU Liwen, GU Wen, et al. Research on online-to-offline clothing customization mode based on consumer perceived value[J]. Journal of Textile Research, 2020, 41(9):128-135.
[2] 江红霞, 黄智威, 刘基宏. 基于模块化划分的旗袍虚拟展示[J]. 纺织学报, 2021, 42(5):138-142.
JIANG Hongxia, HUANG Zhiwei, LIU Jihong. Virtual display of cheongsam based on modularization[J]. Journal of Textile Research, 2021, 42(5):138-142.
[3] 冀艳波, 王玲丽, 刘凯旋. 基于数字化三维人体模型的旗袍定制设计[J]. 纺织学报, 2021, 42(1):133-137.
JI Yanbo, WANG Lingli, LIU Kaixuan. Custom design of cheongsam based on digital 3-D human model[J]. Journal of Textile Research, 2021, 42(1):133-137.
[4] 黎博文, 王萍, 刘玉叶. 基于人体动态特征的三维服装虚拟试穿技术[J]. 纺织学报, 2021, 42(9):144-149.
LI Bowen, WANG Ping, LIU Yuye. 3-D virtual try-on technique based on dynamic feature of body postures[J]. Journal of Textile Research, 2021, 42(9): 144-149.
[5] HAN X, WU Z, WU Z, et al. VITON: an image-based virtual try-on network[C]// LIU W, KAUTZ J, WANG X, et al. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2018:7543-7552.
[6] SONG D, LI T, MAO Z, et al. SP-VITON: shape-preserving image-based virtual try-on network[J]. Multimedia Tools and Applications, 2020, 79(4): 33757-33769.
doi: 10.1007/s11042-019-08363-w
[7] WANG B, ZHENG H, LIANG X, et al. Toward characteristic-preserving image-based virtual try-on network[C]// FERRARI V, HEBERT M, SMINCHISESCU C, et al. Proceedings of the European conference on computer vision (ECCV). Cham: Springer, 2018:589-604.
[8] AYUSH K, JANDIAL S, CHOPRA A, et al. Robust cloth warping via multi-scale patch adversarial loss for virtual try-on framework[C]// TIMODFTE R, GU S, DANELLJAN M, et al. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW). Piscataway: IEEE, 2019:1279-1281.
[9] AYUSH K, JANDIAL S, CHOPRA A, et al. Powering virtual try-on via auxiliary human segmentation learning[C]// TIMODFTE R, GU S, DANELLJAN M, et al. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW). Piscataway: IEEE, 2019:3193-3196.
[10] XIAN W, SANGKLOY P, AGRAWAL V, et al. TextureGAN: controlling deep image synthesis with texture patches[C]// LIU W, KAUTZ J, WANG X, et al. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2018:8456-8465.
[11] 叶晨, 关玮. 生成式对抗网络的应用综述[J]. 同济大学学报(自然科学版), 2020, 48(4):591-601.
YE Chen, GUAN Wei. A review of application of generative adversarial networks[J]. Journal of Tongji University(Natural Science), 2020, 48(4):591-601.
[12] 邵杰, 黄茜, 曹坤涛. 基于深度学习的人体解析研究综述[J]. 电子科技大学学报, 2019, 48(5):644-654.
SHAO Jie, HUANG Xi, CAO Kuntao. A review on deep learning techniques applied to human parsing[J]. Journal of University of Electronic Science and Technology of China, 2019, 48(5):644-654.
[13] GE C, SONG Y, GE Y, et al. Disentangled cycle consistency for highly-realistic virtual try-on[C]// ZHANG L, SUN J, SHEN C, et al. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Piscataway: IEEE, 2021:16923-16932.
[14] JANDIAL S, CHOPRA A, AYUSH K, et al. SieveNet: a unified framework for robust image-based virtual try-on[C]// YUILLE A, RAI P, BISWAS S, et al. 2020 IEEE Winter Conference on Applications of Computer Vision (WACV). Piscataway: IEEE, 2020:2171-2179.
[15] 曹玉东, 刘海燕, 贾旭, 等. 基于深度学习的图像质量评价方法综述[J]. 计算机工程与应用, 2021, 57(23): 27-36.
doi: 10.3778/j.issn.1002-8331.2106-0228
CAO Yudong, LIU Haiyan, JIA Xu, et al. Overview of image quality assessment method based on deep learning[J]. Computer Engineering and Applications, 2021, 57(23): 27-36.
doi: 10.3778/j.issn.1002-8331.2106-0228
[1] LU Hong, SONG Jiayi, LI Yuanyuan, TENG Junfeng. Pattern design of inner rotation styles based on fitted two-piece sleeve [J]. Journal of Textile Research, 2022, 43(08): 140-146.
[2] WANG Chunru, YUAN Yue, CAO Xiaomeng, FAN Yilin, ZHONG Anhua. Influence of structural parameters of stand collar on clothing styling [J]. Journal of Textile Research, 2022, 43(03): 153-159.
[3] WANG Ting, GU Bingfei. 3-D modeling of neck-shoulder part based on human photos [J]. Journal of Textile Research, 2021, 42(01): 125-132.
[4] CHEN Mi, YE Qinwen, ZHANG Gaopeng. Construction of parametric structure model for bias-cut skirt pattern [J]. Journal of Textile Research, 2020, 41(07): 135-140.
[5] XU Qian, CHEN Minzhi. Garment grain balance evaluation system based on deep learning [J]. Journal of Textile Research, 2019, 40(10): 191-195.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!