纺织学报 ›› 2023, Vol. 44 ›› Issue (05): 177-183.doi: 10.13475/j.fzxb.20220403101

• 服装工程 • 上一篇    下一篇

基于纹理特征学习的高精度虚拟试穿智能算法

刘玉叶, 王萍()   

  1. 东华大学 信息科学与技术学院, 上海 201620
  • 收稿日期:2022-04-07 修回日期:2022-11-30 出版日期:2023-05-15 发布日期:2023-06-09
  • 通讯作者: 王萍(1973—),女,教授,博士。主要研究方向为5G物联网、智能视觉、服装虚拟试穿。E-mail:pingwang@dhu.edu.cn。
  • 作者简介:刘玉叶(1996—),女,硕士生。主要研究方向为服装虚拟试穿。

High-precision intelligent algorithm for virtual fitting based on texture feature learning

LIU Yuye, WANG Ping()   

  1. College of Information Science and Technology, Donghua University, Shanghai 201620, China
  • Received:2022-04-07 Revised:2022-11-30 Published:2023-05-15 Online:2023-06-09

摘要:

为更好地适应数字化时尚消费新趋势,针对虚拟试穿中人体姿态、遮挡、非试穿区串扰引起的重构纹理模糊、细节失真、相似度低等难题,提出一种高精度虚拟试穿C-CGAN智能模型。首先,利用服装蒙版定位和纹理约束,采用CGAN网络框架智能学习得到多姿态下的服装重构模型。然后,采用编解码网络融合重构服装与人体特征,并设计多种针对性的损失函数优化网络参数。最后,基于国际流行虚拟试穿模特样本库构建丰富的纹理数据集,在PyTorch环境下开发了服装虚拟试穿系统。对比测试结果表明:C-CGAN较CP-VTON生成图像质量指标,结构相反性提高约25%,弗雷歇距离降低约11%,峰值信噪比提高约78%,重构纹理自然细腻、图像清晰,场景适应性良好,可广泛用于个性化时尚定制、在线试穿等领域。

关键词: 条件生成对抗网络, 编解码网络, 定位重构, 虚拟试穿, 服装个性化定制

Abstract:

Objective Virtual fitting provides users with a digital and interactive fashion fitting experience and meets the requirements for garment customization in the fashion industry by using machine vision, artificial intelligence and other technologies. It has attracted keen attention from international brands and researchers. However, due to the influence of various posture, occlusion and interruption in non-fitting area, the existing virtual fitting methods still have problems, such as distortion, blurring and low accuracy. In order to overcome these problems, this paper proposed a high-precision virtual fitting model named as C-CGAN based on texture feature learning.

Method A garment reconstruction network based on the idea of CGAN was proposed, which used the garment mask positioning and garment texture constraints to learn intelligently the garment reconstruction model under various postures. The encoder-decoder network was utilized to fuse the reconstructed garment and character features. In addition, a variety of comprehensive loss functions were employed to optimize the network performance. A rich texture dataset was eventually constructed based on the international virtual fitting dataset, followed by the development of a garment fitting system in PyTorch environment and its performance evaluation.

Results The results of C-CGAN showed more significant FID (Fréchet distance) and IS (initial score) optimization effect than that of the newly reported VITON and CP-VTON statistical metrics (Tab.2). However, the PSNR (peak signal to noise ratio) accuracy of CP-VTON was still low, which means it had a lot of distortion. Compared with CP-VTON, in the case of comparable IS, the FID of C-CGAN was reduced by about 11%, the SSIM (structural similarity) is increased by about 25%, and the PSNR was increased by about 78%. Therefore, the performance metrics of this network had significant advantages. In order to compare the visual fitting effect, CP-VTON and C-CGAN were both adopted to synthesize the texture of the model's original tops on the test dataset for comparison of the subjective visual similarity between the virtual fitting results and the real sample in dataset. The comparison results of the virtual fitting (Fig.7) in 9 difficult scenes (Tab.1) showed that CP-VTON was prone to large deformation distortion for some complex textures, such as stripes and wave points, and the model's arm was distorted when occluded. In contrast, C-CGAN was shown to be able to suppress effectively the interference of occlusion and garment texture, truly and exquisitely preserve the details of characters and texture, and had a higher similarity with real samples. Furthermore, in order to verify the applicability of this method in practical applications, a model in test dataset was selected whose original top's texture is light pinstripe. There were ups and downs and pleats at the model's front and waist, respectively, relating to her posture. The virtual garment replacement preview results of seven textures (Fig.8) showed that textured details and features varied on the model's chest and waist corresponding to the posture, such as the fold changes of pure color, the density changes of the wave point and the waveform variation of the stripe. In addition, C-CGAN was shown to preserve well the model characteristics of models and clothing characteristics of other areas.

Conclusion This paper presented extensive qualitative and quantitative evaluations on the C-CGAN method. The statistical metrics on the test dataset show that the similarity between the C-CGAN virtual fitting results and the real samples is higher, the accuracy is higher, and the distortion is smaller. The subjective visual comparison results of virtual fitting show that C-CGAN has better adaptability to difficult fitting scenes such as stripes, wave points and occlusion, and the reconstructed texture is more natural and delicate, with high matching sense of human posture and good adaptability. The virtual garment replacement preview test results show that C-CGAN can generate texture deformation adapted to human posture for color, stripe and wave point, and the generated image is clear. C-CGAN can provide a realistic virtual fitting experience, which can be widely used in digital fashion application scenarios such as interactive texture reloading and garment assisted design.

Key words: conditional generative adversarial network, encoder-decoder network, positioning and reconstruction, virtual fitting, garment customization

中图分类号: 

  • TS942.8

图1

其它文献中的试穿结果"

图2

GAN网络结构"

图3

CGAN网络结构"

图4

CGAN服装重构网络"

图5

编解码网络结构"

图6

C-CGAN系统流程图"

图7

虚拟试穿对比结果"

表1

图7中的虚拟试穿场景类型"

列数 有无遮挡 纹理类型 试穿区域
第1列 纯色 长袖
第2列 纯色 短袖
第3、5列 条纹 长袖
第4列 方格 短袖
第6列 波点 短袖
第7列 波点 长袖
第8、9列 条纹 短袖

表2

试穿图像质量的比测结果"

方法 IS FID SSIM PSNR
VITON[13] 2.290 55.710 0.740 /
CP-VTON[14] 2.660 20.331 0.698 14.544
本文C-CGAN 2.535 18.080 0.871 25.907

图8

C-CGAN虚拟换装预览结果"

[1] 李浩, 顾力文, 顾雯, 等. 基于消费者感知价值的线上线下服装定制模式[J]. 纺织学报, 2020, 41(9):128-135.
LI Hao, GU Liwen, GU Wen, et al. Research on online-to-offline clothing customization mode based on consumer perceived value[J]. Journal of Textile Research, 2020, 41(9):128-135.
[2] 江红霞, 黄智威, 刘基宏. 基于模块化划分的旗袍虚拟展示[J]. 纺织学报, 2021, 42(5):138-142.
JIANG Hongxia, HUANG Zhiwei, LIU Jihong. Virtual display of cheongsam based on modularization[J]. Journal of Textile Research, 2021, 42(5):138-142.
[3] 冀艳波, 王玲丽, 刘凯旋. 基于数字化三维人体模型的旗袍定制设计[J]. 纺织学报, 2021, 42(1):133-137.
JI Yanbo, WANG Lingli, LIU Kaixuan. Custom design of cheongsam based on digital 3-D human model[J]. Journal of Textile Research, 2021, 42(1):133-137.
[4] 黎博文, 王萍, 刘玉叶. 基于人体动态特征的三维服装虚拟试穿技术[J]. 纺织学报, 2021, 42(9):144-149.
LI Bowen, WANG Ping, LIU Yuye. 3-D virtual try-on technique based on dynamic feature of body postures[J]. Journal of Textile Research, 2021, 42(9): 144-149.
[5] HAN X, WU Z, WU Z, et al. VITON: an image-based virtual try-on network[C]// LIU W, KAUTZ J, WANG X, et al. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2018:7543-7552.
[6] SONG D, LI T, MAO Z, et al. SP-VITON: shape-preserving image-based virtual try-on network[J]. Multimedia Tools and Applications, 2020, 79(4): 33757-33769.
doi: 10.1007/s11042-019-08363-w
[7] WANG B, ZHENG H, LIANG X, et al. Toward characteristic-preserving image-based virtual try-on network[C]// FERRARI V, HEBERT M, SMINCHISESCU C, et al. Proceedings of the European conference on computer vision (ECCV). Cham: Springer, 2018:589-604.
[8] AYUSH K, JANDIAL S, CHOPRA A, et al. Robust cloth warping via multi-scale patch adversarial loss for virtual try-on framework[C]// TIMODFTE R, GU S, DANELLJAN M, et al. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW). Piscataway: IEEE, 2019:1279-1281.
[9] AYUSH K, JANDIAL S, CHOPRA A, et al. Powering virtual try-on via auxiliary human segmentation learning[C]// TIMODFTE R, GU S, DANELLJAN M, et al. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW). Piscataway: IEEE, 2019:3193-3196.
[10] XIAN W, SANGKLOY P, AGRAWAL V, et al. TextureGAN: controlling deep image synthesis with texture patches[C]// LIU W, KAUTZ J, WANG X, et al. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2018:8456-8465.
[11] 叶晨, 关玮. 生成式对抗网络的应用综述[J]. 同济大学学报(自然科学版), 2020, 48(4):591-601.
YE Chen, GUAN Wei. A review of application of generative adversarial networks[J]. Journal of Tongji University(Natural Science), 2020, 48(4):591-601.
[12] 邵杰, 黄茜, 曹坤涛. 基于深度学习的人体解析研究综述[J]. 电子科技大学学报, 2019, 48(5):644-654.
SHAO Jie, HUANG Xi, CAO Kuntao. A review on deep learning techniques applied to human parsing[J]. Journal of University of Electronic Science and Technology of China, 2019, 48(5):644-654.
[13] GE C, SONG Y, GE Y, et al. Disentangled cycle consistency for highly-realistic virtual try-on[C]// ZHANG L, SUN J, SHEN C, et al. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Piscataway: IEEE, 2021:16923-16932.
[14] JANDIAL S, CHOPRA A, AYUSH K, et al. SieveNet: a unified framework for robust image-based virtual try-on[C]// YUILLE A, RAI P, BISWAS S, et al. 2020 IEEE Winter Conference on Applications of Computer Vision (WACV). Piscataway: IEEE, 2020:2171-2179.
[15] 曹玉东, 刘海燕, 贾旭, 等. 基于深度学习的图像质量评价方法综述[J]. 计算机工程与应用, 2021, 57(23): 27-36.
doi: 10.3778/j.issn.1002-8331.2106-0228
CAO Yudong, LIU Haiyan, JIA Xu, et al. Overview of image quality assessment method based on deep learning[J]. Computer Engineering and Applications, 2021, 57(23): 27-36.
doi: 10.3778/j.issn.1002-8331.2106-0228
[1] 肖伯祥, 刘正东, 郭昱成, 王渊霞. 基于三维人体扫描模型的衬衣版型生成方法[J]. 纺织学报, 2022, 43(12): 151-159.
[2] 张健, 徐凯忆, 赵崧灵, 顾冰菲. 基于二维照片的青年男性颈肩部形态分类与识别[J]. 纺织学报, 2022, 43(05): 143-149.
[3] 黎博文, 王萍, 刘玉叶. 基于人体动态特征的三维服装虚拟试穿技术[J]. 纺织学报, 2021, 42(09): 144-149.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!