纺织学报 ›› 2023, Vol. 44 ›› Issue (01): 171-178.doi: 10.13475/j.fzxb.20211104908

• 服装工程 • 上一篇    下一篇

手绘草图到服装图像的跨域生成

陈佳1,2, 杨聪聪1, 刘军平1(), 何儒汉1,2, 梁金星1,2   

  1. 1.武汉纺织大学 计算机与人工智能学院, 湖北 武汉 430200
    2.湖北省服装信息化工程技术研究中心, 湖北 武汉 430200
  • 收稿日期:2021-11-09 修回日期:2022-09-21 出版日期:2023-01-15 发布日期:2023-02-16
  • 通讯作者: 刘军平(1979—),男,副教授,博士。主要研究方向为人工智能、工业大数据与计算机仿真。E-mail:jpliu@wtu.edu.cn
  • 作者简介:陈佳(1982—),女,副教授,博士。主要研究方向为图像处理与模式识别。
  • 基金资助:
    国家自然科学基金项目(62202345);湖北省自然科学基金计划一般面上项目(2020CFB801)

Cross-domain generation for transferring hand-drawn sketches to garment images

CHEN Jia1,2, YANG Congcong1, LIU Junping1(), HE Ruhan1,2, LIANG Jinxing1,2   

  1. 1. School of Computer Science and Artificial Intelligence, Wuhan Textile University, Wuhan, Hubei 430200, China
    2. Engineering Research Center of Hubei Province for Clothing Information, Wuhan, Hubei 430200, China
  • Received:2021-11-09 Revised:2022-09-21 Published:2023-01-15 Online:2023-02-16

摘要:

针对基于手绘草图的服装图像生成质量较低以及服装图像丰富的语义属性与视觉细节难以利用的问题,提出一种基于手绘草图的服装图像生成方法AGGAN。利用深度学习技术中的强大生成模型生成式对抗网络与注意力机制,对服装草图与服装图像数据进行学习,通过属性融入模块将服装属性进行One-hot编码后得到AdaIN参数并融入到生成对抗网络模型中,训练模型学习服装图像与其视觉属性之间的对应关系,使得模型能够在输入条件为服装属性的情况下生成相应的服装图像。对比了AGGAN与其它图像生成方法在输入为服装草图时生成服装图像的效果,结果表明:AGGAN的弗雷切特初始距离FID值得分相较于无监督图像生成模型CycleGAN降低了26.2%,初始分数IS值则提高了13.8%,明显提升了生成服装图像的多样性与保真度。

关键词: 服装设计, 手绘草图, 深度学习, 图像生成, 生成式对抗网络, 注意力机制

Abstract:

Objective Garment image synthesis is an important part of the garment design and manufacturing process, which uses artificial intelligence technology to automatically generate realistic garment images. Garment design relies heavily on the subjective will of the designer, which often needs to be manually achieved by designers. However, this process is time-consuming and quite inefficient. In the context of artificial intelligence, garment image synthesis can significantly improve efficiency by automatically generating garment images. In addition, it has a wide range of applications in virtual try-on, fashion image manipulation and fashion presentation. Therefore, garment image synthesis has received a lot of attention.
Method The garment sketch was guided to automatically generate the corresponding garment image by entering the garment attributes. A garment image generation method based on hand-drawn sketches was proposed, namely AGGAN. Generative adversarial networks with attention mechanism was adopted to learn garment sketches and garment image data to obtain AdaIN parameters after One-hot encoding of garment attributes through the attribute incorporation module, which are incorporated into the model, and the model was trained to learn the correspondence between garment images and their visual attributes, thus can generate corresponding garment images under the guidance of given garment attributes.
Results AGGAN was qualitatively compared with some existing image generation methods (Fig.2). By comparing with all baselines, the AGGAN proposed not only generates garment images with multiple colors, but also generates images closer to the real situation in terms of visual effects. In addition, IS (inception score), FID (fréchet inception distance), and MOS (mean opinion score) was useds for further quantitatively evaluating the model. The IS value of the garment images generated by the method prosposed is 1.253 (Tab.1), which is 13.8% higher than CycleGAN (cycle-consistent generative adversarial networks ) value, and higher than the values from using other methods. The FID value is 139.634, which is 26.2% lower than CycleGAN, and lower than other methods. In addition to the above two evaluation methods, MOS was adopted to evaluate the quality of garment images generated by each method, the MOS score obtained by the method prosposed is 4.352, which is higher than other image generation methods. In order to control the generation of garment images more flexibly, experiments was conducted on attribute-guided garment image synthesis. The garment sketch was controlled by garment attributes to synthesize the corresponding garment image, and the generated garment image has obvious changes in the sleeve length part, which does not seem to be particularly incongruous (Fig.3). The effect of the color attribute on the generated garment images was also explored. Several common color attributes was chosen in the experiments, and it can be seen that AGGAN can generated almost any corresponding color and high-fidelity garment images under the control of color attributes (Fig.4). Texture is also the most intuitive and the main visual feature of the garment image, and several texture attributes was selected in the experiments (Fig.5). From the figure it can be seen that the generation results are more obvious, basically the required texture can be generated, although further improvement is necessary in terms of realism.
Conclusion The research constructed a garment image generation model based on hand-drawn sketches through the attribute incorporation module, attention mechanism and CycleGAN. Combining the advantages of generative adversarial networks and conditional image generation methods, it took garment attributes as conditions to improve the controllability of the garment image generation process, which helps garment designers to achieve automated garment image synthesis. After a series of experiments, the feasibility and effectiveness of the text method were proved. The method proposed provides new ideas for computer-aided garment design. Some improvements should to be made, for example, the generated garment images cannot generate texture attributes effectively, and there are fewer garment attributes studied.

Key words: fashion design, hand-drawn sketch, deep learning, image generation, generative adversarial network, attention mechanism

中图分类号: 

  • TS942.8

图1

正向的AGGAN框架"

图2

AGGAN与其它方法生成服装图像结果对比"

表1

AGGAN与其它方法比较结果"

图像生成方法 IS值 FID值 MOS值/分
CycleGAN 1.101 189.078 3.021
MUNIT 1.178 211.793 3.222
USPS 1.142 147.486 3.432
AGGAN 1.253 139.634 4.352

图3

袖长属性控制的生成结果"

图4

颜色属性控制的生成结果"

图5

纹理属性控制的生成结果"

[1] CHEN T, CHENG M M, TAN P, et al. Sketch2photo: internet image montage[J]. ACM Transactions on Graphics (TOG), 2009, 28(5): 1-10.
[2] EITZ M, RICHTER R, HILDEBRAND K, et al. Photosketcher: interactive sketch-based image synthesis[J]. IEEE Computer Graphics and Applications, 2011, 31(6): 56-66.
doi: 10.1109/MCG.2011.67 pmid: 24808259
[3] CHEN W, JAMES H. Sketchygan: towards diverse and realistic sketch to image synthesis[C] // 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Salt Lake City: IEEE Computer Society 2018: 9416-9425.
[4] GOODFELLOW I J, POUGET-ABADIE J, MIRZA M, et al. Generative adversarial networks[J]. Communications of the ACM, 2020, 63(11): 139-144.
doi: 10.1145/3422622
[5] ZHU J Y. Unpaired image-to-image translation using cycle-consistent adversarial networks[C] // PARK T, ISOLA P, EFROS A A, et al. 2017 IEEE International Conference on Computer Vision (ICCV). Venice: IEEE Computer Society, 2017: 2223-2232.
[6] HUANG X, BELONGIE S. Arbitrary style transfer in real-time with adaptive instance normalization[C] // 2017 IEEE International Conference on Computer Vision (ICCV). Venice: IEEE Computer Society, 2017: 1510-1519.
[7] PING Q. Fashion-attgan: attribute-aware fashion editing with multi-objective GAN[C]// WU B, DING W Y, YUAN J B, et al. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). Long Beach: IEEE Computer Society, 2019: 323-325.
[8] HEUSEL M. Gans trained by a two time-scale update rule converge to a local nash equilibrium[C] // RAMSAUER H, UNTERTHINER T, NESSLER B, et al. Proceedings of the 31st International Conference on Neural Information Processing Systems. Long Beach: Curran Associates Inc., 2017: 6629-6640.
[9] WANG T C. High-resolution image synthesis and semantic manipulation with conditional gans[C] // LIU M Y, ZHU J Y, TAO A, et al. Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Salt Lake City: IEEE Computer Society, 2018: 8798-8807.
[10] SZEGEDY C. Rethinking the inception architecture for computer vision[C] // VANHOUCKE V, IOFFE S, SHLENS J, et al. Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas: IEEE Computer Society, 2016: 2818-2826.
[11] VAN ERVEN T, HARREMOS P. RÉnyi divergence and kullback-leibler divergence[J]. IEEE Transactions on Information Theory, 2014, 60(7): 3797-3820.
doi: 10.1109/TIT.2014.2320500
[12] HUANG X. Multimodal unsupervised image-to-image translation[C] // LIU M Y, BELONGIE S, KAUTZ J, et al. Proceedings of 15th European Conference on Computer Vision (ECCV). Cham: Springer International Publishing, 2018: 172-189.
[13] LIU R. Unsupervised sketch to photo synthesis[C] // YU Q, YU S X. Proceedings of 16th European Conference on Computer Vision (ECCV). Cham: Springer International Publishing, 2022: 36-52.
[1] 王斌, 李敏, 雷承霖, 何儒汉. 基于深度学习的织物疵点检测研究进展[J]. 纺织学报, 2023, 44(01): 219-227.
[2] 安亦锦, 薛文良, 丁亦, 张顺连. 基于图像处理的纺织品耐摩擦色牢度评级[J]. 纺织学报, 2022, 43(12): 131-137.
[3] 顾梅花, 刘杰, 李立瑶, 崔琳. 结合特征学习与注意力机制的服装图像分割[J]. 纺织学报, 2022, 43(11): 163-171.
[4] 陈金广, 李雪, 邵景峰, 马丽丽. 改进YOLOv5网络的轻量级服装目标检测方法[J]. 纺织学报, 2022, 43(10): 155-160.
[5] 王迪, 柯莹, 王宏付. Voronoi图形在参数化服装造型构建中的应用[J]. 纺织学报, 2021, 42(12): 131-137.
[6] 江慧, 马彪. 基于服装风格的款式相似度算法[J]. 纺织学报, 2021, 42(11): 129-136.
[7] 金鹏, 薛哲彬, 江润恬, 刘丹宇, 张弛. 具有安全防护功能的智能盲人服设计[J]. 纺织学报, 2021, 42(08): 135-143.
[8] 唐茜, 张冰冰, 郑笑雨. 婴幼儿可穿戴智能监测服装设计[J]. 纺织学报, 2021, 42(08): 156-160.
[9] 杨争妍, 薛文良, 张传雄, 丁亦, 马颜雪. 基于生成式对抗网络的用户下装搭配推荐[J]. 纺织学报, 2021, 42(07): 164-168.
[10] 徐增波, 张玲, 张艳红, 陈桂清. 基于复杂网络提取和支持向量机模型分类的服装领型研究[J]. 纺织学报, 2021, 42(06): 146-152.
[11] 张恒. 基于翻领松量结构模型的翻折领结构设计方法[J]. 纺织学报, 2020, 41(11): 128-135.
[12] 杨娟, 张远鹏. 服装设计知识图谱中的服装装饰工艺分类模型[J]. 纺织学报, 2020, 41(08): 95-100.
[13] 刘佟. 基于人体曲面变化的服装几何空间设计[J]. 纺织学报, 2019, 40(11): 155-160.
[14] 许倩, 陈敏之. 基于深度学习的服装丝缕平衡性评价系统[J]. 纺织学报, 2019, 40(10): 191-195.
[15] 吴传彬, 刘骊, 付晓东, 刘利军, 黄青松. 结合显著区域检测和手绘草图的服装图像检索[J]. 纺织学报, 2019, 40(07): 174-181.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!