Journal of Textile Research ›› 2026, Vol. 47 ›› Issue (1): 176-185.doi: 10.13475/j.fzxb.20250600501

• Dyeing and Finishing Engineering • Previous Articles     Next Articles

Cross-attention excited color-semantic consistent generation for textile digital printing patterns

ZHANG Hui1,2, ZHOU Xuan1,2, WANG Junwei2,3, DENG Yongmei2, ZHANG Kaibing2,3()   

  1. 1. School of Textile Science and Engineering, Xi'an Polytechnic University, Xi'an, Shaanxi 710048, China
    2. Shaanxi Key Laboratory of Clothing Intelligence, Xi'an, Shaanxi 710048, China
    3. School of Computer Science, Xi'an Polytechnic University, Xi'an, Shaanxi 710048, China
  • Received:2025-06-03 Revised:2025-11-10 Online:2026-01-15 Published:2026-01-15
  • Contact: ZHANG Kaibing E-mail:zhangkaibing@xpu.edu.cn

Abstract:

Objective Textile digital printing technology is popular in the textile dyeing and printing industry because of its high printing precision, fast production speed, and environmentally friendly processes. However, current digital printing pattern design still relies heavily on manual creation by designers, leading to long cycles, low efficiency, and poor market responsiveness. Leveraging Artificial Intelligence Generated Content (AIGC) technology to achieve rapid text-to-digital printing pattern generation can significantly improve the development efficiency of digital textile products and support user personalized customization, opening up new methods for digital pattern design. Nevertheless, mainstream text-to-image generation models (e.g., Stable Diffusion (SD) trained on generic datasets) have two key limitations, i.e. monotonous generated patterns lacking artistic diversity, and mismatch between the generated colors and the text-described color semantics due to inaccurate color attribute binding. Therefore, it is crucial to develop a specialized generative model for textile digital printing patterns and propose new methods to improve color generation accuracy.

Method A dataset containing 3 090 pairs of digital textile patterns and detailed textual descriptions was first constructed. The SD-V1.5 model was fine-tuned using the Low-Rank Adaptation (LoRA) technology. Subsequently, for the fine-tuned SD model, this study introduced a cross-attention-based color excitation module into the first 70% of the denoising process, without requiring any additional training. This module comprises two loss components, namely the target object color excitation loss and the background color excitation loss. By minimizing these two losses through gradient descent, the model learns to increase the attention weights of color-related keywords in the text prompt, thereby focusing more on the target and background colors of the pattern during generation. Finally, the color semantic consistency was evaluated using the CLIP Score, Text-Text Similarity (T2T-Sim), and a newly proposed Color-Text Accuracy (C2T-Acc) metric.

Results The effectiveness of the proposed method was evaluated through both qualitative and quantitative experiments. Qualitatively, the generated patterns under various text prompts and random seeds showed consistent style, pattern color, and background color with the textual descriptions. Benefiting from the diverse pattern types and richly annotated descriptions in the constructed dataset, the generated results also exhibited strong artistic creativity. Quantitatively, although the introduction of attention excitation caused slight perturbations to noise prediction, resulting in a minor decrease in the CLIP Score metric, the T2T-Sim and C2T-Acc indicators were improved by 2.96% and 8.94%, respectively, compared to the more advanced Structured Diffusion and Attend-and-Excite models. These results indicate that the proposed method not only significantly enhances the color semantic consistency of the patterns but also demonstrates substantial improvements in overall pattern quality and visual fidelity. Furthermore, ablation studies show that both the target color excitation and background color excitation contribute positively to the overall improvement, leading to better color generation, detailed pattern formation, and accurate background rendering.

Conclusion This study investigates cross-modal generation of textile digital printing patterns from textual descriptions. The powerful generative model SD-V1.5 is adopted as the baseline and fine-tuned on a custom-built dataset of textile printing patterns. In order to enhance the consistency between generated pattern colors and color semantics in the text, a cross-attention-based color excitation method is introduced, which increases the attention weights of color-related keywords during generation. Extensive qualitative and quantitative evaluations confirm the effectiveness of the proposed method. With only a textual input, users can efficiently obtain semantically aligned and high-quality textile digital printing patterns, enabling convenient, personalized customization. By leveraging advanced AIGC technologies, the proposed approach establishes a domain-specific generation model for textile digital printing, contributing to the development of new productive forces and facilitating the digital, intelligent, and sustainable transformation of the next-generation textile printing industry.

Key words: digital printing pattern, cross-modal generation, color-semantic consistency, diffusion model, personalized pattern customization, artificial intelligence generated content

CLC Number: 

  • TS101

Fig.1

Partial text-pattern pairs in textile digital printing dataset"

Fig.2

Cross attention color excitation framework"

Fig.3

Visualization results of digital printing patterns generated by proposed method"

Fig.4

Visual comparison of proposed method with other methods under the same seed conditions"

Tab.1

Quantitative comparison results of different methods"

方法 CLIP
分数
T2T-
Sim
C2T-
Acc
TexTile
未微调原始
SD-V1.5
0.245 8 0.620 9 0.737 2 0.539 8
微调
SD-V1.5
0.246 5 0.635 3 0.774 2 0.605 8
Attend-and-
Excite
0.245 6 0.625 8 0.777 8 0.616 7
Structured
Diffusion
0.247 9 0.638 1 0.648 7 0.592 6
本文方法 0.245 6 0.657 0 0.847 3 0.618 5

Tab.2

Ablation study on cross-attention color excitation"

目标对象
色彩激励
背景
色彩激励
CLIP分数 T2T-Sim C2T-Acc
× × 0.246 5 0.635 3 0.774 2
× 0.246 2 0.644 9 0.825 6
× 0.245 8 0.641 5 0.795 7
0.245 6 0.657 0 0.847 3

Fig.5

Comparison of ablation study on target object color exciting and background color exciting. (a) Without color excitation; (b) Only with target-object color excitation; (c) Without background color excitation; (d) With both target-object and background color excitation"

Fig.6

Trend of C2T-Acc indicator with respect to scale factor α"

[1] 丁思佳, 林琳, 董淑秀, 等. 2024年中国纺织品数码喷墨印花发展报告[J]. 染整技术, 2024, 46(12): 6-18.
DING Sijia, LIN Lin, DONG Shuxiu, et al. 2024 China textile digital inkjet printing development report[J]. Textile Dyeing and Finishing Journal, 2024, 46(12): 6-18.
[2] 关芳兰, 王建明, 陈嘉滢, 等. 二醋酯纤维织物分散染料免蒸洗数码印花工艺[J]. 纺织学报, 2024, 45(12): 137-143.
doi: 10.13475/j.fzxb
GUAN Fanglan, WANG Jianming, CHEN Jiaying, et al. Non-steaming digital printing process of diacetate fabric with disperse dyes[J]. Journal of Textile Research, 2024, 45(12): 137-143.
doi: 10.13475/j.fzxb
[3] 陈家辉, 梁跃耀, 陈妮, 等. 棉织物喷墨印花打印方式的调控及其应用[J]. 纺织学报, 2023, 44(7): 159-166.
CHEN Jiahui, LIANG Yueyao, CHEN Ni, et al. Research and application of ink jet printing on cotton fabrics[J]. Journal of Textile Research, 2023, 44(7): 159-166.
[4] 郑畑子, 王建萍. 服装印花图案设计的感性研究[J]. 纺织学报, 2020, 41(8): 101-107.
ZHENG Tianzi, WANG Jianping. Perceptual research on printing pattern design for clothing[J]. Journal of Textile Research, 2020, 41(8): 101-107.
doi: 10.1177/004051757104100203
[5] 郑小虎, 刘正好, 陈峰, 等. 纺织工业智能发展现状与展望[J]. 纺织学报, 2023, 44(8): 205-216.
ZHENG Xiaohu, LIU Zhenghao, CHEN Feng, et al. Current status and prospect of intelligent development in textile industry[J]. Journal of Textile Research, 2023, 44(8): 205-216.
[6] 冉二飞, 贾小军, 喻擎苍, 等. 基于SE注意力CycleGAN的蓝印花布单纹样自动生成[J]. 丝绸, 2024, 61(1): 31-37.
RAN Erfei, JIA Xiaojun, YU Qingcang, et al. Single pattern automatic generation of blue calico based on SE attention CycleGAN[J]. Journal of Silk, 2024, 61(1): 31-37.
[7] SU Z B, ZHAO S Y, ZHANG H H, et al. Digital printing image generation method based on style transfer[J]. Textile Research Journal, 2023, 93(23/24): 5211-5223.
doi: 10.1177/00405175231195367
[8] 张佳伟, 李华军, 王秀丽, 等. 基于扩散模型的印花图案生成方法设计[J]. 计算机测量与控制, 2024, 32(10): 243-249.
ZHANG Jiawei, LI Huajun, WANG Xiuli, et al. Design of printed pattern generation method based on diffusion models[J]. Computer Measurement & Control, 2024, 32(10): 243-249.
[9] 王罕仁, 张华熊. 基于生成对抗网络与稳定扩散模型的花卉丝巾图案生成方法[J]. 浙江理工大学学报(自然科学), 2025, 50(4): 556-570.
WANG Hanren, ZHANG Huaxiong. A generative method for floral scarf patterns using GANs and stable diffusion models[J]. Journal of Zhejiang Sci-Tech University (Natural Sciences), 2025, 50(4): 556-570.
[10] CHEFER H, ALALUF Y, VINKER Y, et al. Attend-and-excite: attention-based semantic guidance for text-to-image diffusion models[J]. ACM Transactions on Graphics, 2023, 42(4): 1-10.
[11] FENG W, HE X, FU T J, et al. Training-free structured diffusion guidance for compositional text-to-image synthesis[C]// The Eleventh International Conference on Learning Representations(ICLR). Kigali: ICLR, 2023:3-12.
[12] AHMAD HASSAN YAR G N, TAHA M, AFZAL Z, et al. TexGAN: textile pattern generation using deep convolutional generative adversarial net-work(DCGAN)[C]// 2023 IEEE International Conference on Emerging Trends in Engineering, Sciences and Technology (ICES&T). New York: IEEE, 2023: 1-6.
[13] KARAGOZ H F, BAYKAL G, EKSI I A, et al. Textile pattern generation using diffusion models[C]// 2023 International Textile & Fashion Congress. Istanbul: ITFC, 2023: 48-53.
[14] HO J, JAIN A, ABBEEL P. Denoising diffusion probabilistic models[J]. Advances in Neural Information Processing Systems, 2020, 33: 6840-6851.
[15] ROMBACH R, BLATTMANN A, LORENZ D, et al. High-resolution image synthesis with latent diffusion models[C]// 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). New York: IEEE, 2022: 10674-10685.
[16] HU E J, SHEN Y, WALLIS P, et al. Lora: low-rank adaptation of large language models[J]. International Conference on Learning Representations, 2022, 1(2): 3.
[17] RADFORD A, KIM J W, HALLACY C, et al. Learning transferable visual models from natural language supervision[C]// International Conference on Machine Learning, 2021: 8748-8763.
[18] LI Q, FU X M, WANG X, et al. Unveiling structural memorization: structural membership inference attack for text-to-image diffusion models[C]// Proceedings of the 32nd ACM International Conference on Multimedia. New York: ACM, 2024: 10554-10562.
[19] RODRIGUEZ-PARDO C, CASAS D, GARCES E, et al. Textile:a differentiable metric for texture tile-ability[C]// 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). New York: IEEE, 2024: 4439-4449.
[20] XIE J H, LI Y X, HUANG Y W, et al. BoxDiff:text-to-image synthesis with training-free box-constrained diffusion[C]// 2023 IEEE/CVF International Conference on Computer Vision (ICCV). New York: IEEE, 2023: 7418-7427.
[1] ZHAO Zewen, LÜ Kuan, SU Xuzhong, SUN Fengxin. Macro-mesoscopic coupled analysis study and numerical simulation of mechanical behavior for staple yarns [J]. Journal of Textile Research, 2026, 47(1): 106-114.
[2] SHAO Jianbo, YUE Xinyan, CHEN Yu, HAN Xiao, HONG Jianhan. Construction and sensing performance of all knitted multi-modal flexible capacitive sensor [J]. Journal of Textile Research, 2026, 47(1): 123-131.
[3] WANG Jiaxi, YUAN Guoshu, CHEN Xiaoyu, CHEN Shuangting, MIAO Ying, FU Chiyu, TANG Wenyang. Research status and development trends of fiber-based colorimetric gas sensors [J]. Journal of Textile Research, 2026, 47(1): 259-267.
[4] ZHANG Ziyue, JIANG Hongxia, LIU Jihong. Design and evaluation of digital texture color cards for monochromatic woven fabrics [J]. Journal of Textile Research, 2025, 46(10): 167-175.
[5] WU Jinyao, ZHONG Yi, ZHANG Linping, XU Hong, MAO Zhiping. Preparation and properties of flexible infrared stealth films based on polyaniline [J]. Journal of Textile Research, 2025, 46(09): 1-8.
[6] JI Hui, XIAO Hong. Structure and electromagnetic response properties of glass-coated magnetic amorphous alloy fibers [J]. Journal of Textile Research, 2025, 46(09): 9-18.
[7] ZHANG Zhaodong, WANG Lei, PAN Ruru. Preparation of quasi-zero stiffness fabric isolator based on ABAQUS simulation [J]. Journal of Textile Research, 2025, 46(09): 163-170.
[8] ZENG Yao, LÜ Jinfeng, WANG Jieping, LIU Rongpeng, ZHOU Chan. Progress in application of three-dimensional silk protein scaffolds [J]. Journal of Textile Research, 2025, 46(09): 258-267.
[9] ZHANG Xiaoting, ZHAO Pengyu, PAN Ruru, GAO Weidong. Plaid fabric image retrieval method based on deep feature fusion [J]. Journal of Textile Research, 2025, 46(08): 89-95.
[10] QUAN Ying, ZHANG Aiqin, ZHANG Man, LIU Shuqiang, ZHANG Yujing. Fabrication and characterization of wearable flexible strain sensors based on three-dimensional braided structures [J]. Journal of Textile Research, 2025, 46(08): 136-144.
[11] WANG Zihan, LI Yong, CHEN Xiaochuan, WANG Jun, LIANG Lingjie. Modeling and simulation of waste cotton fabric shredding process [J]. Journal of Textile Research, 2025, 46(07): 136-143.
[12] LI Duo, XIE Xiaowen, ZHANG Difan, WU Jingxia, LU Kai, CHEN Peining. Construction and electromagnetic properties of Wi-Fi dual-band fabric antenna [J]. Journal of Textile Research, 2025, 46(05): 10-16.
[13] LI Pengfei, LUO Yixin, ZHANG Zifan, LU Ning, CHEN Biling, XU Jianmei. Preparation and degradation performance of silk fibroin/chitosan/gelatin embolic microspheres [J]. Journal of Textile Research, 2025, 46(05): 116-124.
[14] GU Mengshang, ZHANG Ning, PAN Ruru, GAO Weidong. Object detection of weaving fabric defects using frequency-domain convolution modules [J]. Journal of Textile Research, 2025, 46(05): 159-168.
[15] LIANG Qimin, YAN Zhuojun, LI Changxin, LIU Zhifeng, HE Sisi. Research progress and prospects in fibers and fabric-based electrochemical sensing and aqueous batteries for smart textiles [J]. Journal of Textile Research, 2025, 46(05): 89-95.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!