Journal of Textile Research ›› 2026, Vol. 47 ›› Issue (02): 73-83.doi: 10.13475/j.fzxb.20250800301

• Fiber Materials • Previous Articles     Next Articles

Classification of down images based on tri-stage residual dynamic focusing network

LÜ Zebin1, LI Ziyin1(), WANG Xiaodong2, YE Fei2, LIU Weihong2   

  1. 1 College of Optical and Electronics Technology, China Jiliang University, Hangzhou, Zhejiang 310018, China
    2 Huzhou Institute of Quality and Technical Supervision and Inspection(Huzhou Fiber Quality Monitoring Center), Huzhou, Zhejiang 313099, China
  • Received:2025-08-04 Revised:2025-11-06 Online:2026-02-15 Published:2026-04-24
  • Contact: LI Ziyin E-mail:liziyin@cjlu.edu.cn

Abstract:

Objective In down quality evaluation systems, key indicators such as color and freshness are decisive factors for product grading and market value. For fine-grained quality classification tasks involving down characteristics like color and freshness, traditional manual inspection methods exhibit low efficiency and strong subjectivity, while existing computational approaches demonstrate inadequate capability in recognizing the intricate texture patterns of down materials. To battle these challenges, this study proposes a Tri-phase residual dynamic focus network (Tri-RDFNet) to enhance the classification accuracy of down images, thereby advancing automated quality assessment in the down industry.

Method Building upon an enhanced ResNet architecture, a novel three-stage residual dynamic focus network was developed for fine-grained classification of down feather images. The network incorporated dilated convolution modules, deformable spatial-convolutional block attention module (DS-CBAM), and dynamic gap-aware attention loss (DGALoss) function to enable in-depth feature learning of down images. Furthermore, a three-stage cascaded training strategy was introduced to significantly improve the model's generalization capability. The experimental dataset, collected using industrial camera systems with a bar-shaped white light source, comprised four categories, i.e. white fresh down, white recycled down, colored fresh down, and colored recycled down. Comprehensive experiments were carried out using the established model.

Result Comparative experimental results demonstrated that traditional CNN models, such as AlexNet and VGG, exhibited certain performance bottlenecks in this task, achieving accuracy rates of only 89.32% and 90.26%, respectively. These models struggled to capture the fine-grained differences inherent in down feather images. Although Transformer-based models possess strong global modeling capabilities, they suffer from overfitting due to the limited dataset size and architectural complexity. The backbone model RDFNet enhanced the learning focus on down image features by incorporating atrous convolution modules, DS-CBAM, and DGALoss. As a result, it achieved a classification accuracy of 95.28% on the collected down image dataset, representing an improvement of 1.11%-5.96% compared to traditional models such as AlexNet, VGG, ResNet, ViT, and Swin Transformer. Furthermore, based on this RDFNet backbone, a three-stage cascaded training strategy was introduced. In the first stage, the model was trained globally using the cross-entropy loss function. The second stage employed DGALoss to reweight and train on easily confused samples, yielding a 1.02% increase in accuracy over the first stage. In the third stage, noise samples were filtered, and sample weights were reassigned to further train the retained samples. This final phase enhanced the model accuracy by an additional 0.71%, achieving a final accuracy of 97.01%. This three-stage process reduced the risk of overfitting while improving precision and generalization. Ablation studies confirmed the effectiveness of each component. The atrous convolution module improved the model's ability to perceive multi-scale features in down images, raising the validation accuracy by 0.99%. The DS-CBAM module enhanced the model's feature selection capability by integrating channel attention with deformable spatial convolution, leading to further improvement in accuracy while introducing minor overfitting. When combined, DS-CBAM and atrous convolution boosted accuracy to 95.28%. Introducing the three-stage training scheme and applying FocalLoss during the second stage to focus on hard examples increased accuracy to 96.11%, thus improving model robustness and stability. Replacing FocalLoss with DGALoss for better focus on confusing samples led to the highest validation accuracy of 97.01%, demonstrating DGALoss's superior capability in distinguishing ambiguous down categories.

Conclusion To address the challenge of fine-grained classification in down feather images, this paper proposes an innovative three-stage residual dynamic focusing network. The core backbone model, RDFNet, enhances feature extraction capabilities by improving the ResNet architecture through the integration of DS-CBAM and atrous convolution modules. Based on RDFNet, a three-stage training strategy is designed, consisting of warm-up training, adaptive weighted training using the novel DGALoss function, and refined sample training, collectively forming the Tri-RDFNet model. This approach effectively improves the recognition of easily confused down feather image samples and enhances the model's generalization ability. Experimental results demonstrate that the proposed method achieves a classification accuracy of 97.01% on a self-constructed dataset of 8 000 down feather images, significantly outperforming traditional methods. This provides an efficient solution for automated down quality assessment and offers a valuable reference for fine-grained image classification tasks.

Key words: down, down image classification, dynamic focusing network, attention mechanism, DGALoss, Tri-stage cascaded training, residual network

CLC Number: 

  • TS107.2

Fig.1

Down image classification and detection process"

Fig.2

Image acquisition device"

Fig.3

Comparison of four types of down. (a)White fresh down; (b)White recycled down; (c)Colored fresh down; (d)Colored recycled down"

Fig.4

RDFNet network structure diagram"

Fig.5

First-layer residual block structure"

Fig.6

DS-CBAM attention mechanism"

Fig.7

Second-layer residual block structure"

Fig.8

Dilated convolution structure"

Fig.9

Third-layer residual block structure"

Fig.10

Tri-RDFNet architecture"

Fig.11

Loss variation curves of a sample during first two training stages. (a) Stage 1; (b) Stage 2"

Tab.1

Model comparison experiments"

模型 准确率/% 参量M
训练集 验证集
AlexNet 90.38 89.32 61
VGG19 92.26 90.26 143
ResNet18 92.09 91.78 11.7
ResNet24 93.16 92.06 12.36
ResNet50 93.67 92.56 25.6
Vision Transformer 95.81 93.77 ≥86
Swin Transformer 95.38 94.17 ≥29
RDFNet 95.69 95.28 23.3

Tab.2

Model ablation experiments"

模型结构组合 准确率/%
训练集 验证集
ResNet24 93.16 92.06
ResNet24+空洞卷积 93.76 93.05
ResNet24+CBAM 94.33 93.09
ResNet24+DS-CBAM 95.09 93.56
ResNet24+空洞卷积+DS-CBAM (RDFNet) 95.69 95.28
Tri-RDFNet+FocalLoss 96.55 96.11
Tri-RDFNet+DGALoss 97.11 97.01

Fig.12

Loss function and accuracy variation curves. (a)Loss function change curves; (b)Accuracy change curve"

Fig.13

Three-stage confusion matrix. (a) Stage 1; (b) Stage 2; (c) Stage 3"

[1] 魏军望. 梭织羽绒服填充材料的保暖性能与可持续发展研究[J]. 西部皮革, 2025, 47(4): 12-14.
WEI Junwang. Research on the thermal performance and sustainable development of woven down jacket filling material[J]. West Leather, 2025, 47(4): 12-14.
[2] 宋庆强. 基于机器视觉的异色羽绒在线分拣硬件系统研制[D]. 上海: 东华大学, 2016: 1-54.
SONG Qingqiang. The development of colored eiderdown online sorting hardware system based on machine vision[D]. Shanghai: Donghua University, 2016: 1-54.
[3] KIM S, KIM E, PARK Y. Thermal insulation and morphology of natural and synthetic filled outdoor sportswear by repeated water washing and dry cleaning[J]. International Journal of Clothing Science and Technology, 2018, 30(3): 428-443.
doi: 10.1108/IJCST-09-2017-0149
[4] OHTA N, IWAMOTO R, AIBA S, et al. Microscopical identification of goose and downs[J]. Textile Science, 2004, 9(2):95-102.
[5] 单婷婷. 基于统计特征值对异色羽绒的识别[D]. 上海: 东华大学, 2016: 1-50.
SHAN Tingting. Detection of color fluff based on statistical characteristic values[D]. Shanghai: Donghua University, 2016: 1-50.
[6] 孙近, 赵瑞方, 宋晨, 等. 近红外光谱法在羽绒羽毛种类鉴别的研究应用[J]. 中国纤检, 2024(12): 50-53.
SUN Jin, ZHAO Ruifang, SONG Chen, et al. Application research on the identification of down feather by near infrared spectroscopy[J]. China Fiber Inspection, 2024(12): 50-53.
[7] 张克和, 陈宇, 俞旭霞, 等. 鹅鸭羽毛羽绒结构特征分析[J]. 中国纤检, 2011(5): 50-52.
ZHANG Kehe, CHEN Yu, YU Xuxia, et al. Study on structrual characterization of down and feather of the duck and goose[J]. China Fiber Inspection, 2011(5): 50-52.
[8] 王艳秋. 羽绒图像阈值分割算法研究[J]. 计算机工程与应用, 2008, 44(34): 246-248.
WANG Yanqiu. Research of image segmentation algorithm of feather and down[J]. Computer Engineering and Applications, 2008, 44(34): 246-248.
doi: 10.3778/j.issn.1002-8331.2008.34.074
[9] 葛洪伟, 杨小艳. 组合内核与优化算法在羽绒识别系统中的应用[J]. 数据采集与处理, 2008, 23(2): 219-223.
GE Hongwei, YANG Xiaoyan. Application of hybrid kernel and optimization algorithm in down category recognition system[J]. Journal of Data Acquisition & Processing, 2008, 23(2): 219-223.
[10] 邢笛, 葛洪伟, 李志伟. 模糊支持张量机图像分类算法及其应用[J]. 计算机应用, 2012, 32(8): 2227-2229, 2234.
XING Di, GE Hongwei, LI Zhiwei. Research and application of fuzzy tensor machine image classification algorithm[J]. Journal of Computer Applications, 2012, 32(8): 2227-2229, 2234.
doi: 10.3724/SP.J.1087.2012.02227
[11] 杨文柱, 刘晴, 王思乐, 等. 基于深度卷积神经网络的羽绒图像识别[J]. 郑州大学学报(工学版), 2018, 39(2): 11-17.
YANG Wenzhu, LIU Qing, WANG Sile, et al. Down image recognition based on deep convolution neural networks[J]. Journal of Zhengzhou Univer-sity (Engineering Science), 2018, 39(2): 11-17.
[12] 郭文, 黄中浩, 龙俊超, 等. 智能识别技术在羽毛羽绒种类鉴别的应用研究[J]. 中国纤检, 2024(8): 56-59.
GUO Wen, HUANG Zhonghao, LONG Junchao, et al. Research on intelligent recognition for identification of feather and down species[J]. China Fiber Inspection, 2024(8): 56-59.
[13] HE K M, ZHANG X Y, REN S Q, et al. Deep residual learning for image recognition[C]// 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). New York: IEEE, 2016: 770-778.
[14] WOO S, PARK J, LEE J Y, et al. CBAM: convolutional block attention module[C]// Computer Vision - ECCV 2018. Cham: Springer, 2018: 3-19.
[15] YU F, KOLTUN V. Multi-scale context aggregation by dilated convolutions[EB/OL]. 2015: arXiv: 1511.07122. https://arxiv.org/abs/1511.07122.
[16] LIN T Y, GOYAL P, GIRSHICK R, et al. Focal loss for dense object detection[C]// 2017 IEEE International Conference on Computer Vision (ICCV). New York: IEEE, 2017: 2999-3007.
[1] FENG Zhengrong, LIU Beifen, CHEN Mengyuan. Style transfer model for floral printed patterns based on multi-scale feature fusion [J]. Journal of Textile Research, 2026, 47(02): 264-272.
[2] HU Xinyang, WANG Hongzhi. Preparation of poly(vinylidene fluoride-trifluoride-trifluoroethylene)copolymer-based triboeletric nanogenerator and enhancement of its output power [J]. Journal of Textile Research, 2025, 46(12): 92-100.
[3] ZHOU Xiaoye, CAO Ailing, SU Rina, SHI Wenjun, ZHAO Xianmei, DU Dasheng, CAI Luyun. Rapid detection method for unwashed and inadequately washed feather and down [J]. Journal of Textile Research, 2025, 46(10): 46-53.
[4] ZHANG Xiaoting, ZHAO Pengyu, PAN Ruru, GAO Weidong. Plaid fabric image retrieval method based on deep feature fusion [J]. Journal of Textile Research, 2025, 46(08): 89-95.
[5] ZHANG Shasha, CAI Muhang, LÜ Xiaojing, HU Dan, LIU Juan, JI Xingzhao, CAO Genyang, WANG Haona. Electrostatic field synergistic conformation of down/silicon dioxide aerogel warmth keeping materials [J]. Journal of Textile Research, 2025, 46(07): 10-18.
[6] LUO Ruiqi, CHANG Dashun, HU Xinrong, LIANG Jinxing, PENG Tao, CHEN Jia, LI Li. Cross-pose virtual try-on based on improved appearance flow network [J]. Journal of Textile Research, 2025, 46(06): 203-211.
[7] LU Yinwen, HOU Jue, YANG Yang, GU Bingfei, ZHANG Hongwei, LIU Zheng. Single dress image video synthesis based on pose embedding and multi-scale attention [J]. Journal of Textile Research, 2024, 45(07): 165-172.
[8] GU Meihua, HUA Wei, DONG Xiaoxiao, ZHANG Xiaodan. Occlusive clothing image segmentation based on context extraction and attention fusion [J]. Journal of Textile Research, 2024, 45(05): 155-164.
[9] HU Xudong, TANG Wei, ZENG Zhifa, RU Xin, PENG Laihu, LI Jianqiang, WANG Boping. Structure classification of weft-knitted fabric based on lightweight convolutional neural network [J]. Journal of Textile Research, 2024, 45(05): 60-69.
[10] LU Weijian, TU Jiajia, WANG Junru, HAN Sijie, SHI Weimin. Model for empty bobbin recognition based on improved residual network [J]. Journal of Textile Research, 2024, 45(01): 194-202.
[11] SHI Hongyu, WEI Yingjie, GUAN Shengqi, LI Yi. Cotton foreign fibers detection algorithm based on residual structure [J]. Journal of Textile Research, 2023, 44(12): 35-42.
[12] HUANG Yueyue, CHEN Xiao, WANG Haiyan, YAO Haiyang. Human clothing color recognition based on ClothResNet model [J]. Journal of Textile Research, 2023, 44(10): 143-148.
[13] BU Fan, YING Lili, LI Changlong, WANG Zongqian. Dissolution behavior and mechanism of down in lactic acid/cysteine deep eutectic solvent [J]. Journal of Textile Research, 2023, 44(10): 24-30.
[14] MA Chuangjia, QI Lizhe, GAO Xiaofei, WANG Ziheng, SUN Yunquan. Stitch quality detection method based on improved YOLOv4-Tiny [J]. Journal of Textile Research, 2023, 44(08): 181-188.
[15] FU Han, HU Feng, GONG Jie, YU Lianqing. Defect reconstruction algorithm for fabric defect detection [J]. Journal of Textile Research, 2023, 44(07): 103-109.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!