Enhancing Fabric Defect Detection Using Efficient Pyramid Split Attention in a Lightweight YOLOv5 Framework
Article Sidebar
Main Article Content
Fabric defect detection is a fundamental quality control process in textile manufacturing, yet achieving accurate and reliable automated inspection remains difficult because of complex background textures, subtle defect patterns, and substantial variation in defect scale and shape. Although deep learning–based detectors have improved inspection performance, many lightweight models still suffer from limited feature discrimination, particularly in real-time industrial environments where computational efficiency is critical. To address this limitation, this study proposes an enhanced fabric defect detection framework by integrating an Efficient Pyramid Split Attention (EPSA) mechanism into a YOLOv5-based convolutional network. The EPSA module is designed to adaptively recalibrate multi-scale feature responses, enabling the network to emphasize defect-relevant information more effectively while preserving inference efficiency. A quantitative experimental design was employed using a labeled fabric defect image dataset, and the proposed model was evaluated through comparative and ablation analyses against baseline and alternative attention-based configurations. Experimental results indicate that the EPSA-enhanced model achieves superior detection performance in terms of mean Average Precision while maintaining real-time processing capability. The improvement is especially evident for small, low-contrast, and irregular defects embedded in repetitive fabric textures. These findings confirm that pyramid-based attention can substantially improve feature representation without imposing significant computational overhead. The proposed approach offers a practical and efficient solution for automated textile inspection and provides a useful foundation for future research on lightweight attention modeling for industrial vision systems.
Downloads
References
X. Xie, “A review of recent advances in surface defect detection using texture analysis techniques,” Electronic Letters on Computer Vision and Image Analysis, vol. 7, no. 3, pp. 1–22, 2008, doi: 10.5565/rev/elcvia.108.
Y. Liu, K. Zhang, J. Zhang, and Q. Wang, “Automatic fabric defect detection using convolutional neural networks,” Textile Research Journal, vol. 89, no. 23–24, pp. 5147–5160, 2019, doi: 10.1177/0040517519849985.
Kumar, “Computer-vision-based fabric defect detection: A survey,” IEEE Transactions on Industrial Electronics, vol. 55, no. 1, pp. 348–363, Jan. 2008, doi: 10.1109/TIE.2007.896476.
Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, May 2015, doi: 10.1038/nature14539.
J. Redmon and A. Farhadi, “YOLOv3: An incremental improvement,” arXiv preprint arXiv:1804.02767, 2018.
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 2016, pp. 779–788, doi: 10.1109/CVPR.2016.91.
G. Jocher et al., “YOLOv5,” GitHub repository, 2020. [Online]. Available: https://github.com/ultralytics/yolov5
S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “CBAM: Convolutional block attention module,” in Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 2018, pp. 3–19, doi: 10.1007/978-3-030-01234-2_1.
J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 2018, pp. 7132–7141, doi: 10.1109/CVPR.2018.00745.
T.-Y. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and S. Belongie, “Feature pyramid networks for object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 2017, pp. 2117–2125, doi: 10.1109/CVPR.2017.106.
X. Wang, R. Girshick, A. Gupta, and K. He, “Non-local neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 2018, pp. 7794–7803, doi: 10.1109/CVPR.2018.00813.
H. Zhang, C. Wu, Z. Zhang, Y. Zhu, Z. Lin, and Y. Sun, “ResNeSt: Split-attention networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 2020, pp. 2736–2746, doi: 10.1109/CVPR42600.2020.00281.
Z. Zhu, H. Liang, H. Zhang, and R. Zhang, “Efficient pyramid split attention for convolutional neural networks,” Pattern Recognition, vol. 123, Art. no. 108377, 2022, doi: 10.1016/j.patcog.2021.108377.
Szegedy et al., “Going deeper with convolutions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 2015, pp. 1–9, doi: 10.1109/CVPR.2015.7298594.
International Organization for Standardization, Textiles—Quality Control and Inspection Systems, Geneva, Switzerland: ISO, 2015.

This work is licensed under a Creative Commons Attribution 4.0 International License.
All articles published in our journal are licensed under CC-BY 4.0, which permits authors to retain copyright of their work. This license allows for unrestricted use, sharing, and reproduction of the articles, provided that proper credit is given to the original authors and the source.