Defect detection algorithm of strip surface based on STCS-YOLO
ZHOU Yaluo1, WU Xianchao1, LIU Wenguang2, ZHANG Ruicheng1
1. Collage of Electrical Engineering, North China University of Science and Technology, Tangshan 063210, Hebei, China; 2. Shougang Jingtang Limited Iron and Steel Co., Ltd., Tangshan 063200, Hebei, China
Abstract:Strip steel surface defect detection technology is one of the important technologies for the production of high quality strip steel products. Aiming at the problems of misdetection, inaccurate positioning, and poor detection ability of small-scale defect targets in previous strip surface defect detection, a strip surface defect detection algorithm (STCS-YOLO) of improved YOLOv5 was proposed. Firstly, the Swin Transformer module was used to fuse with the original C3 module in the output part of feature fusion network to enhance the interaction and reuse of global feature information, which significantly improved the detection ability of small-scale defect targets. Secondly, a lightweight upsampling operator CARAFE was used to replace the traditional sampling operation to better recover defect information, which improved the recognition accuracy of strip surface defects. Finally, embedding 3-D weight attention mechanism SimAM in feature extraction network to enhance the ability of focusing on the foreground feature information, which improved the strong identification ability of defect target. The experimental results demonstrate that the proposed algorithm achieves 79.7% of PmA(mean average precision) on the NEU-DET dataset, which is improvement of 3.9 percentage points over the original network. Additionally, while maintaining model weight and computational complexity nearly unchanged, the single-frame detection time reaches 10.9 ms, which can basically meet the requirements of accurate and rapid detection of strip surface defects. The strip surface defect detection algorithm proposed in this paper lays the technological foundation for improving neat and flawless high quality strip products.
REN Z,FANG F,YAN N,et al. State of the art in defect detection based on machine vision[J]. International Journal of Precision Engineering and Manufacturing-Green Technology,2022,9(2): 661.
FANG X, LUO Q, ZHOU B,et al. Research progress of automated visual surface defect detection for industrial metal planar materials[J]. Sensors,2020,20(18): 5136.
GIRSHICK R. Fast R-CNN[C]//Proceedings of the IEEE international conference on computer vision. Santiago:IEEE,2015: 1440.
[22]
REN S, HE K, GIRSHICK R, et al. Faster R-CNN: Towards real-time object detection with region proposal networks[J]. Advances in neural information processing systems, 2015, 28.
[23]
HE K, GKIOXARI G, DOLLÁR P, et al. Mask R-CNN[C]//Proceedings of the IEEE international conference on computer vision. Venice: IEEE, 2017: 2961.
[24]
LIU W, ANGUELOV D, ERHAN D, et al. Ssd: Single shot multibox detector[C]//Computer Vision-ECCV 2016: 14th European Conference. Amsterdam, Netherlands:Springer International Publishing, 2016: 21.
[25]
REDMON J, DIVVALA S, GIRSHICK R, et al. You only look once: Unified, real-time object detection[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. Las Vegas:IEEE,2016: 779.
[26]
REDMON J, FARHADI A. YOLO9000: better, faster, stronger[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. Honolulu:IEEE,2017:7263.
[27]
REDMON J,FARHADI A. YOLOv3: An incremental improvement[J]. arXiv preprint arXiv:1804. 02767, 2018.
LIN T Y, DOLLÁR P, GIRSHICK R, et al. Feature pyramid networks for object detection[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. Honolulu:IEEE,2017: 2117.
[32]
LIU S, QI L, QIN H, et al. Path aggregation network for instance segmentation[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. Salt Lake City: IEEE,2018: 8759.
[33]
LIU Z, LIN Y, CAO Y, et al. Swin transformer: Hierarchical vision transformer using shifted windows[C]//Proceedings of the IEEE/CVF international conference on computer vision.Montreal:IEEE,2021: 10012.
[34]
WANG J, CHEN K, XU R, et al. Carafe: Content-aware reassembly of features[C]//Proceedings of the IEEE/CVF international conference on computer vision. Seoul:IEEE,2019: 3007.
[35]
YANG L, ZHANG R Y, LI L, et al. Simam: A simple, parameter-free attention module for convolutional neural networks[C]//International conference on machine learning. Virtual:PMLR, 2021: 11863.