摘要(英) |
Amidst the rhythm of rapid and large-scale production in the era of Industry 4.0, effective and precise inspection is pivotal to boost process quality. Traditionally, defect detection hinged on manual visual inspection, which proved not only time-consuming but also vulnerable to fluctuating misjudgment rates due to the inspector′s mental condition. Moreover, with an escalating complexity of metal components, the misjudgment rates of human visual inspection have an increasing tendency, leading manufacturers to adopt automated optical inspection systems in lieu of conventional manual labor. However, present automated optical inspection systems, primarily driven by image recognition, often incur high misjudgment rates in the pursuit of near-perfect detection rates, resulting in a majority of machine-selected candidate images being false defects.
This study attempts to apply convolutional neural network techniques to enhance the surface defect detection of metal components, aiming to alleviate the high misjudgment rates caused by automated optical systems. It aspires to adopt the revolutionary outcomes from the field of image classification driven by convolutional neural networks into the domain of detection technology. The Faster R-CNN model serves as the main architecture for this research. The overall Mean Average Precision (mAP) of the original model is 89.720. By employing the COCO (ResNet-50) pre-trained model and subsequent transfer learning to share the well-trained model and its parameters to a new model, the aim is to optimize the Faster R-CNN algorithm. Consequently, the overall mAP of the optimized model achieves 94.223, marking an approximate improvement of 5%. |
參考文獻 |
[1] V. Riffo and D. Mery, “Automated Detection of Threat Objects Using Adapted Implicit Shape Model, ” IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2015, Vol. 46, No. 4, pp. 472-482.
[2] X. Li, C. Liu, S. Dai, H. Lian and G. Ding, “Scale Specified Single Shot Multibox Detector, ” IET Computer Vision, 2020, Vol. 14, No. 2, pp. 59-64.
[3] D. T. Nguyen, T. N. Nguyen, H. Kim and H. Lee, “A High-Throughput and Power-Efficient FPGA Implementation of YOLO CNN for Object Detection, ” IEEE Transactions on Very Large Scale Integration (VLSI) Systems, 2019, Vol. 27, No. 8, pp. 1861-1873.
[4] K. Han, M. Sun, X. Zhou, G. Zhang, H. Dang and Z. Liu, "A new method in wheel hub surface defect detection: Object detection algorithm based on deep learning," 2017 International Conference on Advanced Mechatronic Systems (ICAMechS), Xiamen, China, 2017, pp. 335-338, doi: 10.1109/ICAMechS.2017.8316494.
[5] S. Ren, K. He, R. Girshick and J. Sun, "Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks," in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, pp. 1137-1149, 1 June 2017, doi: 10.1109/TPAMI.2016.2577031.
[6] 北美智權報第285期 ,深度學習神經網路之運作 ,芮嘉瑋 ,2021年5月26日
[7] Y. Lecun, L. Bottou, Y. Bengio and P. Haffner, "Gradient-based learning applied to document recognition," in Proceedings of the IEEE, vol. 86, no. 11, pp. 2278-2324, Nov. 1998, doi: 10.1109/5.726791.
[8] Krizhevsky, A., I. Sutskever and G. E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 1097-1105.
[9] K. Simonyan, and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in CoRR, abs/1409.1556, 2014.
[10] He, K., X. Zhang, S. Ren and J. Sun. 2016. Deep residual learning for image recognition. In "2016 IEEE Conference on Computer.
[11] R. Girshick, J. Donahue, T. Darrell and J. Malik, "Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation," 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 2014, pp. 580-587, doi: 10.1109/CVPR.2014.81.
[12] R. Girshick, "Fast R-CNN," 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 2015, pp. 1440-1448, doi: 10.1109/ICCV.2015.169. |