LwDetFormer: A Hybrid Mobile and Transformer-Based Deep Learning Model for Lockwire Object Detection
DOI:
https://doi.org/10.54691/pn0y9a15Keywords:
Lockwire Object Detection, Small Target Detection, Complex Backgrounds.Abstract
Lockwire plays a crucial role in various industries, including aviation, automotive, power, oil, and gas, by ensuring the secure fastening of critical components. However, detecting Lockwire objects poses significant challenges due to their reflective metallic materials and surface textures that closely resemble the surrounding environment. Traditional object detection methods struggle with these characteristics, especially when working with small datasets and variable lighting conditions. To address these issues, we propose LwDetFormer, a hybrid CNN-Transformer model designed for high-precision detection of Lockwire objects in complex backgrounds. LwDetFormer integrates MobileNet's local feature extraction capabilities with Transformer's global feature modeling abilities, enhancing accuracy and robustness. The model includes innovative modules such as Spatial Pyramid Pooling Focus, Feature Enhancement Module, Feature Fusion Module, and Spatial Context Awareness Module. Experimental results on a test set containing 2,000 images show that LwDetFormer outperforms advanced models like YOLOv8, MobileNetv3, and MobileFormer in terms of precision and recall, achieving a precision rate of 95.9% and a recall rate of 92.0%. These findings highlight LwDetFormer's potential for improving safety and efficiency in the inspection of aircraft engine fuse parts.
Downloads
References
[1] Zhang Fengfei, Sun Junhua. An Instance Segmentation Method of Aviation Engine Fuses Based on Improved Mask R-CNN[J/OL]. Measurement & Control Technology, 1-9 [2025-03-14]. http://kns.cnki.net/kcms/detail/11.5347.TB.20250122.1826.002.html.
[2] Howard A, Sandler M, Chu G, et al. Searching for MobilenetV3[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). 2019: 1314-1324.
[3] Yuan L, Chen Y, Wang T, et al. Tokens-to-token ViT: Training Vision Transformers from Scratch on ImageNet [C]//Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). 2021: 558-567.
[4] Chen Y, Dai X, Chen D, et al. Mobile-former: Bridging mobilenet and transformer[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022: 5270-5279.
[5] Zhao X, Wang L, Zhang Y, et al. A review of convolutional neural networks in computer vision[J]. Artificial Intelligence Review, 2024, 57(4): 99.
[6] Mehta S, Rastegari M. Mobilevit: light-weight, general-purpose, and mobile-friendly vision transformer[J]. arXiv preprint arXiv:2110.02178, 2021.
[7] Guo D, Zhang C, Yang G, et al. Siamese-RCNet: Defect Detection Model for Complex Textured Surfaces with Few Annotations[J]. Electronics, 2024, 13(24): 4873.
[8] Zhang Y, Ye M, Zhu G, et al. FFCA-YOLO for small object detection in remote sensing images[J]. IEEE Transactions on Geoscience and Remote Sensing, 2024, 62: 1-15.
[9] Liu Y, Li H, Hu C, et al. Learning to aggregate multi-scale context for instance segmentation in remote sensing images[J]. IEEE Transactions on Neural Networks and Learning Systems, 2024, 36(1): 595-609.
[10] Liu Jinqiu. Fault Analysis and Repair of a Small Component in an Aero Engine [J]. Value Engineering, 2018, 37(2): 100-102.
[11] Feng Maoxing. An Object Detection Algorithm for Aero-Engine Fasteners Based on Improved YOLOv5s [J]. Science and Technology & Innovation, 2024, (24): 20-22+27. DOI: 10.15913/j.cnki.kjycx.2024.24.006.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Frontiers in Science and Engineering

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.