PDF (12.3 MB)
Collect
Submit Manuscript
Show Outline
Outline
Abstract
Keywords
References
Show full outline
Hide outline

Improved YOLOv5s Algorithm for Small Target Engineering Vehicles Detection with Fixed-Point Cameras

Shiliang XU1Minquan LAI2Yu LEI1Jizhong LIU2()
Jiangxi Natural Resources Development Center, Nanchang Jiangxi 330025, China
School of Advanced Manufactoring, Nanchang University, Nanchang Jiangxi 330031, China
Show Author Information

Abstract

In order to effectively reduce and prevent the damage to the natural environment caused by illegal land reclamation and mineral excavation, an ETS-YOLO small target monitoring and recognition algorithm for the detection of various types of engineering vehicles in complex environments is proposed using cameras deployed to high towers. Firstly, the EfficientViT network is used to replace the backbone feature extraction network of YOLOv5s in order to improve the attention diversity and significantly reduce the number of model parameters. Secondly, a small target detection layer is added to enhance the network’s extraction of shallow semantic information to improve the performance of small target detection. Finally, the original NMS function is replaced with the soft non-maximal suppression algorithm (soft-NMS) to effectively recognize occluded and overlapped targets. The experimental results show that the improved model has a mean average precision (mAP) of 93.3%, a parameter count of 5.90 M, and a detection speed of 52 f/s. Compared with the YOLOv5s model, the mAP is improved by 2.6% and the parameter count is decreased by 16.1%.

CLC number: TP391 Document code: A Article ID: 2096-7675(2025)01-0099-08

References

[1]
ZHOU S. Study on the implementation dilemma of the normal management of illegal building in Shanghai: A case study of P district[D]. Shanghai: East China Normal University, 2022. (in Chinese)
[2]
REN S Q, HE K M, GIRSHICK R, et al. Faster R-CNN: Towards real-time object detection with region proposal networks[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 39(6): 1137-1149.
[3]
BOCHKOVSKIY A, WANG C Y, LIAO H Y M. YOLOv4: Optimal speed and accuracy of object detection[EB/OL]. (2020-04-23)[2023-05-20]. https://doi.org/10.48550/arXiv.2004.10934.
[4]
GE Z, LIU S T, WANG F, et al. YOLOX: Exceeding YOLO series in 2021[EB/OL]. (2021-08-06)[2023-05-20]. https://doi.org/10.48550/arXiv.2107.08430.
[5]
LI C Y, LI L L, JIANG H L, et al. YOLOv6: A single-stage object detection framework for industrial applications[EB/OL]. (2022-09-07)[2023-06-25]. https://doi.org/10.48550/arXiv.2209.02976.
[6]
XU S L, WANG X X, LYU W Y, et al. PP-YOLOE: An evolved version of YOLO[EB/OL]. (2022-03-30)[2023-06-01]. https://doi.org/10.48550/arXiv.2203.16250.
[7]
LIU W, ANGUELOV D, ERHAN D, et al. SSD: single shot multibox detector[C]//European Conference on Computer Vision. Cham: Springer, 2016: 21-37.
[8]
GAO Y S, CHEN W Z, WANG J, et al. Detection of transmission lines construction vehicles for UAV based on Android[J]. Computer Systems & Applications, 2020, 29(2): 257-261. (in Chinese)
[9]
GUO Y P, XU Y, LI S L. Dense construction vehicle detection based on orientation-aware feature fusion convolutional neural network[J]. Automation in Construction, 2020, 112: 103124.
[10]
LI Z X, YANG L, ZHOU F Q. FSSD: Feature fusion single shot multibox detector[EB/OL]. (2017-12-04)[2023-06-01]. https://doi.org/10.48550/arXiv.1712.00960.
[11]
GU C L, YANG H, LIU Y B, et al. Object detection of engineering vehicles based on self-adaptive local exclusion loss and normalized area loss[J]. China Production Safety Science and Technology, 2021, 17(11): 40-47. (in Chinese)
[12]
YAN W J. Research of key technologies of 3D object recognition on construction sites[D]. Chengdu: University of Electronic Science and Technology of China, 2022. (in Chinese)
[13]
LI J X, LIU Z Y, MA Y, et al. Design and application of improved YOLOv5 in the quality control system of Amblyseius cucumeris Oudemans[J]. Journal of Nanchang University(Engineering & Technology), 2023, 45(1): 85-95. (in Chinese)
[14]
PU J P, WANG X M, GAO H W. Research of aerial image object detection method based on improved YOLOv4[J]. Journal of Shenyang Ligong University, 2023, 42(3): 46-53. (in Chinese)
[15]
ZHENG Z H, WANG P, LIU W, et al. Distance-IoU loss: Faster and better learning for bounding box regression[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 2020, 34(7): 12993-13000.
[16]
HOWARD A, SANDLER M, CHEN B, et al. Searching for MobileNetV3[C]//2019 IEEE/CVF International Conference on Computer Vision(ICCV). Seoul, Korea(South). IEEE, 2019: 1314-1324.
[17]
KATHAROPOULOS A, VYAS A, PAPPAS N, et al. Transformers are RNNs: Fast autoregressive transformers with linear attention[EB/OL]. (2020-08-31)[2023-10-01]. https://doi.org/10.48550/arXiv.2006.16236.
[18]
XU Y, XIONG J J, LI L, et al. Detection of pepper clusters using improved YOLOv5s[J]. Transactions of the Chinese Society of Agricultural Engineering, 2023, 39(16): 283-290. (in Chinese)
[19]
BODLA N, SINGH B, CHELLAPPA R, et al. Soft-NMS-improving object detection with one line of code[C]//2017 IEEE International Conference on Computer Vision(ICCV). Venice, Italy. IEEE, 2017: 5562-5570.
[20]
HAN W X, ALIFU K, HUANG Z T. Small-object fast detection in optical remote sensing images based on a modified SSD[J]. Journal of Xinjiang University(Natural Science Edition in Chinese and English), 2020, 37(2): 163-169. (in Chinese)
[21]
TAN M X, LE Q V. EfficientNet: Rethinking model scaling for convolutional neural networks[EB/OL]. (2019-05-24)[2023-06-08]. http://arxiv.org/abs/1905.11946v5.
[22]
WANG R, LIU D M, ZHANG J. Wear-YOLO: Research on detection methods of safety equipment for power personnel in substations[J]. Computer Engineering and Applications, 2024, 60(9): 111-121. (in Chinese)
Journal of Xinjiang University(Natural Science Edition in Chinese and English)
Pages 99-106
Cite this article:
XU S, LAI M, LEI Y, et al. Improved YOLOv5s Algorithm for Small Target Engineering Vehicles Detection with Fixed-Point Cameras. Journal of Xinjiang University(Natural Science Edition in Chinese and English), 2025, 42(1): 99-106. https://doi.org/10.13568/j.cnki.651094.651316.2024.04.28.0001
Metrics & Citations  
Article History
Copyright
Return