AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (2 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Research Article | Open Access

BaAM-YOLO: a balanced feature fusion and attention mechanism based vehicle detection network in aerial images

Xunxun Zhang1Xu Zhu2
Xi'an University of Architecture & Technology, Xi'an Shaanxi 710055, China
Chang'an University, Xi'an Shaanxi 710064, China
Show Author Information

Abstract

Vehicle detection in aerial imagery is of paramount importance for various intelligent transportation applications, including vehicle tracking, traffic control, and traffic behavior analysis. However, this task presents significant challenges due to factors such as scale variance, the presence of small vehicles, and complex scenes conditions. The demand for high-quality and precise vehicle detection exacerbates these challenges, particularly in complex environments characterized by low light and occlusion. To address these issues, we propose a balanced feature fusion and attention mechanism-based vehicle detection network, termed BaAM-YOLO, specifically designed for aerial images. Firstly, to improve the vehicle detection accuracy, we didn't directly fuse deep and shallow features. Instead, we focused on achieving a balanced contribution from different feature layers. Secondly, to enhance the feature extraction capabilities of our model, we incorporated a simple parameter-free attention mechanism (SimAM), which aims to improve detection accuracy. Additionally, to effectively model the target confidence loss and classification loss, we emplogyed focal loss to derive a loss function that facilitates dynamic adjustment for multi-class classification. The proposed vehicle detection method was evaluated using the VisDrone2019 dataset. Comparative analyses with state-of-the-art research indicate that the proposed method can obtain superior and competitive results in vehicle detection.

References

[1]

Mao, X. Y.; Yu, B. G.; Shi, Y. J.; et al. Real-Time online trajectory planning and guidance for terminal area energy management of unmanned aerial vehicle[J]. Journal of Industrial & Management Optimization, 2023, 19(3): 1945-1962.

[2]

Abdullah, A.; Constantine, E.; Frederic, B. The detection of clandestine graves in an arid environment using thermal imaging deployed from an unmanned aerial vehicle[J]. Journal of Forensic Sciences, 2023, 68(4): 1286-1291.

[3]

Zhao, X. Y.; Wang, P.; Zhu, S. Y.; et al. Spatial distribution of traffic conflicts in interchange merging area based on video recognition[J]. Journal of Highway and Transportation, Research and Development, 2021, 38(5): 90-99.

[4]

Ma, Y.; Zou, L. P.; Zhang, L. Q. Application technology of uav equipped with optical camera and digital infrared imagery in inspecting quality of pylon and stay cable of sea-crossing bridge[J]. Journal of Highway and Transportation Research and Development, 2018, 35(8): 89-105.

[5]

Li, X. W.; Shi, L. X.; Tang, J. Q.; et al. Determinants of passengers' ticketing channel choice in rail transit systems: New evidence of e-payment behaviors from Xi'an, China[J]. Transport Policy, 2023, 140(1): 30-41.

[6]

Huang, P. C.; Liu, S.; Zou, Z.; et al. Helmet wearing detection method based on improved centernet with enhanced associations[J]. Computer Engineering and Applications, 2023, 59(17): 250-256.

[7]

Niyigena, G.; Lee, S. J.; Kwon, S. K.; et al. Real-time detection and classification of scirtothrips dorsalis on fruit crops with smartphone-based deep learning system: preliminary results[J]. Insects, 2023, 14(6): 523-536.

[8]

Wen, X. P.; Lai, H. C.; Gao, G. X.; et al. Video abnormal behaviour detection based on Pseudo-3D encoder and multi-cascade memory mechanism[J]. IET image processing, 2023, 17(3): 709-721.

[9]
Sun, P.; Wang, W. Y.; Chai, Y. N.; et al. RSN: Range sparse net for efficient, accurate lidar 3D object detection[J]. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021: 5721-5730.
[10]

Zhang Y. J.; Yin, Y.; Shao, Z. Y. An enhanced target detection algorithm for maritime search and rescue based on aerial images[J]. Remote Sensing, 2023, 15(19): 4818-4829.

[11]

Liu, C. G.; Cheng, H. D.; Zhang, Y. T.; et al. Robust multiple cue Fusion-Based High-Speed and nonrigid object tracking algorithm for short track speed skating[J]. Journal of Electronic Imaging, 2016, 25(1): 013014.

[12]
Redmon, J.; Divvala, S.; Girshick, R.; et al. You Only Look Once: Unified, Real-Time object detection[C]. IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas: IEEE Press, 2016: 779-788.
[13]

Yue, B. Y.; Chen, L.; Shi, H.; et al. Ship detection in sar images based on improved retinanet[J]. Journal of Signal Processing, 2022, 38(1): 128-136.

[14]
Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger[C]. IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Hawaii: IEEE Press, 2017: 7263-7271.
[15]
Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement[C]. IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City: IEEE Press, 2018: 1-7.
[16]
Bochkovskiy, A.; Wang, C. Y.; Liao, H. Y. M. YOLOv4: Optimal speed and accuracy of object detection[C]//IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Seattle: IEEE Press, 2020: 1-10.
[17]
Jocher, G.; Chaurasia, A.; Stoken, A.; et al. ultralytics/yolov5: v7. 0-yolov5 sota realtime instance segmentation. Zenodo, 2022zndo. . 7347926J[CP/OL]. DOI:10.5281/zenodo.7347926.2022.
[18]
Li, C. Y.; Li, L. L.; Jiang, H. L.; et al. YOLOv6: A Single-Stage object detection framework for industrial applications[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans: IEEE Press, 2022: 1-17.
[19]
Wang, C. Y.; Bochkovskiy, A.; Liao, H. Y. M. YOLOv7: Trainable Bag-Of-Freebies sets new State-Of-The-Art for Real-Time object detectors[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver: IEEE Press, 2023: 7464-7475.
[20]

Lu, W. J.; Niu, C. Y.; Lan, C. Z.; et al. High-Quality object detection method for uav images based on improved dino and masked image modeling[J]. Remote Sensing, 2023, 15(19): 4740-4751.

[21]
Hu, Z. C.; Liu, J.; Jiang, C.; et al. A high-precision detection method for coated fuel particles based on improved faster Region-Based convolutional neural network[J]. Computers in Industry, 2022, 143.
[22]

Sun, Y. M.; Cao, B.; Zhu, P. F.; et al. Drone-Based RGB-Infrared Cross-Modality vehicle detection via uncertainty-aware learning[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2022, 32(10): 6700-6713.

[23]
Wen, F.; Wang, M.; Hu, X. J. DFAM-DETR: Deformable feature based attention mechanism DETR on slender object detection[J]. IEICE Transactions on Information and Systems, 2023, E106. D(3): 401-409.
Journal of Highway and Transportation Research and Development (English Edition)
Pages 48-60
Cite this article:
Zhang X, Zhu X. BaAM-YOLO: a balanced feature fusion and attention mechanism based vehicle detection network in aerial images. Journal of Highway and Transportation Research and Development (English Edition), 2024, 18(3): 48-60. https://doi.org/10.26599/HTRD.2024.9480022

226

Views

19

Downloads

0

Crossref

Altmetrics

Received: 08 October 2023
Revised: 25 May 2024
Accepted: 06 June 2024
Published: 30 September 2024
© The Author(s) 2024.

This is an open access article under the terms of the Creative Commons Attribution 4.0 International License (CC BY 4.0, http://creativecommons.org/licenses/by/4.0/).

Return