Enhancing YOLOv5s with Attention Mechanisms for Object Detection in Complex Backgrounds Environment
Abstract
Full Text:
PDF (140-147)References
H. Da, “Complex Environment Road Object Detection Algorithm Based on Improved YOLOv5s,” in 2024 6th International Conference on Data-driven Optimization of Complex Systems (DOCS), 2024, pp. 625–630. doi: 10.1109/DOCS63458.2024.10704511.
J. Zhong, Q. Cheng, X. Hu, and Z. Liu, “YOLO Adaptive Developments in Complex Natural Environments for Tiny Object Detection,” Electronics (Switzerland), vol. 13, no. 13, Jul. 2024, doi: 10.3390/electronics13132525.
J. Ruan, H. Cui, Y. Huang, T. Li, C. Wu, and K. Zhang, “A review of occluded objects detection in real complex scenarios for autonomous driving,” Green Energy and Intelligent Transportation, vol. 2, no. 3, p. 100092, 2023, doi: https://doi.org/10.1016/j.geits.2023.100092.
E. Yurtsever, J. Lambert, A. Carballo, and K. Takeda, “A Survey of Autonomous Driving: Common Practices and Emerging Technologies,” IEEE Access, vol. 8, pp. 58443–58469, 2020, doi: 10.1109/ACCESS.2020.2983149.
C. Baoyuan, L. Yitong, and S. Kun, “Research on Object Detection Method Based on FF-YOLO for Complex Scenes,” IEEE Access, vol. 9, pp. 127950–127960, 2021, doi: 10.1109/ACCESS.2021.3108398.
D. Peng, W. Ding, and T. Zhen, “A novel low light object detection method based on the YOLOv5 fusion feature enhancement,” Sci Rep, vol. 14, no. 1, p. 4486, 2024, doi: 10.1038/s41598-024-54428-8.
W.-Y. Hsu and W.-Y. Lin, “Adaptive Fusion of Multi-Scale YOLO for Pedestrian Detection,” IEEE Access, vol. 9, pp. 110063–110073, 2021, doi: 10.1109/ACCESS.2021.3102600.
X. Ren, W. Zhang, M. Wu, C. Li, and X. Wang, “Meta-YOLO: Meta-Learning for Few-Shot Traffic Sign Detection via Decoupling Dependencies,” Applied Sciences, vol. 12, no. 11, 2022, doi: 10.3390/app12115543.
F. Cao et al., “An Efficient Object Detection Algorithm Based on Improved YOLOv5 for High-Spatial-Resolution Remote Sensing Images,” Remote Sens (Basel), vol. 15, no. 15, Aug. 2023, doi: 10.3390/rs15153755.
Y. Li, M. Zhang, C. Zhang, H. Liang, P. Li, and W. Zhang, “YOLO-CCS: Vehicle detection algorithm based on coordinate attention mechanism,” Digit Signal Process, 2024, doi: 10.1016/j.dsp.2024.104632.
Q. Su and J. Mu, “Complex Scene Occluded Object Detection with Fusion of Mixed Local Channel Attention and Multi-Detection Layer Anchor-Free Optimization,” Automation, vol. 5, no. 2, pp. 176–189, Jun. 2024, doi: 10.3390/automation5020011.
C. Conversion, “Citypersons Dataset,” Dec. 2022, Roboflow. [Online]. Available: https://universe.roboflow.com/citypersons-conversion/citypersons-woqjq
J. R. Terven and D. M. Cordova-Esparza, “A COMPREHENSIVE REVIEW OF YOLO ARCHITECTURES IN COMPUTER VISION: FROM YOLOV1 TO YOLOV8 AND YOLO-NAS PUBLISHED AS A JOURNAL PAPER AT MACHINE LEARNING AND KNOWLEDGE EXTRACTION.”
N. Jegham, C. Y. Koh, M. Abdelatti, and A. Hendawi, “YOLO Evolution: A Comprehensive Benchmark and Architectural Review of YOLOv12, YOLO11, and Their Previous Versions,” Mar. 2025, [Online]. Available: http://arxiv.org/abs/2411.00201
S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “CBAM: Convolutional Block Attention Module,” Jul. 2018, [Online]. Available: http://arxiv.org/abs/1807.06521
Q. Wang, B. Wu, P. Zhu, P. Li, W. Zuo, and Q. Hu, “ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks,” Oct. 2019, [Online]. Available: http://arxiv.org/abs/1910.03151
Q. Hou, D. Zhou, and J. Feng, “Coordinate Attention for Efficient Mobile Network Design,” Mar. 2021, [Online]. Available: http://arxiv.org/abs/2103.02907
M.-H. Guo et al., “Attention Mechanisms in Computer Vision: A Survey,” Nov. 2021, doi: 10.1007/s41095-022-0271-y.
Z. Ren, H. Zhang, and Z. Li, “Improved YOLOv5 Network for Real-Time Object Detection in Vehicle-Mounted Camera Capture Scenarios,” Sensors, vol. 23, no. 10, May 2023, doi: 10.3390/s23104589.
S. Hao, W. Li, X. Ma, and Z. Tian, “SSE-YOLOv5: a real-time fault line selection method based on lightweight modules and attention models,” J. Real-Time Image Process., vol. 21, no. 4, May 2024, doi: 10.1007/s11554-024-01480-2.
Refbacks
- There are currently no refbacks.