[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3687488.3687544acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccirConference Proceedingsconference-collections
research-article

Pedestrian Detection in Underground Coal Mines with an Improved YOLOv7 Algorithm

Published: 18 November 2024 Publication History

Abstract

To achieve intelligent monitoring and unmanned driving in coal mines, accurate pedestrian detection plays a crucial role. However, due to various complex factors such as uneven lighting, dense dust, cable interference, and obstacles underground in coal mines, traditional detection methods often struggle to accurately detect pedestrians. To address this challenge, an improved algorithm based on the YOLOv7 network is proposed. This algorithm introduces a deformable attention mechanism in the Backbone, allowing it to dynamically adjust the shape and size of regions to focus on important features of the image. Additionally, in the channel boosting structure (CBS), an activate or not (ACON) function is used after batch normalization (BN), which adaptively adjusts the activation output based on the input amplitude. This helps to better maintain gradient flow and promote more stable and efficient training. Specifically, the Wise IoU (WIoU) loss function is introduced into the model. This loss function provides a comprehensive approach that considers both spatial overlap and confidence scores, enabling the model to learn more robust and accurate predictions. Experimental results show that the [email protected] reaches 96.7%, and the precision rate reaches 97.2%. These values are 1.5% and 1.9% higher than those of YOLOv7, respectively.

References

[1]
Kai, H.; Xianmin, M. 2017. Research on avoidance obstacle strategy of coal underground inspection robot based on binocular vision. In 2017 29th Chinese Control and Decision Conference (CCDC), Chongqing, China, 28-30 (May 2017); pp. 6732-6737.
[2]
Ribeiro, D.; Nascimento, J. C.l; Bernardino, A.; Carneiro, G. 2017. Improving the performance of pedestrian detectors using convolutional learning. Pattern Recogn (2017), 61, 641-649.
[3]
Ren, S.; He, K.; Girshick, R.; Sun, J. 2015. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 7–12 (December 2015); pp. 91-99.
[4]
He, K.; Gkioxari, G.; Dollár, P., Girshick, R. 2017. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision (ICCV), Venice, Italy, 22-27 (October 2017); pp. 2961-2969.
[5]
Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. 2016. Ssd: Single shot multibox detector. In Proceedings of the 14th European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 (October 2016); Springer: Cham, Switzerland; pp. 21–37.
[6]
Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. 2016. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), Las Vegas, Nevada, USA, 26 June -1 (July 2016); pp. 779-788.
[7]
Mathew, M. P.; Mahesh, T. Y. 2022. Leaf-based disease detection in bell pepper plant using YOLO v5. Signal, Image and Video Processing (SIVP) (2022), 1-7.
[8]
Norkobil Saydirasulovich, S., Abdusalomov, A., Jamil, M. K., Nasimov, R., Kozhamzharova, D., Cho, Y. I. 2023. A YOLOv6-based improved fire detection approach for smart city environments. Sensors (2023), 23, 3161.
[9]
Wang, C. Y., Bochkovskiy, A., Liao, H. Y. M. 2023. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, Canada. 18 -22 (June 2023); pp. 7464-7475.
[10]
Xu, Z., Li, J., Zhang, M. 2021. A surveillance video real-time analysis system based on edge-cloud and fl-yolo cooperation in coal mine. IEEE Access (2021), 9, 68482-68497.
[11]
Xu, Z., Li, J., Meng, Y., Zhang, X. 2022. Cap-yolo: Channel attention based pruning yolo for coal mine real-time intelligent monitoring. Sensors (2022), 22, 4331.
[12]
Zhang, Y., Zhou, Y. 2021. YOLOv5 Based Pedestrian Safety Detection in Underground Coal Mines. In 2021 IEEE International Conference on Robotics and Biomimetics (ROBIO), Sanya, China, 6-9 (December 2021); pp. 1700-1705.
[13]
Wei, X., Zhang, H., Liu, S., Lu, Y. 2020. Pedestrian detection in underground mines via parallel feature transfer network. Pattern Recogn (2020), 103, 107195.
[14]
Xia, Z., Pan, X., Song, S., Li, L. E., Huang, G. 2022. Vision transformer with deformable attention. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (CVPR), New Orleans, Louisiana, USA, 19 -24 (June 2022); pp. 4794-4803.
[15]
Tong, Z., Chen, Y., Xu, Z., Yu, R. 2023. Wise-IoU: Bounding Box Regression Loss with Dynamic Focusing Mechanism. arXiv (2023), arXiv:2301.10051.

Index Terms

  1. Pedestrian Detection in Underground Coal Mines with an Improved YOLOv7 Algorithm

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICCIR '24: Proceedings of the 2024 4th International Conference on Control and Intelligent Robotics
    June 2024
    399 pages
    ISBN:9798400709937
    DOI:10.1145/3687488
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 18 November 2024

    Check for updates

    Author Tags

    1. activate or not (ACON)
    2. deformable attention mechanism
    3. pedestrian detection
    4. wise IoU (WIoU) loss function

    Qualifiers

    • Research-article

    Conference

    ICCIR 2024

    Acceptance Rates

    Overall Acceptance Rate 131 of 239 submissions, 55%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 13
      Total Downloads
    • Downloads (Last 12 months)13
    • Downloads (Last 6 weeks)13
    Reflects downloads up to 12 Dec 2024

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media