[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Forest fire detection utilizing ghost Swin transformer with attention and auxiliary geometric loss

Published: 21 November 2024 Publication History

Abstract

Forest fires are a devastating natural disaster. Existing fire detection models face limitations in dataset availability, multi-scale feature extraction, and locating obscured or small flames and smoke. To address these issues, we develop a dataset containing real and synthetic forest fire images, sourced from a UAV (Unmanned Aerial Vehicle) perspective. Additionally, we propose the Ghost Convolution Swin Transformer (GCST) module to extract multi-scale flame and smoke features from different receptive fields by integrating parallel Ghost convolution and Swin Transformer. Subsequently, we design a lightweight reparameterized rotation attention module, which captures interactions across channel and spatial dimensions to suppress background noise and focus on obscured flames and smoke. Finally, we introduce a loss function, called Efficient Auxiliary Geometric Intersection over Union (EAGIoU), which employs an auxiliary bounding box to accelerate the model's convergence while integrating the geometrical principles of the predicted and real bounding boxes to accurately locate small flames and smoke. Extensive experimental results demonstrate that our method achieves 75.2 % [email protected] and 42 % [email protected]:0.95 with a frame rate of 239 frames per second, indicating a significant improvement in accuracy and real-time performance compared to state-of-the-art techniques. The code and datasets are available at https://github.com/luckylil/forest-fire-detection.

Reference

[1]
D. Gutmacher, U. Hoefer, J. Wollenstein, Gas sensor technologies for fire detection, Sens. Actuators B 175 (2012) 40–45.
[2]
D. Huang, X. Zhao, Q. Li, et al., Real-Time ultraviolet flame detection system based on 4H-SiC phototransistor, IEEE Trans. Electron Devices (2024).
[3]
H. Yar, W. Ullah, Z.A. Khan, et al., An effective attention-based CNN model for fire detection in adverse weather conditions, ISPRS J. Photogramm. Remote Sens. 206 (2023) 335–346.
[4]
Y. Ahn, H. Choi, B.S Kim, Development of early fire detection model for buildings using computer vision-based CCTV, J. Build.Eng. 65 (2023).
[5]
C. Li, G. Li, Y. Song, et al., Fast forest fire detection and segmentation application for UAV-assisted mobile edge computing system, IEEE IoT J. (2023).
[6]
M. Zhao, L. Li, W. Li, R. Tao, L. Li, W. Zhang, Infrared small-target detection based on multiple morphological profiles, IEEE Trans. Geosci. Remote Sens. 59 (7) (2020) 6077–6091.
[7]
L. Huang, G. Liu, Y. Wang, H. Yuan, T. Chen, Fire detection in video surveillances using convolutional neural networks and wavelet transform, Eng. Appl. Artif. Intell. 110 (2022).
[8]
F. Yuan, K. Li, C. Wang, et al., A lightweight network for smoke semantic segmentation, Pattern Recognit. 137 (2023).
[9]
L. Zhao, L. Zhi, C. Zhao, W. Zheng, Fire-yolo: a small target object detection method for fire inspection, Sustainability 14 (9) (2022) 4930.
[10]
W. Bo, J. Liu, X. Fan, T. Tjahjadi, Q. Ye, L. Fu, Basnet: burned area segmentation network for real-time detection of damage maps in remote sensing images, IEEE Trans. Geosci. Remote Sens. 60 (2022) 1–13.
[11]
X. Wang, Z. Pan, H. Gao, et al., An efficient model for real-time wildfire detection in complex scenarios based on multi-head attention mechanism, J. Real-Time Image Process. 20 (4) (2023) 66.
[12]
J. Huang, Z. He, Y. Guan, H. Zhang, Real-time forest fire detection by ensemble lightweight yolox-l and defogging method, Sensors 23 (4) (2023) 1894.
[13]
J. Qian, D. Bai, W. Jiao, et al., A high-precision ensemble model for forest fire detection in large and small targets, Forests 14 (10) (2023) 2089.
[14]
X. Jia, Y. Wang, T. Chen, Forest fire detection and recognition using yolov8 algorithms from UAVS images, in: Proceedings of the 2023 IEEE 5th International Conference on Power, Intelligent Computing and Systems (ICPICS), IEEE, 2023, pp. 646–651.
[15]
G. Chen, R. Cheng, X. Lin, et al., LMDFS: a lightweight model for detecting forest fire smoke in UAV images based on YOLOv7, Remote Sens. 15 (15) (2023) 3790.
[16]
S. Wu, X. Zhang, R. Liu, B. Li, A dataset for fire and smoke object detection, Multimed. Tools Appl. 82 (5) (2023) 6707–6726.
[17]
G. Wang, H. Ding, M. Duan, Y. Pu, Z. Yang, H. Li, Fighting against terrorism: a real-time cctv autonomous weapons detection based on improved yolo v4, Digit Signal Process 132 (2023).
[18]
Z. Song, et al., Synthetic datasets for autonomous driving: a survey, IEEE Trans. Intell. Veh. 9 (1) (Jan. 2024) 1847–1864,.
[19]
J. Song, X. Qin, J. Lei, et al., A fault detection method for transmission line components based on synthetic dataset and improved YOLOv5, Int. J. Electr. Power Energy Syst. 157 (2024).
[20]
G. Wang, H. Li, P. Li, X. Lang, Y. Feng, Z. Ding, S. Xie, M4sfwd: a multi-faceted synthetic dataset for remote sensing forest wildfires detection, Expert Syst. Appl. (2024).
[21]
L. Zhang, C. Lu, H. Xu, A. Chen, L. Li, G“MMFNet: forest fire smoke detection using multiscale convergence coordinated pyramid network with mixed attention and fast-robust NMS, IEEE IoT J. 10 (20) (2023) 18168–18180,. 15 Oct.15.
[22]
L. M. Jin, Y. Yu, J. Zhou, et al., SWVR: a Lightweight Deep Learning Algorithm for Forest Fire Detection and Recognition, Forests 15 (1) (2024) 204.
[23]
X. Geng, Y. Su, X. Cao, et al., YOLOFM: an improved fire and smoke object detection algorithm based on YOLOv5n, Sci. Rep. 14 (1) (2024) 4543.
[24]
K. Han, Y. Wang, Q. Tian, J. Guo, C. Xu, C. Xu, Ghostnet: more features from cheap operations, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 1580–1589.
[25]
Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, B. Guo, Swin transformer: hierarchical vision transformer using shifted windows, in: Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 10012–10022.
[26]
K.W. Lau, L.M. Po, Y.A.U Rehman, Large separable kernel attention: rethinking the large kernel attention design in cnn, Expert Syst. Appl. 236 (2024).
[27]
J. Hu, L. Shen, G. Sun, Squeeze-and-excitation networks, in: Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 7132–7141.
[28]
Q. Hou, D. Zhou, J. Feng, Coordinate attention for efficient mobile network design, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 13713–13722.
[29]
Z. Zheng, P. Wang, W. Liu, J. Li, R. Ye, D. Ren, Distance-iou loss: faster and better learning for bounding box regression, in: Proceedings of the AAAI conference on artificial intelligence, 34, 2020, pp. 12993–13000.
[30]
Z. Tong, Y. Chen, Z. Xu, R. Yu, Wise-iou: bounding box regression loss with dynamic focusing mechanism, arXiv preprint (2023).
[31]
M. Siliang, X. Yong, Mpdiou: a loss for efficient and accurate bounding box regression, arXiv preprint (2023).
[32]
H. Zhang, C. Xu, S. Zhang, Inner-iou: more effective intersection over union loss with auxiliary bounding box, arXiv preprint (2023).
[33]
Jocher, G., Chaurasia, A., & Qiu, J. (2023). Ultralytics YOLO (Version 8.0.0) [Computer software]. https://github.com/ultralytics/ultralytics.
[34]
D. Misra, T. Nalamada, A.U. Arasanipalai, et al., Rotate to attend: convolutional triplet attention module, in: Proceedings of the IEEE/CVF winter conference on applications of computer vision, 2021, pp. 3139–3148.
[35]
X. Ding, X. Zhang, N. Ma, et al., Repvgg: making VGG-style convnets great again, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 13733–13742.
[36]
J. Redmon, A. Farhadi, Yolov3: an incremental improvement, arXiv preprint (2018).
[37]
Bochkovskiy A., Wang C.Y., Liao H Y M. Yolov4: optimal speed and accuracy of object detection. arXiv preprint, 2020.
[38]
G. Jocher, A. Stoken, J. Borovec, A. Chaurasia, L. Changyu, A. Hogan, J. Hajek, L. Diaconu, Y. Kwon, Y. Defretin, et al., ultralytics/yolov5: v5. 0-yolov5-p6 1280 models, aws, supervise. Ly and Youtube Integrations, Zenodo, 2021.
[39]
C. Li, L. Li, H. Jiang, K. Weng, Y. Geng, L. Li, Z. Ke, Q. Li, M. Cheng, W. Nie, et al., Yolov6: a single-stage object detection framework for industrial applications, arXiv preprint (2022).
[40]
C.-Y. Wang, A. Bochkovskiy, H.-Y.M. Liao, Yolov7: trainable bag-offreebies sets new state-of-the-art for real-time object detectors, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 7464–7475.
[41]
Wang C.Y., Yeh I.H., Liao H Y M. YOLOv9: learning What You Want to Learn Using Programmable Gradient Information. arXiv preprint, 2024.
[42]
S. Ren, K. He, R. Girshick, J. Sun, Faster r-cnn: towards real-time object detection with region proposal networks, Adv. Neural Inf. Process. Syst. 28 (2015).
[43]
Z. Cai, N. Vasconcelos, Cascade r-cnn: delving into high quality object detection, in: Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 6154–6162.
[44]
E.J.Y. Koh, E. Amini, G.J. McLachlan, et al., Utilising convolutional neural networks to perform fast automated modal mineralogy analysis for thin-section optical microscopy, Miner. Eng. 173 (2021).
[45]
W. Liu, D. Anguelov, D. Erhan, et al., Ssd: single shot multibox detector, in: Computer Vision–ECCV 2016: 14th European Conference, Springer International Publishing, Amsterdam, The Netherlands, 2016, pp. 21–37. October 11–14, 2016, Proceedings, Part I 14.
[46]
Zhu X., Su W., Lu L., et al. Deformable detr: deformable transformers for end-to-end object detection. arXiv preprint, 2020.
[47]
N. Carion, F. Massa, G. Synnaeve, et al., End-to-end object detection with transformers, in: Proceedings of the European Conference on Computer Vision, Springer International Publishing, Cham, 2020, pp. 213–229.
[48]
Z. Tian, C. Shen, H. Chen, et al., FCOS: a simple and strong anchor-free object detector, IEEE Trans. Pattern Anal. Mach. Intell. 44 (4) (2020) 1922–1933.
[49]
Li Y., Ren F. Light-weight retinanet for object detection, arXiv preprint, 2019.
[50]
Zhao Y., Lv W., Xu S., et al. Detrs beat yolos on real-time object detection, arXiv preprint, 2023.
[51]
J. Cai, J. Hu, X. Tang, et al., Deep historical long short-term memory network for action recognition, Neurocomputing 407 (2020) 428–438.
[52]
J. Cai, Y. Li, B. Liu, et al., Developing deep LSTMs with later temporal attention for predicting COVID-19 severity, clinical outcome, and antibody level by screening serological indicators over time, IEEE J. Biomed. Health Inform. (2024).

Index Terms

  1. Forest fire detection utilizing ghost Swin transformer with attention and auxiliary geometric loss
            Index terms have been assigned to the content through auto-classification.

            Recommendations

            Comments

            Please enable JavaScript to view thecomments powered by Disqus.

            Information & Contributors

            Information

            Published In

            cover image Digital Signal Processing
            Digital Signal Processing  Volume 154, Issue C
            Nov 2024
            623 pages

            Publisher

            Academic Press, Inc.

            United States

            Publication History

            Published: 21 November 2024

            Author Tags

            1. Dataset
            2. Efficient auxiliary geometric intersection over union
            3. Multi-scale feature extraction
            4. Obscured
            5. Reparameterized rotation attention
            6. Small flames and smoke

            Qualifiers

            • Research-article

            Contributors

            Other Metrics

            Bibliometrics & Citations

            Bibliometrics

            Article Metrics

            • 0
              Total Citations
            • 0
              Total Downloads
            • Downloads (Last 12 months)0
            • Downloads (Last 6 weeks)0
            Reflects downloads up to 30 Dec 2024

            Other Metrics

            Citations

            View Options

            View options

            Media

            Figures

            Other

            Tables

            Share

            Share

            Share this Publication link

            Share on social media