Fast Personal Protective Equipment Detection for Real Construction Sites Using Deep Learning Approaches
<p>Main kinds of fatal accident for workers in 2018/19 in the UK [<a href="#B3-sensors-21-03478" class="html-bibr">3</a>].</p> "> Figure 2
<p>Helmet colors and the roles on construction sites [<a href="#B7-sensors-21-03478" class="html-bibr">7</a>].</p> "> Figure 3
<p>The proposed framework of PPE detection based on YOLO models.</p> "> Figure 4
<p>Architectures of YOLO v4 and YOLO v5s.</p> "> Figure 5
<p>Number of images and instances in the training, validation and test datasets.</p> "> Figure 6
<p>Number of instances in each scale category (calculated on resizing each image to 408 × 408).</p> "> Figure 7
<p>Examples of intersection and union.</p> "> Figure 8
<p>Precision × Recall curves.</p> "> Figure 9
<p>Examples of detection results.</p> "> Figure 10
<p>Process time (GPU) and mAP of each YOLO v5 models.</p> "> Figure 11
<p>Wrong detections in YOLO v5x.</p> "> Figure 12
<p>Mean average precision in each model.</p> "> Figure 13
<p>Average time for processing one image in each model.</p> "> Figure 14
<p>Blurring face test on YOLO v5x.</p> ">
Abstract
:1. Introduction
2. Methodology
2.1. Framework
2.2. YOLO Family
2.2.1. YOLO v3 and Previous Versions
2.2.2. YOLO v4 and YOLO v5
2.3. CHV Dataset
2.4. PPE Detector
2.4.1. Experiment Goals
2.4.2. Blurring Face Test
2.4.3. Training Process
2.5. Metrics for Performance Evaluation
3. Results and Discussions
3.1. Results
3.2. Different Layers
3.3. Different Train Size
3.4. Different Model Sizes
3.5. Where Are Wrong Detections?
3.6. All YOLO Models
3.7. Blurring Face Test
3.8. Comparison with the State-of-the-Art
4. Conclusions
- A new dataset, CHV dataset, was constructed to detect the desired six classes of PPE. It contained 1330 high-quality images from the real construction sites, selected from more than 10,000 images, based on strict criteria. To foster future innovative research work, the dataset was publicly available on the GitHub page at https://github.com/ZijianWang1995/ppe_detection (accessed on 13 May 2021);
- This paper expanded the detection scope to more PPE characteristics, including people, four helmet colors and the vest. The detection results could measure the safety compliance, and contribute to the management;
- Different versions of YOLO with different parameters were tested systematically in terms of accuracy and speed. For YOLO v3 models, different detection layers were tested, while an increased number of layers could not improve performance. For YOLO v4 models, the increase in training image size could not contribute to a better performance. Additionally, YOLO v5x owned the best mAP (86.55%), and YOLO v5s had a faster processing speed for one single image (52 FPS). Even compared with previous studies, YOLO v5 models still had the best performance;
- Detection errors for YOLO v5x were analyzed in more depth. In the False Negative (miss detection) cases, it was found that small, blocked, and strange-gesture instances were hard to detect. In the False Positive (wrong detection) cases, on the one hand, the detector mis-classified the helmet colors or mistakenly regarded green clothes as vests. On the other hand, while the detection class was correct, the detection size was not appropriate for the ground truth size;
- YOLO v5x was adopted to test blurring face images. The detection of the vest and person was not affected by the blurring areas, while the average precision of four helmet colors decreased by around 7% when the blurring area covered the helmet part.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- U.S. Bureau of Labor Statistics. Industries at A Glance: Retail. 2018. Available online: https://www.bls.gov/iag/tgs/iag23.htm (accessed on 10 May 2020).
- Ministry of Housing and Urban-Rural Development. The Report of Fatal Accidents in China’s Building Construction Activities in 2018. 2019. Available online: http://www.mohurd.gov.cn/wjfb/201903/t20190326_239913.html (accessed on 10 May 2020).
- Fatal Injuries in Great Britain. Technical Report. 2017. Available online: https://www.hse.gov.uk/statistics/fatals.htm (accessed on 10 June 2020).
- OSHA. Worker Safety Series Construction. 2005. Available online: https://www.osha.gov/Publications/OSHA3252/3252.html (accessed on 10 May 2020).
- Hume, A.; Mills, N.J. Industrial head injuries and the performance of the helmets. Biomech. Impacts 1995, 217–231. Available online: https://doi.org/http://www.ircobi.org/wordpress/downloads/irc1995/pdf_files/1995_15.pdf (accessed on 10 May 2020).
- Suderman, B.L.; Hoover, R.W.; Ching, R.P.; Scher, I.S. The effect of hardhats on head and neck response to vertical impacts from large construction objects. Accid. Anal. Prev. 2014, 73, 116–124. [Google Scholar] [CrossRef] [PubMed]
- Safety Helmet Colours. Technical Report. 2015. Available online: https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_data/file/573103/B31_Safety_Helmet_Colours_Sept_16_V2.pdf (accessed on 5 June 2020).
- Zhang, S.; Teizer, J.; Pradhananga, N.; Eastman, C.M. Workforce location tracking to model, visualize and analyze workspace requirements in building information models for construction safety planning. Autom. Constr. 2015, 60, 74–86. [Google Scholar] [CrossRef]
- Kelm, A.; Laußat, L.; Meins-Becker, A.; Platz, D.; Khazaee, M.J.; Costin, A.M.; Helmus, M.; Teizer, J. Mobile passive Radio Frequency Identification (RFID) portal for automated and rapid control of Personal Protective Equipment (PPE) on construction sites. Autom. Constr. 2013, 36, 38–52. [Google Scholar] [CrossRef]
- Zhang, H.; Yan, X.; Li, H.; Jin, R.; Fu, H.F. Real-Time Alarming, Monitoring, and Locating for Non-Hard-Hat Use in Construction. J. Constr. Eng. Manag. 2019, 145, 1–13. [Google Scholar] [CrossRef]
- Seo, J.; Han, S.; Lee, S.; Kim, H. Computer vision techniques for construction safety and health monitoring. Adv. Eng. Inform. 2015, 29, 239–251. [Google Scholar] [CrossRef]
- Han, S.; Lee, S. A vision-based motion capture and recognition framework for behavior-based safety management. Autom. Constr. 2013, 35, 131–141. [Google Scholar] [CrossRef]
- Cheng, T.; Teizer, J. Modeling tower crane operator visibility to minimize the risk of limited situational awareness. J. Comput. Civ. Eng. 2014, 28, 04014004. [Google Scholar] [CrossRef]
- Zhu, Z.; Park, M.W.; Elsafty, N. Automated monitoring of hardhats wearing for onsite safety enhancement. In Proceedings of the 11th Construction Specialty Conference, Vancouver, UK, 8–10 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
- Rubaiyat, A.H.; Toma, T.T.; Kalantari-Khandani, M.; Rahman, S.A.; Chen, L.; Ye, Y.; Pan, C.S. Automatic detection of helmet uses for construction safety. In Proceedings of the 2016 IEEE International Conference on Web Intelligence Workshops, Omaha, NE, USA, 13–16 October 2016; pp. 135–142. [Google Scholar] [CrossRef]
- Shrestha, K.; Shrestha, P.P.; Bajracharya, D.; Yfantis, E.A. Hard-Hat Detection for Construction Safety Visualization. J. Constr. Eng. 2015, 2015, 1–8. [Google Scholar] [CrossRef]
- Du, S.; Shehata, M.; Badawy, W. Hard hat detection in video sequences based on face features, motion and color information. In Proceedings of the 2011 3rd International Conference on Computer Research and Development, Shanghai, China, 11–13 March 2011; Volume 4, pp. 25–29. [Google Scholar] [CrossRef]
- Nath, N.D.; Behzadan, A.H. Deep Learning Detection of Personal Protective Equipment to Maintain Safety Compliance on Construction Sites. In Construction Research Congress 2020: Computer Applications; American Society of Civil Engineers: Reston, VA, USA, 2020; pp. 181–190. [Google Scholar] [CrossRef]
- Wu, H.; Zhao, J. Automated Visual Helmet Identification Based on Deep Convolutional Neural Networks; Elsevier Masson SAS: Amsterdam, The Netherlands, 2018; Volume 44, pp. 2299–2304. [Google Scholar] [CrossRef]
- Pradana, R.D.W.; Adhitya, R.Y.; Syai’in, M.; Sudibyo, R.M.; Abiyoga, D.R.A.; Jami’in, M.A.; Subiyanto, L.; Herijono, B.; Wahidin, A.; Budianto, A.; et al. MIdentification System of Personal Protective Equipment Using Convolutional Neural Network (CNN) Method. In Proceedings of the 2019 IEEE International Symposium on Electronics and Smart Devices, Rourkela, India, 16–18 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
- Akbarzadeh, M.; Zhu, Z.; Hammad, A. Nested Network for Detecting PPE on Large Construction Sites Based on Frame Segmentation. In Creative Construction e-Conference 2020; Budapest University of Technology and Economics: Budapest, Hungary, 2020; pp. 33–38. [Google Scholar] [CrossRef]
- Wu, J.; Cai, N.; Chen, W.; Wang, H.; Wang, G. Automatic detection of hardhats worn by construction personnel: A deep learning approach and benchmark dataset. Autom. Constr. 2019, 106, 102894. [Google Scholar] [CrossRef]
- Chen, S.; Demachi, K. A Vision-Based Approach for Ensuring Proper Use of Personal Protective Equipment (PPE) in Decommissioning of Fukushima Daiichi Nuclear Power Station. Appl. Sci. 2020, 10, 5129. [Google Scholar] [CrossRef]
- Xie, Z.; Liu, H.; Li, Z.; He, Y. A convolutional neural network based approach towards real-time hard hat detection. In Proceedings of the 2018 IEEE International Conference on Progress in Informatics and Computing, Suzhou, China, 14–16 December 2018; pp. 430–434. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef] [Green Version]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar] [CrossRef] [Green Version]
- Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
- Ultralytics. yolov5. Available online: https://github.com/ultralytics/yolov5 (accessed on 15 May 2020).
- Wang, C.Y.; Mark Liao, H.Y.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar] [CrossRef]
- Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Curran Associates, Inc.: Nice, France, 2019; pp. 8024–8035. [Google Scholar]
- njvisionpower. Safety-Helmet-Wearing-Dataset. 2019. Available online: https://github.com/njvisionpower/Safety-Helmet-Wearing-Dataset (accessed on 10 May 2020).
- Qipeng, D.; Ping, K.; Li, F.; He, M.; Yu, G. Method of Safety Helmet Wearing Detection based on Key-Point Estimation without Anchor. In Proceedings of the 2020 17th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), Chengdu, China, 18–20 December 2020; pp. 93–96. [Google Scholar] [CrossRef]
- Tzutalin. LabelImg. 2015. Available online: https://github.com/tzutalin/labelImg (accessed on 20 May 2020).
- Alexey, A.B. Darknet. Available online: https://github.com/AlexeyAB/darknet#how-to-train-to-detect-your-custom-objects (accessed on 10 May 2020).
- Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2001. [Google Scholar] [CrossRef]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–26 June 2005; Volume 1, pp. 886–893. [Google Scholar] [CrossRef] [Green Version]
- Wang, H.; Li, Z.; Ji, X.; Wang, Y. Face r-cnn. arXiv 2017, arXiv:1706.01061. [Google Scholar]
- Chen, W.; Huang, H.; Peng, S.; Zhou, C.; Zhang, C. YOLO-face: A real-time face detector. Vis. Comput. 2020, 1–9. [Google Scholar] [CrossRef]
- Padilla, R.; Netto, S.L.; da Silva, E.A.B. A Survey on Performance Metrics for Object-Detection Algorithms. In Proceedings of the 2020 International Conference on Systems, Signals and Image Processing, Marrakech, Morocco, 4–6 June 2020; pp. 237–242. [Google Scholar] [CrossRef]
- Rafaelpadilla. Object Detection Metrics. 2019. Available online: https://github.com/rafaelpadilla/Object-Detection-Metrics (accessed on 30 May 2020).
- Nath, N.D.; Behzadan, A.H.; Paal, S.G. Deep learning for site safety: Real-time detection of personal protective equipment. Autom. Constr. 2020, 112, 103085. [Google Scholar] [CrossRef]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar] [CrossRef] [Green Version]
No. | Goal | Model | Note |
---|---|---|---|
1 | Layer | YOLO v3 (3 layers) | Default configure |
2 | YOLO v3 (5 layers) | Two more layers | |
3 | Training image size | YOLO v4 (416) | Default configure |
4 | YOLO v4 (608) | Larger training image size | |
5 | Model size | YOLO v5s | Small |
6 | YOLO v5m | Medium | |
7 | YOLO v5l | Large | |
8 | YOLO v5x | Extra large | |
9 | Blurring face | YOLO v5x | Protect private information |
Model | DL Library | Optimizer | Initial Learning Rate | Momentum | Decay | Batch Size |
---|---|---|---|---|---|---|
YOLO v3/4 | Darknet | SGD | 0.001 | 0.949 | 0.0005 | 64 |
YOLO v5 | PyTorch | SGD | 0.01 | 0.937 | 0.0005 | 32 |
Platform | Location | OS | GPU | GPU Size | CPU Cores | RAM |
---|---|---|---|---|---|---|
Traning | Cloud | Ubuntu 16.04 | Tesla P40 | 24 GB | 8 | 48 GB |
Test | Local | macOS Catalina | None | None | 4 | 8 GB |
Model | Weight Size | AP | mAP | Time (ms) | ||||||
---|---|---|---|---|---|---|---|---|---|---|
Person | Vest | Blue | Red | White | Yellow | GPU | CPU | |||
YOLO v3 (3layers) | 246 MB | 79.76% | 79.30% | 80.65% | 86.35% | 86.77% | 83.04% | 82.65% | 27.15 | 5375.02 |
YOLO v3 (5layers) | 248 MB | 77.98% | 77.95% | 77.51% | 83.31% | 88.19% | 87.00% | 81.99% | 37.04 | 6338.74 |
Model | Train Image Size | AP | mAP | Time (ms) | ||||||
---|---|---|---|---|---|---|---|---|---|---|
Person | Vest | Blue | Red | White | Yellow | GPU | CPU | |||
YOLO v4 (416) | 416 × 416 | 85.17% | 82.86% | 78.14% | 79.63% | 90.57% | 88.61% | 84.16% | 30.18 | 7773.10 |
YOLO v4 (608) | 608 × 608 | 86.10% | 81.54% | 85.31% | 83.33% | 88.15% | 84.61% | 84.84% | 30.43 | 7648.62 |
Model | Weight Size | AP | |||||
---|---|---|---|---|---|---|---|
Person | Vest | Blue | Red | White | Yellow | ||
YOLO v5x | 178 MB | 83.77% | 81.47% | 80.76% | 91.91% | 89.96% | 91.41% |
YOLO v5l | 95 MB | 83.11% | 83.64% | 78.09% | 92.04% | 89.19% | 90.74% |
YOLO v5m | 43 MB | 83.05% | 82.14% | 67.07% | 84.68% | 88.88% | 90.49% |
YOLO v5s | 15 MB | 82.96% | 76.56% | 74.81% | 87.73% | 85.06% | 88.79% |
Category | Person | Vest | Blue | Red | White | Yellow | |
---|---|---|---|---|---|---|---|
TP | 385 | 182 | 36 | 47 | 161 | 135 | |
FN | 65 | 35 | 8 | 3 | 15 | 12 | |
FP | IoU = 0 | 13 | 15 | 1 | 6 | 16 | 7 |
0 < IoU < 0.5 | 32 | 19 | 2 | 3 | 4 | 3 |
Models | Person | Vest | Blue | Red | White | Yellow | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
TP | FP | FN | TP | FP | FN | TP | FP | FN | TP | FP | FN | TP | FP | FN | TP | FP | FN | |
YOLO v3 (3 layers) | 369 | 39 | 81 | 179 | 26 | 38 | 36 | 2 | 8 | 46 | 8 | 4 | 155 | 17 | 21 | 125 | 16 | 22 |
YOLO v3 (5 layers) | 363 | 47 | 87 | 177 | 35 | 40 | 35 | 5 | 9 | 44 | 9 | 6 | 158 | 16 | 18 | 129 | 14 | 18 |
YOLO v4 (416) | 393 | 56 | 57 | 189 | 44 | 28 | 35 | 5 | 9 | 42 | 12 | 8 | 163 | 23 | 13 | 132 | 23 | 15 |
YOLO v4 (608) | 401 | 79 | 49 | 189 | 64 | 28 | 38 | 5 | 6 | 43 | 16 | 7 | 160 | 35 | 16 | 128 | 25 | 19 |
YOLO v5s | 382 | 39 | 68 | 174 | 30 | 43 | 34 | 5 | 10 | 45 | 6 | 5 | 152 | 21 | 24 | 133 | 12 | 14 |
YOLO v5m | 381 | 42 | 69 | 185 | 27 | 32 | 30 | 4 | 14 | 43 | 7 | 7 | 158 | 20 | 18 | 136 | 19 | 11 |
YOLO v5l | 382 | 48 | 68 | 188 | 24 | 29 | 35 | 6 | 9 | 47 | 8 | 3 | 160 | 13 | 16 | 135 | 11 | 12 |
YOLO v5x | 385 | 45 | 65 | 182 | 34 | 35 | 36 | 3 | 8 | 47 | 9 | 3 | 161 | 20 | 15 | 135 | 10 | 12 |
Face | Models | Person | Vest | Blue | Red | White | Yellow | mAP |
---|---|---|---|---|---|---|---|---|
Clear | YOLO v5x | 83.77% | 81.47% | 80.76% | 91.91% | 89.96% | 91.41% | 86.55% |
Blurring | YOLO v5x | 82.78% | 82.46% | 73.78% | 74.59% | 78.46% | 85.17% | 79.55% |
Dataset | Pictor-v3 (Crowd-Sourced Part) | GDUT-HWD | |||
---|---|---|---|---|---|
Method | YOLO v3 [42] | YOLO v5x * | Pelee-RPA [22] | YOLO v5x * | |
Class | Person | × | 85.74% | × | 86.71% |
Vest | 84.96% | × | × | 81.58% | |
Blue | × | × | 77.35% | 82.74% | |
Red | × | 87.39% | 68.58% | 81.87% | |
White | × | 72.98% | 74.64% | 86.56% | |
Yellow | × | 81.05% | 78.48% | 81.10% | |
Helmet | 79.81% | 80.47% | 72.34% | 83.07% | |
mAP | 72.30% | 81.79% | 72.34% | 83.43% |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, Z.; Wu, Y.; Yang, L.; Thirunavukarasu, A.; Evison, C.; Zhao, Y. Fast Personal Protective Equipment Detection for Real Construction Sites Using Deep Learning Approaches. Sensors 2021, 21, 3478. https://doi.org/10.3390/s21103478
Wang Z, Wu Y, Yang L, Thirunavukarasu A, Evison C, Zhao Y. Fast Personal Protective Equipment Detection for Real Construction Sites Using Deep Learning Approaches. Sensors. 2021; 21(10):3478. https://doi.org/10.3390/s21103478
Chicago/Turabian StyleWang, Zijian, Yimin Wu, Lichao Yang, Arjun Thirunavukarasu, Colin Evison, and Yifan Zhao. 2021. "Fast Personal Protective Equipment Detection for Real Construction Sites Using Deep Learning Approaches" Sensors 21, no. 10: 3478. https://doi.org/10.3390/s21103478
APA StyleWang, Z., Wu, Y., Yang, L., Thirunavukarasu, A., Evison, C., & Zhao, Y. (2021). Fast Personal Protective Equipment Detection for Real Construction Sites Using Deep Learning Approaches. Sensors, 21(10), 3478. https://doi.org/10.3390/s21103478