A Recognition Method for Rice Plant Diseases and Pests Video Detection Based on Deep Convolutional Neural Network
<p>As shown in the red boxes, the lesion spots of rice sheath blight, rice stem borer symptoms, rice brown spot are visualized: (<b>a</b>) rice sheath blight, (<b>b</b>) rice stem borer symptoms, (<b>c</b>) rice brown spot. The images are from our image dataset, which were collected in Anhui, Jiangxi and Hunan Province, China, between June and August, 2018.</p> "> Figure 2
<p>The challenges of rice video detection. Subgraphs (<b>a</b>,<b>b</b>,<b>c</b>) are frames of our video dataset (Video 1 illuminated in <a href="#sec2dot1-sensors-20-00578" class="html-sec">Section 2.1</a>). (<b>a</b>) Video defocus, (<b>b</b>) motion blur, (<b>c</b>) part occlusion. Video 1 was captured in Hunan Province, China, July 2018.</p> "> Figure 3
<p>Images of the three kinds of rice diseases and pests. (<b>a</b>) Rice sheath blight, (<b>b</b>) rice stem borer symptoms, (<b>c</b>) rice brown spot. These images were collected in Anhui Province, etc., China, between June and August, 2018.</p> "> Figure 4
<p>The annotation of rice sheath blight in our datasets. Only lesion spots were annotated, and no withered leaves were annotated. Different colors only denote different annotation boxes, the lesions in the figure are all rice sheath blight. The images were captured in Anhui Province, China, June 2018.</p> "> Figure 5
<p>The architecture of the still-image detector, which is using the framework of Faster-RCNN, where the “CNN backbone” is the proposed deep convolutional neural network (DCNN).</p> "> Figure 6
<p>The custom DCNN video detection system results, (<b>a</b>) Video 1, which mainly has rice sheath blight. (<b>b</b>) Video 2, which mainly has rice stem borer symptoms. (<b>c</b>) Video 3, which mainly has rice brown spot. Each video has three categories of rice disease and pest symptoms. Red, purple and blue boxes denote rice sheath blight, rice stem borer symptoms, and rice brown spot respectively. Indigo ellipses magnify the lesion spots, which is realized by Adobe Photoshop CS6. The same kind of lesions is magnified once in each subgraph. Videos 1–3 were not used in the model training. Video 1 and Video 3 were captured in Hunan Province, China, July 2018, Video 2 was captured in Jiangxi Province, China, July 2018.</p> "> Figure 7
<p>Confusion matrix of Video 1 which contains rice sheath blight, rice stem borer symptoms and rice brown spot. Video 1 was captured in Hunan Province, China, July 2018, which was illustrated in the caption of <a href="#sensors-20-00578-f006" class="html-fig">Figure 6</a>. The input data of the confusion matrix was counted manually from Video 1.</p> "> Figure 8
<p>The custom DCNN video detection system qualitative results of Videos 4 and 5. (<b>a</b>) Video 4, which mainly has rice sheath blight. (<b>b</b>) Video 5, which mainly has rice stem borer symptoms. Red, purple and blue boxes denote rice sheath blight, rice stem borer symptoms, and rice brown spot respectively, and indigo ellipses magnify the lesion spots, which is realized by Adobe Photoshop CS6. The same kind of lesions is magnified once in each subgraph. Video 4 and Video 5 did not participate in model training. Video 4 was captured in Hunan Province, China, July 2018, Video 5 was captured in Jiangxi Province, China, July 2018.</p> "> Figure 9
<p>Precision-Recall curve of the still-image detector of the three other backbones and our custom DCNN backbone. Blue, yellow and green line denotes rice sheath blight, rice stem borer symptoms and rice brown spot respectively. (<b>a</b>) VGG16 with pre-trained model, (<b>b</b>) ResNet-50 with pre-trained model, (<b>c</b>) ResNet-101 with pre-trained model, (<b>d</b>) the custom DCNN training from scratch.</p> "> Figure 10
<p>Detection results of YOLOv3 using Video 1. The red box denotes rice sheath blight, green box denotes rice brown spot. Video 1 was captured in Hunan Province, China, July 2018, which was illustrated in <a href="#sensors-20-00578-f006" class="html-fig">Figure 6</a>. The detection results of YOLOv3 was not good as the custom DCNN system from the comparison of the figure and <a href="#sensors-20-00578-f006" class="html-fig">Figure 6</a>.</p> "> Figure 11
<p>Loss curve of YOLOv3. We set the iterations to 50,000. Other parameters are described in <a href="#sec3dot1-sensors-20-00578" class="html-sec">Section 3.1</a>. The training was stopped when achieving the max iterations.</p> ">
Abstract
:1. Introduction
2. Materials and Methods
2.1. Image and Video Datasets
2.2. Video Detection Metrics
2.3. The Video Detection System
2.3.1. Frame Extraction Module
2.3.2. Still-Image Object Detector
2.3.3. The Custom DCNN Backbone
2.3.4. Video Synthesis Module
3. Experiments
3.1. Training
3.2. Setting of the Detection Threshold
4. Results and Discussion
4.1. Results of the Rice Video Detection
4.2. Confusion Matrix
4.3. Analysis and Validation of the Results
4.4. Study on Different Depth Architecture of the Proposed DCNN Backbone
4.5. Comparison with Other Backbones
4.6. Comparison with the State-of-the-Art Method YOLOv3
4.7. Supplementary Evaluation of Our System
4.8. Difficulties in Rice Video Detection
4.9. Shortcomings of Our System
5. Conclusions
Supplementary Materials
Author Contributions
Funding
Conflicts of Interest
References
- Mutka, A.M.; Bart, R.S. Image-based phenotyping of plant disease symptoms. Front. Plant Sci. 2015, 5, 734. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Rush, M.; Lee, F. Rice sheath blight: A major rice disease. Plant Dis. 1983, 67, 829–832. [Google Scholar]
- Chaudhary, R.; Khush, G.; Heinrichs, E. Varietal resistance to rice stem-borers in Asia. Int. J. Trop. Insect Sci. 1984, 5, 447–463. [Google Scholar] [CrossRef]
- Liu, Z.-Y.; Huang, J.-F.; Tao, R.-X. Characterizing and estimating fungal disease severity of rice brown spot with hyperspectral reflectance data. Rice Sci. 2008, 15, 232–242. [Google Scholar] [CrossRef]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 July 2005. [Google Scholar]
- Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1997; pp. 1150–1157. [Google Scholar]
- Cireşan, D.; Meier, U.; Masci, J.; Schmidhuber, J. A committee of neural networks for traffic sign classification. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 1918–1921. [Google Scholar]
- Le Cun, Y.; Matan, O.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jacket, L.; Baird, H.S. Handwritten zip code recognition with multilayer networks. In Proceedings of the 10th International Conference on Pattern Recognition, Atlantic City, NJ, USA, 16–21 June 1990; pp. 35–40. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
- Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
- Ghazi, M.M.; Yanikoglu, B.; Aptoula, E. Plant identification using deep neural networks via optimization of transfer learning parameters. Neurocomputing 2017, 235, 228–235. [Google Scholar] [CrossRef]
- Ding, W.; Taylor, G. Automatic moth detection from trap images for pest management. Comput. Electron. Agric. 2016, 123, 17–28. [Google Scholar] [CrossRef] [Green Version]
- Grinblat, G.L.; Uzal, L.C.; Larese, M.G.; Granitto, P.M. Deep learning for plant identification using vein morphological patterns. Comput. Electron. Agric. 2016, 127, 418–424. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Lake Tahoe, NV, USA, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
- Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
- Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 779–788. [Google Scholar]
- Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
- Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Fei-Fei, L. Large-scale video classification with convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1725–1732. [Google Scholar]
- Neubeck, A.; Van Gool, L. Efficient non-maximum suppression. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; pp. 850–855. [Google Scholar]
- Han, W.; Khorrami, P.; Paine, T.L.; Ramachandran, P.; Babaeizadeh, M.; Shi, H.; Li, J.; Yan, S.; Huang, T.S. Seq-NMS for Video Object Detection. arXiv 2016, arXiv:1602.08465. [Google Scholar]
- Galteri, L.; Seidenari, L.; Bertini, M.; Del Bimbo, A. Spatio-temporal closed-loop object detection. IEEE Trans. Image Process. 2017, 26, 1253–1263. [Google Scholar] [CrossRef] [PubMed]
- Kang, K.; Li, H.; Yan, J.; Zeng, X.; Yang, B.; Xiao, T.; Zhang, C.; Wang, Z.; Wang, R.; Wang, X. T-cnn: Tubelets with convolutional neural networks for object detection from videos. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 2896–2907. [Google Scholar] [CrossRef] [Green Version]
- Zhu, X.; Wang, Y.; Dai, J.; Yuan, L.; Wei, Y. Flow-guided feature aggregation for video object detection. In Proceedings of the IEEE International Conference on Computer Vision(ICCV), Venice, Italy, 22–29 October 2017; pp. 408–417. [Google Scholar]
- Mahlein, A.-K. Plant disease detection by imaging sensors–parallels and specific demands for precision agriculture and plant phenotyping. Plant Dis. 2016, 100, 241–251. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Pethybridge, S.J.; Nelson, S.C. Leaf Doctor: A new portable application for quantifying plant disease severity. Plant Dis. 2015, 99, 1310–1316. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Stewart, E.L.; McDonald, B.A. Measuring quantitative virulence in the wheat pathogen Zymoseptoria tritici using high-throughput automated image analysis. Phytopathology 2014, 104, 985–992. [Google Scholar] [CrossRef] [Green Version]
- Bock, C.; Cook, A.; Parker, P.; Gottwald, T. Automated image analysis of the severity of foliar citrus canker symptoms. Plant Dis. 2009, 93, 660–665. [Google Scholar] [CrossRef]
- Bai, X.; Li, X.; Fu, Z.; Lv, X.; Zhang, L. A fuzzy clustering segmentation method based on neighborhood grayscale information for defining cucumber leaf spot disease images. Comput. Electron. Agric. 2017, 136, 157–165. [Google Scholar] [CrossRef]
- Dorj, U.-O.; Lee, M.; Yun, S.-S. An yield estimation in citrus orchards via fruit detection and counting using image processing. Comput. Electron. Agric. 2017, 140, 103–112. [Google Scholar] [CrossRef]
- Ma, J.; Li, X.; Wen, H.; Fu, Z.; Zhang, L. A key frame extraction method for processing greenhouse vegetables production monitoring video. Comput. Electron. Agric. 2015, 111, 92–102. [Google Scholar] [CrossRef]
- Zhang, S.; Wu, X.; You, Z.; Zhang, L. Leaf image based cucumber disease recognition using sparse representation classification. Comput. Electron. Agric. 2017, 134, 135–141. [Google Scholar] [CrossRef]
- Zhang, S.; Zhu, Y.; You, Z.; Wu, X. Fusion of superpixel, expectation maximization and PHOG for recognizing cucumber diseases. Comput. Electron. Agric. 2017, 140, 338–347. [Google Scholar] [CrossRef]
- Du, K.; Sun, Z.; Li, Y.; Zheng, F.; Chu, J.; Su, Y. Diagnostic model for wheat leaf conditions using image features and a support vector machine. Trans. ASABE 2016, 59, 1041–1052. [Google Scholar]
- Geirhos, R.; Rubisch, P.; Michaelis, C.; Bethge, M.; Wichmann, F.A.; Brendel, W. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv 2018, arXiv:1811.12231. [Google Scholar]
- Ferentinos, K.P. Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 2018, 145, 311–318. [Google Scholar] [CrossRef]
- Fuentes, A.; Yoon, S.; Kim, S.; Park, D. A robust deep-learning-based detector for real-time tomato plant diseases and pests recognition. Sensors 2017, 17, 2022. [Google Scholar] [CrossRef] [Green Version]
- Barbedo, J.G.A. A review on the main challenges in automatic plant disease identification based on visible range images. Biosyst. Eng. 2016, 144, 52–60. [Google Scholar] [CrossRef]
- Raghavan, V.; Bollmann, P.; Jung, G.S. A critical investigation of recall and precision as measures of retrieval system performance. ACM Trans. Inf. Syst. (Tois) 1989, 7, 205–229. [Google Scholar] [CrossRef]
- Aradhye, H.; Toderici, G.; Yagnik, J. Video2text: Learning to annotate video content. In Proceedings of the 2009 IEEE International Conference on Data Mining Workshops, Miami, FL, USA, 6–9 December 2009; pp. 144–151. [Google Scholar]
- Ciresan, D.C.; Meier, U.; Masci, J.; Gambardella, L.M.; Schmidhuber, J. Flexible, high performance convolutional neural networks for image classification. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011. [Google Scholar]
- Scherer, D.; Müller, A.; Behnke, S. Evaluation of pooling operations in convolutional architectures for object recognition. In Proceedings of the International Conference on Artificial Neural Networks, Thessaloniki, Greece, 15–18 September 2010; pp. 92–101. [Google Scholar]
- Tur, G.; Deng, L.; Hakkani-Tür, D.; He, X. Towards deeper understanding: Deep convex networks for semantic utterance classification. In Proceedings of the 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 25–30 March 2012; pp. 5045–5048. [Google Scholar]
- Li, Z.; Peng, C.; Yu, G.; Zhang, X.; Deng, Y.; Sun, J. Detnet: Design backbone for object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 Septemper 2018; pp. 334–350. [Google Scholar]
- Girshick, R.; Radosavovic, I.; Gkioxari, G.; Dollár, P.; He, K. Detectron. 2018. Available online: https://github.com/facebookresearch/detectron (accessed on 21 January 2020).
- Wong, S.C.; Gatt, A.; Stamatescu, V.; McDonnell, M.D. Understanding data augmentation for classification: When to warp? In Proceedings of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia, 30 November–2 December 2016; pp. 1–6. [Google Scholar]
- Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P. Natural language processing (almost) from scratch. J. Mach. Learn. Res. 2011, 12, 2493–2537. [Google Scholar]
Symptom | No. of Images | No. for Training | No. for Test |
---|---|---|---|
Rice sheath blight | 1800 | 1620 | 180 |
Rice stem borer symptoms | 1760 | 1584 | 176 |
Rice brown spot | 1760 | 1584 | 176 |
Total | 5320 | 4788 | 532 |
Video Name | Main Symptom | Frames |
---|---|---|
Video 1 | Rice sheath blight | 450 |
Video 2 | Rice stem borer symptoms | 1650 |
Video 3 | Rice brown spot | 2190 |
Video 4 | Rice sheath blight | 390 |
Video 5 | Rice stem borer symptoms | 2220 |
Total | 4290 |
Layer | Kernel Size | Stride | Repeat | Output Channels | |
---|---|---|---|---|---|
Image | 3 | ||||
block 1 | Conv. 1 | 3 × 3 | 1 | 3 | 64 |
ReLU | |||||
LRN | |||||
MaxPool | 2 × 2 | 2 | |||
block 2 | Conv. 2 | 3 × 3 | 1 | 3 | 128 |
ReLU | |||||
LRN | |||||
MaxPool | 2 × 2 | 2 | |||
block 3 | Conv. 3 | 3 × 3 | 1 | 3 | 256 |
ReLU | |||||
LRN | |||||
MaxPool | 2 × 2 | 2 | |||
block 4 | Conv. 4 | 3 × 3 | 1 | 3 | 512 |
ReLU | |||||
LRN | |||||
MaxPool | 2 × 2 | 2 |
Video 1 | Video 2 | Video 3 | |
---|---|---|---|
No. of true spots | 50 | 67 | 38 |
No. of spots detected | 39 | 91 | 40 |
No. of true spots detected | 34 | 60 | 28 |
Video recall | 68.0% | 89.6% | 73.7% |
Video precision | 87.2% | 65.9% | 70.0% |
F1 score | 76.4 | 75.9 | 71.8 |
Video Metrics | Blight Recall | Blight Precision | Borer Recall | Borer Precision | Spot Recall | Spot Precision |
---|---|---|---|---|---|---|
Video 4 | 78.6 | 84.6 | 66.7 | 50.0 | 75.0 | 85.7 |
Video 5 | 66.7 | 50.0 | 83.3 | 76.9 | 70.0 | 63.6 |
Video Metrics | Blight Recall | Blight Precision | Borer Recall | Borer Precision | Spot Recall | Spot Precision |
---|---|---|---|---|---|---|
14 conv. layers | 74.1 | 64.5 | 9.1 | 3.8 | 8.3 | 5.3 |
11 conv. layers | 29.6 | 72.7 | 18.2 | 15.4 | 33.3 | 40.0 |
Custom DCNN | 74.1 | 90.9 | 45.5 | 71.4 | 75.0 | 90.0 |
Backbone | Blight Recall | Blight Precision | Borer Recall | Borer Precision | Spot Recall | Spot Precision |
---|---|---|---|---|---|---|
VGG16 | 51.9 | 100.0 | 0 | / | 8.3 | 100.0 |
ResNet-50 | 55.6 | 93.8 | 0 | / | 0 | / |
ResNet-101 | 59.3 | 100.0 | 0 | / | 41.7 | 60.0 |
Custom DCNN | 70.4 | 90.5 | 9.1 | 100.0 | 0 | 0 |
Backbone | Blight Recall | Blight Precision | Borer Recall | Borer Precision | Spot Recall | Spot Precision |
---|---|---|---|---|---|---|
VGG16 | 18.5 | 100.0 | 9.1 | 8.3 | 16.7 | 8.7 |
ResNet-50 | 59.3 | 94.1 | 45.5 | 83.3 | 66.7 | 88.9 |
ResNet-101 | 66.7 | 94.7 | 0 | / | 16.7 | 40.0 |
Custom DCNN | 74.1 | 90.9 | 45.5 | 71.4 | 75.0 | 90.0 |
Video Metrics | Blight Recall | Blight Precision | Borer Recall | Borer Precision | Spot Recall | Spot Precision |
---|---|---|---|---|---|---|
YOLOv3 | 29.6 | 100.0 | 9.1 | 100.0 | 25.0 | 100.0 |
Custom DCNN System | 74.1 | 90.9 | 45.5 | 71.4 | 75.0 | 90.0 |
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, D.; Wang, R.; Xie, C.; Liu, L.; Zhang, J.; Li, R.; Wang, F.; Zhou, M.; Liu, W. A Recognition Method for Rice Plant Diseases and Pests Video Detection Based on Deep Convolutional Neural Network. Sensors 2020, 20, 578. https://doi.org/10.3390/s20030578
Li D, Wang R, Xie C, Liu L, Zhang J, Li R, Wang F, Zhou M, Liu W. A Recognition Method for Rice Plant Diseases and Pests Video Detection Based on Deep Convolutional Neural Network. Sensors. 2020; 20(3):578. https://doi.org/10.3390/s20030578
Chicago/Turabian StyleLi, Dengshan, Rujing Wang, Chengjun Xie, Liu Liu, Jie Zhang, Rui Li, Fangyuan Wang, Man Zhou, and Wancai Liu. 2020. "A Recognition Method for Rice Plant Diseases and Pests Video Detection Based on Deep Convolutional Neural Network" Sensors 20, no. 3: 578. https://doi.org/10.3390/s20030578
APA StyleLi, D., Wang, R., Xie, C., Liu, L., Zhang, J., Li, R., Wang, F., Zhou, M., & Liu, W. (2020). A Recognition Method for Rice Plant Diseases and Pests Video Detection Based on Deep Convolutional Neural Network. Sensors, 20(3), 578. https://doi.org/10.3390/s20030578