An Accurate Classification of Rice Diseases Based on ICAI-V4
<p>Process of acquisition, transmission, enhancement (flip, crop), and preprocessing of four rice disease images.</p> "> Figure 2
<p>Main flow chart of rice disease classification.</p> "> Figure 3
<p>Main process of rice leaf image processing with Candy algorithm.</p> "> Figure 4
<p>Picture utilizing the gravitational edge detection algorithm and Candy algorithm for processing rice leaf images.</p> "> Figure 5
<p>Block diagram of ICAI-V4 model. It includes INCV blocks and coordinate attention network structure.</p> "> Figure 6
<p>Schematic diagram of involution.</p> "> Figure 7
<p>Accuracy of 10-fold cross-validation training results.</p> "> Figure 8
<p>Classification confusion matrix of citrus diseases by different models (the models corresponding to the (<b>A</b>–<b>G</b>) confusion matrix are: AlexNet, ResNet50, Inceptionv4, ResNeXt, MobileNetv3, DenseNet121, and ICAI-V4).</p> ">
Abstract
:1. Introduction
- This paper proposes an algorithm for enhancing Candy images. The proposed method involves extracting the edge mask through ideal high-pass filtering and gravitational field intensity operator. The simple detail enhancement image is obtained by combining smooth filtering and a Laplace operator. The resulting image is then multiplied by the mask and added to the input image to obtain a better edge detail enhanced image.
- This paper proposes the ICAI-V4 model, which incorporates a coordinated attention mechanism into the backbone network. This enhancement is designed to improve the feature capture ability and overall performance of the network model. The INCV backbone structure incorporates both Inception-iv and Reduction-iv structures, while also integrating involution to enhance the network’s feature extraction capabilities from a channel perspective. This enables the network to better classify similar images of rice pests and diseases. Leaky ReLU is utilized as an alternative to the ReLU activation function to mitigate the issue of neuron death caused by the latter and enhance the model’s resilience.
2. Materials and Methods
2.1. Data Acquisition
- Rice bacterial blight is caused by the bacterium Xanthomonas, which produces water-stained lesions starting from the leaf margin a few centimeters from the tip of the leaf and spreading to the leaf base. The affected area increases in length and width, and changes color from yellow to light brown due to dryness [19].
- Rice blast disease is characterized by the appearance of green-gray spots on infected leaves, which are surrounded by dark green borders. As the disease progresses, the spots or lesions become oval or spindle-shaped with reddish-brown borders, and some may even become rhombus-shaped. The lesions can expand and merge, ultimately leading to the complete destruction of the leaves.
- Rice brown spot is a fungal disease caused by Bipolaris oryzae that primarily damages the aboveground parts of rice plants, particularly the leaves. It is prevalent in all rice-growing regions, especially when there is a shortage of fertilizer and water, leading to poor growth of the rice. It often occurs in combination with rice blight. The disease causes leaf blight, resulting in reduced 1000-grain weight and an increase in empty grains, which adversely affects the yield and quality of rice. Although the harm of disease in rice production is decreasing with improved fertilization and water conservation, certain areas still experience increased incidence of ear blight in late rice due to prolonged seedling age, resulting in significant harvest losses.
- Rice tungro disease causes infected leaves to turn from orange to yellow, with the staining spreading from the tip to the lower part of the leaf. Infected leaves may exhibit a striped appearance and rusty spots, and may contain planthoppers [17].
2.2. Candy Image Enhancement Algorithm
- To remove noise, a Gaussian filter is applied to smooth the image. The filter is selected based on the appropriate Gaussian function to smooth the image according to the rows and columns, which is achieved by convoluting the image matrix. The Canny algorithm typically uses a two-dimensional Gaussian function (Equation (1)) to achieve this, as the convolution operation satisfies the commutative and associative laws.
- The parameter σ in the Gaussian filter determines the extent to which the image is smoothed, affecting its expansion.
- The intensity gradient of an image can be determined by analyzing the change in pixels. When there is a significant difference between adjacent pixels, it can suggest the presence of an edge in that particular area. This edge is visually represented as a transition from black to white. The first-order partial derivative in the x and y directions can be approximated using the following formula (Equations (2) and (3)):
- 4.
- Non-maximum suppression is a technique used to eliminate false edge detections. It involves suppressing pixels whose gradient is not large enough and retaining only the maximum gradient, resulting in thin edges. This classical thin-edge algorithm is applied after obtaining the gradient magnitude image to accurately locate the edge. The Canny algorithm utilizes 3 × 3 adjacent regions, consisting of eight directions to interpolate the gradient amplitude along the gradient direction. If the magnitude is greater than the two interpolation results in the gradient direction, it is identified as a candidate edge point; otherwise, it is marked as a non-edge point. This process generates a candidate edge image.
- 5.
- To determine the possible boundary, the double threshold method is applied. Despite non-maximum suppression, there may still be noise points in the image. Therefore, the Canny algorithm employs a technique in which a threshold upper bound and a threshold lower bound are set. The process involves marking pixels with a gradient amplitude higher than the high threshold as edge points and those with a lower gradient amplitude as non-edge points. The remaining pixels are marked as candidate edge points. Candidate edge points connected to edge points are then marked as edge points, reducing the impact of noise on the final edge image.
- 6.
- The boundary is tracked using the lag technique. The candidate edge points are re-evaluated and the 8-connected domain pixels of a weak edge point are examined. If there are any strong edge points present, the weak edge point is considered to be retained as part of the edge. However, if there are no strong edge points, the weak edge is suppressed.
- The first step in the list number image processing pipeline involves preprocessing the original image. This includes normalization and conversion to grayscale if the original image is in color. The resulting normalized and grayscale image is then used as input for the subsequent steps.
- The process of extracting the mask and detail enhancement images involves two steps. Firstly, the input image is processed using the improved canny algorithm to extract the edge details. The detected image edge information is then utilized as the mask. In the second step, the input image undergoes a process of smoothing and denoising. To deal with the noise, the Laplace operator is used, which can have a strong response to the noise and cause negative effects. Therefore, denoising processing should be carried out prior to using the Laplacian operator. This allows the denoised image to better highlight small details in the image.
- A preliminary detail-enhanced image can be obtained by adding the result of the Laplacian operator to the input image. However, since the Laplacian operator is isotropic, it detects isolated points effectively, but may cause loss of the edges of the square. Therefore, further operations are required to preserve the information.
- To extract the image well, the processed mask image is multiplied with the preliminary enhanced image. The image enhancement algorithm results from adding the edge and detail information to the input image [33].
2.3. Identification of Rice Diseases Based on ICAI-V4 Model
- The initial layer of ICAI-V4 is the Stem layer, which primarily serves to rapidly decrease the resolution of the feature map. This reduction in resolution allows for subsequent inceptions to decrease the amount of computation required. Additionally, the activation function of the standard convolution layer has been modified from ReLU to Leaky ReLU, which enhances the network’s robustness.
- The second module in ICAI-V4 is the coordination attention module. This module is capable of capturing cross-channel information, as well as direction-aware and position-sensitive information. This enables the model to more accurately locate and identify objects of interest and enhance features by emphasizing information representation.
- The third component of ICAI-V4 is the INCV structure, which consists of 1 × 1 convolution, 3 × 3 convolution, pooling layer, and asymmetric convolution decomposition. The involution layer is incorporated into various structures to enhance the flexibility of different channels and improve the overall network learning effect. The network comprises three Inception-iv structures and two Reduction-iv structures used multiple times. Refer to Figure 5 for a detailed illustration of the specific structure.
- ICAI-V4’s fourth component consists of an adaptive pooling layer, data dimensionality reduction, linear layer, and a discard layer. The classification results of input rice pest images are determined by a softmax activation function.
2.3.1. Coordinate Attention
2.3.2. INCV Blocks
- Sharing kernels on the channel improves performance by allowing us to use large spatial spans while maintaining design efficiency through the channel dimension. This is true even if weights are not shared in different spatial locations, as it does not significantly increase the number of parameters and calculations.
- While the kernel parameters of each pixel in the space are not directly shared, involution does share meta-weights at a higher level, specifically the parameters of the kernel generating function. This allows us to still share and migrate knowledge across different spatial locations. However, freeing the limitation of the convolution sharing kernel in space and allowing each pixel to learn its corresponding kernel parameters does not necessarily lead to better results, despite addressing the issue of large parameter increase.
2.3.3. Leaky ReLU
- When the input is positive, there is no gradient saturation problem.
- The calculation speed is much faster. There is only a linear relationship in the ReLU function, so its calculation speed is faster than the Sigmoid function and the tanh function [41].
- The Dead ReLU problem arises when the input to the ReLU activation function is negative, rendering it completely invalid. This is not a concern during forward propagation as only certain areas are sensitive to negative input. However, during back propagation, if a negative number is encountered, the gradient will be zero, resulting in non-convergence of calculation results and neuronal death. This prevents weight updates, leading to the problem of non-updating neurons.
- The output of the ReLU function is zero or positive, which means that the ReLU function is not a zero-centered function.
- Some neurons will not die. The Dead ReLU problem of the ReLU function is solved;
- The linear and unsaturated form of Leaky ReLU can converge quickly in SGD;
- The Leaky ReLU function has a faster calculation speed compared to the sigmoid and tanh functions. This is because it has a linear relationship and does not require exponential calculations. Nonlinearity typically requires more computation and thus results in slower execution speeds.
2.4. Evaluating Indicator
3. Results
3.1. Experimental Environment and Data Preparation
3.2. Experimental Design and Results
3.3. Verifying the Effectiveness of Data Preprocessing
3.4. Ablation Experiment
3.5. Comparison with Other Networks
3.6. Classification Performance Evaluation of ICAI-V4 and Other Networks
3.7. Formatting of Mathematical Components
4. Discussion
- This study utilized the Candy algorithm to enhance images prior to input classification. The results, as shown in Figure 3, demonstrate that this method significantly improves the visual quality of images compared to traditional filtering methods. The algorithm enhances edge details while suppressing noise, which maintains the authenticity of the image. The results of Table 7 indicate that the classification accuracy of the model for rice diseases improved by 5.21% when using this method compared to no image enhancement.
- This academic paper introduces a coordinated attention mechanism into the backbone network to enhance the feature capture capability and overall performance of the network model. The experiment in Table 8 shows that the classification accuracy of Inception-V4, AlexNet, and ResNeXt for rice diseases improved after the addition of the coordinated attention mechanism (+7.79%, +1.57%, +2.14%).
- The INCV trunk structure is a combination of Inception-iv and Reduction-iv structures, with the addition of involution to enable more flexible modeling in spatial dimensions. This approach also incorporates a channel dimension shared kernel with a space-specific kernel to enhance the feature extraction capability of the entire network. As a result, the classification of similar rice disease images is improved. This paper incorporates involution into Inception-V4 and conducts experiments by adding involution to each part. The results, presented in Table 4, demonstrate that this matching enhances the network’s feature extraction ability and improves its classification accuracy (81.27%, 82.51%, 82.93%, 85.14%, 85.94%, 87.74%).
- The use of Leaky ReLU as a replacement for the ReLU activation function has been shown to alleviate the problem of neuron death caused by the latter, while also enhancing the model’s elasticity. In fact, Table 8 demonstrates that changing the activation function to Leaky ReLU resulted in a 2.87% improvement in the classification accuracy of rice diseases.
- According to our research, this method has shown superior classification accuracy compared to other deep neural network models. The average accuracy is 95.57%, with a recall rate of 95.54% and an F1-score of 95.81%.
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Yang, X.; Pan, Y.; Qing, D.; Xia, X.; Zhou, W.; Chen, W.; Li, J.; Liang, H.; Yang, X.; Li, D.; et al. Dissecting the Molecular Basis for Genetic Improvement of Rice, 22 June 2022, PREPRINT (Version 1) Available at Research Square. Available online: https://doi.org/10.21203/rs.3.rs-1749870/v1 (accessed on 27 January 2023).
- Li, P.; Chen, Y.H.; Lu, J.; Zhang, C.Q.; Liu, Q.Q.; Li, Q.F. Genes and their molecular functions determining seed structure, components, and quality of rice. Rice 2022, 15, 18. [Google Scholar] [CrossRef] [PubMed]
- Yang, X.; Xu, H.; Zang, H.; Feng, X.; Pan, R.; Gu, C.; Gao, T. Current situation and countermeasures of rice disease prevention and control. Mod. Pestic. 2022, 21, 1–5. [Google Scholar] [CrossRef]
- Guo, D. Research on Identification Method of Rice Sheath Blight Based on Deep Learning. Ph.D. Thesis, Heilongjiang Bayi Agricultural University, Mishan, China, 2020. [Google Scholar] [CrossRef]
- Khanal, S.; Fulton, J.; Shearer, S. An overview of current and potential applications of thermal remote sensing in precision agriculture. Comput. Electron. Agric. 2017, 139, 22–32. [Google Scholar] [CrossRef]
- Jagyasi, B.; Mohite, J.; Pappula, S. Applications of mobile sensing technologies in precision agriculture. CSI Commun. 2013, 18, 21–23. [Google Scholar]
- Sartori, D.; Brunelli, D. A smart sensor for precision agriculture powered by microbial fuel cells. In Proceedings of the 2016 IEEE Sensors Applications Symposium (SAS), Catania, Italy, 20–22 April 2016; pp. 1–6. [Google Scholar] [CrossRef]
- Puri, V.; Nayyar, A.; Raja, L. Agriculture drones: A modern breakthrough in precision agriculture. J. Stat. Manag. Syst. 2017, 20, 507–518. [Google Scholar] [CrossRef]
- Mekala, M.S.; Viswanathan, P. A Survey: Smart agriculture IoT with cloud computing. In Proceedings of the 2017 International Conference on Microelectronic Devices, Circuits and Systems (ICMDCS), Vellore, India, 10–12 August 2017; pp. 1–7. [Google Scholar] [CrossRef]
- Jha, K.; Doshi, A.; Patel, P.; Shah, M. A comprehensive review on automation in agriculture using artificial intelligence. Artif. Intell. Agric. 2019, 2, 1–12. [Google Scholar] [CrossRef]
- Ahmed, N.; De, D.; Hussain, I. Internet of Things (IoT) for Smart Precision Agriculture and Farming in Rural Areas. IEEE Internet Things J. 2018, 5, 4890–4899. [Google Scholar] [CrossRef]
- Huynh-The, T.; Gadekallu T, R.; Wang, W.; Yenduri, G.; Ranaweera, P.; Pham, Q.V.; da Costa, D.B.; Liyanage, M. Blockchain for the metaverse: A Review. Future Gener. Comput. Syst. 2023, 143, 401–419. [Google Scholar] [CrossRef]
- Khattab, A.; Habib, S.E.; Ismail, H.; Zayan, S.; Fahmy, Y.; Khairy, M.M. An IoT-based cognitive monitoring system for early plant disease forecast. Comput. Electron. Agric. 2019, 1, 105028. [Google Scholar] [CrossRef]
- Yamagishi, Y.; Kato, Y.; Ninomiya, S.; Guo, W. Image-Based Phenotyping for Non-Destructive In Situ Rice (Oryza sativa L.) Tiller Counting Using Proximal Sensing. Sensors 2022, 22, 5547. [Google Scholar] [CrossRef]
- Lv, L.; Peng, M.; Wang, X.; Wu, Y. Multi-scale information fusion network with label smoothing strategy for corneal ulcer classification in slit lamp images. Front. Neurosci. 2022, 16, 993234. [Google Scholar] [CrossRef] [PubMed]
- Liu, X.; Hu, Y.; Zhou, G.; Cai, W.; He, M.; Zhan, J.; Hu, Y.; Li, L. DS-MENet for the classification of citrus disease. Front. Plant Sci. 2022, 13, 88446. [Google Scholar] [CrossRef] [PubMed]
- Bera, T.; Das, A.; Sil, J.; Das, A.K. A Survey on Rice Plant Disease Identification Using Image Processing and Data Mining Techniques. Emerg. Technol. Data Min. Inf. Secur. 2018, 3, 365–376. [Google Scholar] [CrossRef]
- Polesel, A.; Ramponi, G.; Mathews, V.J. Image enhancement via adaptive unsharp masking. IEEE Trans. Image Process. 2000, 9, 505–510. [Google Scholar] [CrossRef] [Green Version]
- Phadikar, S.; Sil, J.; Das, A.K. Rice diseases classification using feature selection and rule generation techniques. Comput. Electron. Agric. 2013, 90, 76–85. [Google Scholar] [CrossRef]
- Kumawat, A.; Panda, S. A robust edge detection algorithm based on feature-based image registration (FBIR) using improved canny with fuzzy logic (ICWFL). Vis. Comput. 2022, 38, 3681–3702. [Google Scholar] [CrossRef]
- Xuan, L.; Hong, Z. An improved canny edge detection algorithm. In Proceedings of the 2017 8th IEEE International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 24–26 November 2017; pp. 275–278. [Google Scholar] [CrossRef]
- Kalbasi, M.; Nikmehr, H. Noise-robust, reconfigurable canny edge detection and its hardware realization. IEEE Access 2020, 8, 39934–39945. [Google Scholar] [CrossRef]
- Wang, Y.Y.; Zhou, Z.G.; Luo, L.K. Image enhancement algorithm based on sobel operator filtering. Comput. Appl. Softw. 2019, 36, 184–188. [Google Scholar] [CrossRef]
- Lu, Y.; Yi, S.; Zeng, N.; Liu, Y.; Zhang, Y. Identification of rice diseases using deep convolutional neural networks. Neurocomputing 2017, 267, 378–384. [Google Scholar] [CrossRef]
- Atole, R.R.; Park, D. A multiclass deep convolutional neural network classifier for detection of common rice plant anomalies. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 1. [Google Scholar]
- Zhou, G.; Zhang, W.; Chen, A.; He, M.; Ma, X. Rapid detection of rice disease based on FCM-KM and faster R-CNN fusion. IEEE Access 2019, 7, 143190–143206. [Google Scholar] [CrossRef]
- Rahman, C.R.; Arko, P.S.; Ali, M.E.; Khan, M.A.I.; Apon, S.H.; Nowrin, F.; Wasif, A. Identification and recognition of rice diseases and pests using convolutional neural networks. Biosyst. Eng. 2020, 194, 112–120. [Google Scholar] [CrossRef] [Green Version]
- Wang, Y.; Wang, H.; Peng, Z. Rice diseases detection and classification using attention based neural network and bayesian optimization. Expert Syst. Appl. 2021, 178, 114770. [Google Scholar] [CrossRef]
- Haque, M.E.; Rahman, A.; Junaeid, I.; Hoque, S.U.; Paul, M. Rice leaf disease classification and detection using yolov5. arXiv 2022, arXiv:2209.01579. [Google Scholar] [CrossRef]
- Venkatesan, R.; Er, M.J.; Dave, M.; Pratama, M.; Wu, S. A novel online multi-label classifier for high-speed streaming data applications. Evol. Syst. 2017, 8, 303–315. [Google Scholar] [CrossRef] [Green Version]
- Rong, W.; Li, Z.; Zhang, W.; Sun, L. An improved Canny edge detection algorithm. In Proceedings of the 2014 IEEE International Conference on Mechatronics and Automation, Tianjin, China, 3–6 August 2014; pp. 577–582. [Google Scholar] [CrossRef]
- Li, R.; Song, B.; Liu, X.; Pan, X.; Han, L.; Xu, Y. Smoking Behavior Detection Based on TF-YOLOv5. In Proceedings of the 2022 3rd International Conference on Pattern Recognition and Machine Learning (PRML), Chengdu, China, 15–17 July 2022; pp. 14–19. [Google Scholar] [CrossRef]
- Qaddoori, A.S.; Saud, J.H. Comparison of edge discrimination of microbubbles between Sobel and improved mathematics of Canny edge detection. Int. J. Nonlinear Anal. Appl. 2022, 13, 817–829. [Google Scholar] [CrossRef]
- Li, D.; Hu, J.; Wang, C.; Li, X.; She, Q.; Zhu, L.; Zhang, T.; Chen, Q. Involution: Inverting the inherence of convolution for visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12321–12330. [Google Scholar] [CrossRef]
- Pan, C.; Fei, S. Violence detection based on attention mechanism. In Proceedings of the 2022 41st Chinese Control Conference (CCC), Hefei, China, 25–27 July 2022; pp. 6438–6443. [Google Scholar] [CrossRef]
- Cheng, S.; Cheng, H.; Yang, R.; Zhou, J.; Li, Z.; Shi, B.; Lee, M.; Ma, Q. A High Performance Wheat Disease Detection Based on Position Information. Plants 2023, 12, 1191. [Google Scholar] [CrossRef]
- Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar] [CrossRef]
- Guo, G.; Zhang, Z. Road damage detection algorithm for improved YOLOv5. Sci. Rep. 2022, 12, 15523. [Google Scholar] [CrossRef]
- Li, M. Dual Channel with Involution for Long-Tailed Visual Recognition. Open J. Appl. Sci. 2022, 12, 421–433. [Google Scholar] [CrossRef]
- Lei, X.; Lin, X.; Zhong, Y.; Chen, Q. Optimization of neural network structure using involution operator based on particle swarm optimization for image classification. In Proceedings of the 2021 11th International Conference on Information Technology in Medicine and Education (ITME), Wuyishan, China, 19–21 November 2021; pp. 168–173. [Google Scholar] [CrossRef]
- Liu, J.; Yin, L.; Hu, Y.; Lv, S.; Sun, L. A Novel Intrusion Detection Algorithm for Industrial Control Systems Based on CNN and Process State Transition. In Proceedings of the 2018 IEEE 37th International Performance Computing and Communications Conference (IPCCC), Orlando, FL, USA, 17–19 November 2018; pp. 1–8. [Google Scholar] [CrossRef]
- Tarek, H.; Aly, H.; Eisa, S.; Abul-Soud, M. Optimized deep learning algorithms for tomato leaf disease detection with hardware deployment. Electronics 2022, 11, 140. [Google Scholar] [CrossRef]
- Chen, C.-C.; Ba, J.Y.; Li, T.J.; Ki Chan, C.C.; Wang, K.C.; Liu, Z. EfficientNet: A low-bandwidth iot image sensor framework for cassava leaf disease classification. Sens. Mater. 2021, 33, 4031–4044. [Google Scholar] [CrossRef]
- Huang, G.; Liu, Z.; Laurens, V.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef] [Green Version]
- Sunil, C.K.; Jaidhar, C.D.; Patil, N. Cardamom plant disease detection approach using EfficientNetV2. IEEE Access 2022, 10, 789–804. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar] [CrossRef] [Green Version]
- Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef] [Green Version]
- Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar] [CrossRef]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
Disease Type | Kaggle | Mendeley Data | Expanded | Sum | Percentage (%) |
---|---|---|---|---|---|
Bacterial blight | 1584 | 935 | 145 | 2551 | 26.70% |
Blast | 1440 | 894 | 289 | 2531 | 24.28% |
Brown spot | 1600 | 872 | 569 | 2845 | 26.97% |
Tungro | 1308 | 654 | 525 | 2314 | 22.05% |
Total | 5932 | 3355 | 1528 | 10,241 | 100% |
Symbol | Description |
---|---|
The parameters of the Gaussian filter | |
Approximation of the first partial derivative in the X direction | |
Approximation of the first partial derivative in the Y direction | |
The image gradient operator in the x direction | |
The image gradient operator in the y direction | |
Image gradient size | |
The azimuth of the image gradient | |
Intensity of gravitational field | |
The component of the gradient in the X direction | |
The component of the gradient in the Y direction | |
Gradient magnitude | |
Gradient azimuth | |
The average of the magnitude of the gradient | |
The threshold of a pixel | |
Involution kernel height | |
Involution kernel width | |
kernel size | |
An index set in the neighborhood of coordinates | |
A patch in the feature map contains | |
Coordinate Attention Block output Y | |
TP | The number of samples that were predicted to be positive |
FP | The number of samples that are actually negative, but predicted to be positive |
FN | The number of samples that are actually positive, but predicted to be negative |
TN | The number of samples that are predicted to be negative |
Acronym | Full Term |
---|---|
AI | Artificial Intelligence |
CA | Coordinate Attention |
FN | False Negative |
FP | False Positive |
GIS | Geographic Information System |
IoT | Internet of Things |
IT | Information Technology |
TN | True Negative |
TP | True Positive |
Network Structure | Accuracy |
---|---|
Baseline | 81.27 |
Inception_ivA | 82.51 |
Reduction_ivA | 82.93 |
Inception_ivB | 85.14 |
Reduction_ivB | 85.94 |
Inception_ivC | 87.74 |
Confusion Matrix | Prediction | ||
---|---|---|---|
True | False | ||
Practical | True | TP | FN |
False | FP | TN |
Hyperparameters | Value |
---|---|
Learning rate | 0.001 |
Epoch | 20 |
Momentum | 0.9 |
Batch size | 32 |
Optimizer | RMSprop |
Image Enhancement Methods | Original Dataset (%) | Extended Dataset (%) |
---|---|---|
No image enhancement | 81.27 | 89.26 |
Image enhancement | 86.48 | 95.57 |
Model | InceptionV4 | AlexNet | ResNeXt |
---|---|---|---|
Baseline | 81.27 | 89.68 | 89.07 |
CA | 89.06 | 91.25 | 91.21 |
INCV | 87.74 | 92.14 | 90.13 |
Leaky ReLU | 84.14 | 90.23 | 89.91 |
CA+INCV | 94.25 | 94.35 | 91.25 |
CA+Leaky ReLU | 91.24 | 93.13 | 90.63 |
INCV+Leaky ReLU | 90.25 | 92.53 | 91.25 |
CA+INCV+Leaky ReLU | 95.57 | 93.23 | 92.14 |
Methods | Accuracy (Test Set without Noise) (%) | Accuracy (Test Set Added Noise) (%) |
---|---|---|
AlexNet | 89.68 | 84.72 |
ResNet50 | 84.48 | 80.67 |
Inception V4 | 81.27 | 79.31 |
ResNeXt50 | 89.07 | 87.45 |
MobileNetV2 | 91.31 | 89.14 |
DenseNet121 | 93.31 | 93.23 |
ICAI-V4 | 95.57 | 95.25 |
Methods | Evaluation Indicators | Rice Disease | |||
---|---|---|---|---|---|
Bacterial Blight | Blast | Brown Spot | Tungro | ||
AlexNet | Precision (%) | 93.41% | 88.40% | 89.16% | 87.05% |
Recall (%) | 91.35% | 86.31% | 88.43% | 92.50% | |
F1-score (%) | 92.36% | 87.34% | 88.79% | 89.69% | |
ResNet50 | Precision (%) | 86.49% | 81.29% | 85.05% | 84.19% |
Recall (%) | 84.91% | 81.18% | 83.25% | 88.37% | |
F1-score (%) | 85.69% | 81.23% | 84.14% | 86.22% | |
Inception V4 | Precision (%) | 81.63% | 77.97% | 82.51% | 82.56% |
Recall (%) | 83.08% | 79.16% | 80.25% | 82.18% | |
F1-score (%) | 82.34% | 78.56% | 81.36% | 82.36% | |
ResNeXt50 | Precision (%) | 91.84% | 84.95% | 89.20% | 89.67% |
Recall (%) | 90.34% | 86.25% | 87.31% | 92.27% | |
F1-score (%) | 91.08% | 85.59% | 88.24% | 90.95% | |
MobileNetV2 | Precision (%) | 93.07% | 91.06% | 90.99% | 89.55% |
Recall (%) | 92.48% | 89.16% | 90.31% | 93.11% | |
F1-score (%) | 92.77% | 90.09% | 90.64% | 91.29% | |
DenseNet121 | Precision (%) | 95.01% | 91.73% | 91.50% | 95.18% |
Recall (%) | 96.21% | 90.20% | 91.50% | 95.18% | |
F1-score (%) | 95.06% | 90.95% | 91.50% | 95.18% | |
ICAI-V4 | Precision (%) | 97.00% | 94.58% | 96.39% | 96.47% |
Recall (%) | 98.10% | 92.22% | 93.62% | 98.24% | |
F1-score (%) | 97.54% | 93.38% | 94.98% | 97.34% |
Dataset | Category | Total | Available |
---|---|---|---|
PlantVillage | 38 | 55,400 | https://www.kaggle.com/datasets/hiyash99/plantvillage (accessed on 24 March 2023) |
Stanford cars | 196 | 16,185 | https://www.kaggle.com/datasets/jutrera/stanford-car-dataset-by-classes-folder (accessed on 24 March 2023) |
ImageNetDogs | 120 | 20,580 | http://vision.stanford.edu/aditya86/ImageNetDogs/ (accessed on 24 March 2023) |
Dataset | Accuracy (%) |
---|---|
PlantVillage | 96.25 |
Stanford cars | 94.67 |
ImageNetDogs | 97.24 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zeng, N.; Gong, G.; Zhou, G.; Hu, C. An Accurate Classification of Rice Diseases Based on ICAI-V4. Plants 2023, 12, 2225. https://doi.org/10.3390/plants12112225
Zeng N, Gong G, Zhou G, Hu C. An Accurate Classification of Rice Diseases Based on ICAI-V4. Plants. 2023; 12(11):2225. https://doi.org/10.3390/plants12112225
Chicago/Turabian StyleZeng, Nanxin, Gufeng Gong, Guoxiong Zhou, and Can Hu. 2023. "An Accurate Classification of Rice Diseases Based on ICAI-V4" Plants 12, no. 11: 2225. https://doi.org/10.3390/plants12112225
APA StyleZeng, N., Gong, G., Zhou, G., & Hu, C. (2023). An Accurate Classification of Rice Diseases Based on ICAI-V4. Plants, 12(11), 2225. https://doi.org/10.3390/plants12112225