[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
A Comparison Study of Person Identification Using IR Array Sensors and LiDAR
Previous Article in Journal
Channel Code-Book (CCB): Semantic Image-Adaptive Transmission in Satellite–Ground Scenario
Previous Article in Special Issue
Non-Destructive Monitoring of External Quality of Date Palm Fruit (Phoenix dactylifera L.) During Frozen Storage Using Digital Camera and Flatbed Scanner
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention Score-Based Multi-Vision Transformer Technique for Plant Disease Classification

Department of AI & Big Data, Honam University, Gwangju 62399, Republic of Korea
Sensors 2025, 25(1), 270; https://doi.org/10.3390/s25010270
Submission received: 23 October 2024 / Revised: 26 December 2024 / Accepted: 2 January 2025 / Published: 6 January 2025
(This article belongs to the Special Issue Artificial Intelligence and Key Technologies of Smart Agriculture)

Abstract

:
This study proposes an advanced plant disease classification framework leveraging the Attention Score-Based Multi-Vision Transformer (Multi-ViT) model. The framework introduces a novel attention mechanism to dynamically prioritize relevant features from multiple leaf images, overcoming the limitations of single-leaf-based diagnoses. Building on the Vision Transformer (ViT) architecture, the Multi-ViT model aggregates diverse feature representations by combining outputs from multiple ViTs, each capturing unique visual patterns. This approach allows for a holistic analysis of spatially distributed symptoms, crucial for accurately diagnosing diseases in trees. Extensive experiments conducted on apple, grape, and tomato leaf disease datasets demonstrate the model’s superior performance, achieving over 99% accuracy and significantly improving F1 scores compared to traditional methods such as ResNet, VGG, and MobileNet. These findings underscore the effectiveness of the proposed model for precise and reliable plant disease classification.

1. Introduction

Plant diseases pose significant challenges to agricultural productivity and food security worldwide. Leaf diseases, in particular, can cause substantial yield losses and economic damages if left unmanaged. The timely and accurate identification of these diseases is crucial for implementing effective control measures to mitigate their impact. Traditional methods of leaf disease diagnoses often rely on visual inspection by trained experts. This process can be labor-intensive, time-consuming, and prone to subjective interpretation [1,2,3]. Moreover, the increasing complexity and variability of disease symptoms present further challenges for accurate diagnoses using conventional approaches. Visual inspections are not only resource-intensive but also limited by the availability and expertise of human diagnosticians, which can vary significantly.
To address these challenges, researchers have increasingly turned to advanced computational techniques, particularly artificial intelligence (AI) and machine learning (ML) algorithms, to develop automated systems for leaf disease detection and classification. By leveraging computer vision and pattern recognition, these approaches analyze digital images of plant leaves to identify disease symptoms with high accuracy and efficiency. AI and ML algorithms can process large volumes of image data rapidly and consistently, providing scalable solutions for timely and reliable disease diagnoses.

1.1. Materials and Methods

Recent advances in deep learning, particularly the use of convolutional neural networks (CNNs) and Vision Transformers (ViTs), have significantly enhanced the accuracy of leaf disease classification [4,5,6]. CNNs excel at capturing spatial hierarchies within images, enabling them to automatically learn and extract intricate features from raw image data. ViTs introduce self-attention mechanisms that handle larger image contexts, allowing the model to discern subtle patterns associated with various diseases more effectively. By exploiting these powerful architectures, researchers can achieve higher diagnostic precision, thereby providing critical data for early intervention and management strategies.
Building upon these developments, our approach integrates multiple leaves as input and incorporates a novel attention score mechanism for calculating the importance between embeddings extracted by Multi-Vision Transformer models. This mechanism dynamically adjusts attention weights based on the input images, prioritizing those that offer the most relevant information for disease classification. Through this method, we aim to overcome the limitations associated with conventional single-leaf examinations and leverage the spatial distribution of symptoms across multiple leaves for a more robust and comprehensive diagnostic outcome.

1.2. Challenges in Tree Disease Diagnosis

However, diagnosing diseases solely from individual leaves is often impractical, especially for tree diseases with symptoms that may be localized. Relying only on a single leaf overlooks the possibility that certain parts of the plant may show signs of disease while others appear healthy. To address this issue, our proposed method uses multiple leaves as input, providing a more complete assessment of the plant’s health and increasing the overall accuracy of disease diagnoses.
In this paper, we present a novel approach that leverages deep learning algorithms and a specialized attention scoring technique. By examining multiple leaves and dynamically adjusting attention weights, this method offers a more comprehensive, reliable, and actionable strategy for diagnosing plant diseases, ultimately contributing to more effective agricultural management practices.

2. Related Works

Several methodologies have been explored, ranging from traditional image processing techniques to modern machine learning algorithms. Notable approaches include the utilization of Support Vector Machines (SVMs) for sugarcane leaf disease detection and severity estimation based on segmented spot images, as proposed by Ratnasari et al., achieving an accuracy of 80% [7]. Gavhale et al. present an approach for unhealthy region detection in citrus leaves using statistical GLCM and SVM, achieving a recognition rate of 96% with an SVMRBF classifier [8]. Zhang et al. introduce a genetic algorithm-supported SVM (GA-SVM) for maize leaf disease recognition, outperforming traditional SVM with an accuracy of 92.59% [9]. Furthermore, Padol and Yadav employ SVM classifiers for grape leaf disease detection, achieving an accuracy of 88.89% [10]. Parikh et al. propose a cascaded classifier system for disease detection and severity estimation in cotton plants from unconstrained images, achieving an accuracy of 82.5% [11]. Thilagavathi and Abirami explore the application of SIFT features and SVM for diagnosing guava leaf diseases with an accuracy of 84% [12]. Other notable works include Sureesha et al. utilizing the KNN classifier for paddy leaf disease recognition with an accuracy of 76.59% [13]. Finally, Vaishnnave et al. employ the KNN classifier and Random Forest, respectively, for groundnut leaf disease and tomato plant disease detection, achieving commendable accuracies [14]. These studies collectively illustrate the evolving landscape of plant disease detection methodologies, showcasing the advancement from traditional techniques to sophisticated machine learning algorithms.
Several studies have been conducted in the field of plant disease classification and identification using artificial intelligence techniques, particularly deep learning models. In 2016, Ishak et al. utilized an Artificial Neural Network (ANN) to achieve an accuracy of 90.3% in classifying leaf diseases, demonstrating the effectiveness of ANN in this domain [15]. Subsequent research by Grinblat et al. in 2016 introduced the use of convolutional neural networks (CNNs) for plant identification based on vein morphological patterns [16]. Their study demonstrated superior performance compared to traditional machine learning algorithms, achieving accuracies of 93% and 98.8% with different CNN configurations. Ghosal et al. proposed an explainable deep machine vision framework in 2018, utilizing CNNs for plant stress phenotyping [17]. With an accuracy of 94.13%, their framework enabled the identification of various plant stresses, including bacterial blight and nutrient deficiencies. In 2019, Militante and Gerardo employed various deep learning models, including VGGNet, to detect sugarcane diseases, achieving an accuracy of 95.40% [18]. Their study highlighted the efficacy of deep learning in accurately identifying diseased sugarcane plants. Further advancements were made by Sethy et al. in 2020, who combined CNNs (ResNet50) with Support Vector Machines (SVMs) for rice leaf disease identification, achieving an impressive F1 score of 98.38% [19]. Continuing the trend of innovation, Hassan et al. in 2021 explored the use of CNN-based transfer learning models for plant disease detection, achieving remarkable accuracy rates, with EfficientNetB0 reaching 99.56% accuracy [20]. Moreover, recent studies have focused on hybrid approaches, such as the integration of Vision Transformers (ViTs) with CNNs, as demonstrated by Thakur et al., who achieved high accuracies in plant disease classification using a hybrid ViT-CNN model [21].

3. ViT (Vision Transformer)

A Vision Transformer (ViT) is a type of neural network architecture designed for image recognition tasks. The principles of the Transformer model, originally developed for natural language processing (NLP), are applied to the field of computer vision. The Vision Transformer comprises several key components, including patch embedding, positional encoding, the Transformer encoder, and the classification head.
The process begins with patch embedding, where the input image is divided into fixed-size patches. Each patch is flattened and linearly projected into a fixed-dimensional embedding. Formally, given an input image { P 1 , P 2 ,   ,   P n } , each patch P i is embedded into a vector E P i . This transformation converts the 2D image data into a sequence of 1D token embeddings, preparing it for processing by the Transformer model. Positional encoding is then added to these patch embeddings to provide spatial information, ensuring that the model can recognize the relative positions of the patches within the original image.
Central to the ViT is the Transformer encoder, which consists of multiple layers of multi-head self-attention mechanisms and feed-forward networks. Self-attention allows the model to weigh the importance of each patch relative to the others. To achieve this, the mechanism involves three key components such as query Q, key K, and value V vectors. The self-attention mechanism calculates the attention scores by taking the dot product of the query Qi with all keys Kj from other patches, followed by a softmax operation to obtain attention weights. These weights are then used to compute a weighted sum of the value vectors Vj, which represents how much information from each patch should contribute to the current patch. The equation for the self-attention mechanism is as follows:
A t t e n t i o n Q i ,   K j ,   V j = S o f t m a x Q i K j T d k V j
where dk is the dimensionality of the key vectors, used to scale the dot product for numerical stability. This self-attention process captures complex relationships between patches, allowing the model to focus on relevant parts of the image. Each attention layer is followed by a position-wise feed-forward network, which is applied independently to each patch embedding. Residual connections and layer normalization are used to stabilize and improve the training process, ensuring efficient and effective learning of image representations. From this process, we obtain a 768-dimensional vector for each image, which is then exploited in the next stage for further tasks.

4. Attention Score-Based Multi-VIT

The proposed Attention Score-Based Multi-ViT model integrates multiple Vision Transformer (ViT) instances to enhance image classification performance. While ViTs capture global features of images using self-attention mechanisms, a single ViT may find it challenging to fully comprehend the complex characteristics of an image. This makes it difficult to diagnose a tree’s overall disease by examining just one leaf. Relying on individual pieces of information has limitations in grasping the overall state or patterns, so it is important to gather information from various perspectives. Therefore, we combine outputs from multiple pre-trained ViTs to learn diverse representations of the input image, extracting richer features in the process as shown in Figure 1.
Each ViT independently processes the input image and extracts a unique representation hi from the first (classification) token of its output, where i = 1, 2, …, M. These representations capture different features and patterns due to differences in each ViT’s architecture or initialization, allowing the model to understand the image from various perspectives.
To evaluate the importance of each ViT’s output, we introduce learnable attention weight vectors Wi. The attention score ai is calculated through the dot product of the ViT’s representation hi and the attention weight Wi as follows:
a i = h i T W i
The calculated attention scores are then normalized using the softmax function to produce normalized attention weights αi, representing the relative importance of each ViT. The normalization is performed as follows:
a i = e x p a i j = 1 M e x p a j
where M = 5 is the number of ViTs used. The normalized attention weights αi reflect each ViT’s contribution to the final prediction. Using these normalized attention weights, we aggregate the individual ViT representations to form the final representation. The aggregation is performed through a weighted summation, as expressed in the following equation:
H = i = 1 M a i h i
This aggregated representation integrates the diverse features obtained from multiple ViTs, more accurately reflecting the complex characteristics of the input image. To refine this combined representation, we apply a linear transformation layer, which can be mathematically described as
H = L i n e a r H
The refined representation H′ is then passed to the classification head to generate the final logits.
The classification head typically consists of a fully connected layer with a softmax activation function for multi-class classification, allowing the model to output a probability distribution over predefined classes. This approach aims to improve the overall performance and generalization ability of the model by integrating diverse representations from multiple ViTs. By combining information obtained from various perspectives, the model can more effectively handle complex patterns and variations in images, which contributes to more accurate classification.

5. Experimental Results

This section presents a comprehensive quantitative evaluation of the proposed method compared to other models, utilizing metrics such as the F1 score, accuracy, mIoU, and confusion matrix. Information about the dataset and the input methods is also explained.

5.1. Dataset and Input Methods

To thoroughly assess the robustness of the proposed method, we utilized a variety of tree leaf datasets, including those related to apple trees and several other species. Each dataset comprises images of leaves exhibiting different disease conditions as well as healthy states. In the preprocessing stage, five leaf images with the same label were randomly selected from the dataset to train the proposed method. This approach ensures that the model learns accurate representations of each disease class from consistent input data.
During data composition, we further enhanced the dataset by systematically selecting four additional random images based on a sequential criterion ranging from 1 to n. This process was repeated iteratively, ensuring diverse combinations and providing the model with comprehensive learning opportunities. This iterative augmentation, performed over and over, generated a robust dataset, which was subsequently used for model training.
The input images were standardized to match the required format for the models. Specifically, images were resized to a suitable resolution compatible with the ViT architecture and other convolutional neural networks employed in the comparison. Data augmentation techniques, such as rotation, flipping, and scaling, were applied to enhance the diversity of the training data, thereby improving the model’s generalization capabilities.

5.2. Evaluation Metrics

This paper presents a comprehensive quantitative evaluation of the proposed method in comparison to other models, utilizing a range of metrics including the F1 score, accuracy, mean Intersection over Union (mIoU), and confusion matrix. The F1 score serves as a harmonic mean of precision and recall, offering a balanced evaluation when the dataset is imbalanced. It is defined by the equation
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
This metric is particularly useful in contexts where both false positives and false negatives carry significant costs, such as in disease detection tasks. A higher F1 score reflects the model’s ability to accurately balance the trade-offs between precision and recall, ensuring that both types of errors are minimized. Precision is defined as the ratio of true positive predictions to the total number of positive predictions made by the model:
P r e c i s i o n = T P T P + F P
A high precision score indicates that the model is effective at minimizing false positives, making it particularly valuable in scenarios where incorrect positive predictions can lead to costly outcomes. Recall measures the model’s ability to correctly identify actual positive cases:
R e c a l l = T P T P + F N
High recall is essential in tasks where false negatives must be minimized, such as medical diagnoses, where failing to identify a positive case can have serious consequences. Accuracy provides an overall measure of the model’s performance, reflecting the proportion of correct predictions (both true positives and true negatives) across all cases:
A c c u r a c y = T P + T N T P + T N + F P + F N
Although accuracy is a widely used metric, it can be misleading in imbalanced datasets, where the model may achieve high accuracy by predominantly predicting the majority class correctly, while underperforming on the minority class. Therefore, accuracy should be used in conjunction with metrics like precision, recall, and the F1 score to provide a more holistic view of the model’s performance.
mIoU is commonly used in segmentation and object detection tasks to measure the overlap between the predicted region and the ground truth:
I o U = P r e d i c t i o n G r o u n d   T r u t h P r e d i c t i o n G r o u n d   T r u t h ,       m I o U = 1 C C = 1 C I o U C
where C is the number of classes. A higher mIoU indicates a more accurate localization of diseased regions—an essential aspect for plant disease diagnoses.
To further understand the model’s performance, we utilize the confusion matrix, which breaks down the predictions into four categories: true positives (TPs), true negatives (TNs), false positives (FPs), and false negatives (FNs). This matrix helps identify specific areas where the model excels or requires improvement, such as reducing false positives (Type I errors) or false negatives (Type II errors). The confusion matrix also serves as the foundation for calculating precision, recall, and accuracy, offering a deeper insight into the types of errors the model is making and how they impact the overall performance.

5.3. Performance Evaluation

The experimental results reveal several key insights into the effectiveness of the proposed Attention Score-Based Multi-ViT model for plant disease classification, particularly in terms of accuracy and F1 scores. Across all datasets, the proposed model consistently outperforms traditional models, as evidenced by the figures in Table 1 and Table 2, demonstrating its ability to capture complex patterns in leaf images. This suggests that the integration of multiple ViT instances, each contributing diverse feature representations, significantly enhances classification performance. The introduction of attention score mechanisms further refines this process, allowing the model to prioritize the most informative image regions, resulting in fewer misclassifications, which can be observed in the confusion matrices presented in Figure 2.
The proposed model achieved remarkable accuracy across different plant diseases, notably with apple and grape leaf diseases, where it reached accuracies of 99.116% and 99.49%, respectively, as shown in Table 1 and Table 2. These figures surpass the performance of other models, underscoring the robustness of the Multi-ViT approach. ViT, ResNet50, and MobileNetV2, which are commonly used in image classification tasks, consistently underperformed compared to the proposed model, particularly in handling more subtle or diverse disease symptoms, as illustrated by the confusion matrices in Figure 3.
One key observation from these results is the significant improvement in F1 scores, particularly in datasets with overlapping symptoms or complex disease patterns, such as apple and grape leaf diseases. The attention score mechanism enables the model to dynamically weigh the importance of different ViT outputs, leading to more nuanced and precise classification decisions. This is clearly illustrated by the reduced misclassification rates, especially for classes that are traditionally harder to distinguish due to visual similarity, as shown in Figure 2 and Figure 3.
While the proposed model maintained competitive performance in tomato leaf disease classification, with an accuracy of 96.6924% and an F1 score of 0.9672, as demonstrated in Table 3, the results show a slight performance dip compared to its performance in other datasets. This trend likely reflects the complexity of tomato leaf disease symptoms, which might be more varied or less visually distinguishable. Nonetheless, the proposed model still outperformed the ViT and ResNet50 models, which struggled significantly in this task, with ViT achieving only 80.1963% accuracy as indicated in Table 3. The confusion matrices in Figure 4 highlight that, while MobileNetV2 and VGG performed well, the proposed model consistently reduced error rates across classes.
The confusion matrices presented in Figure 2, Figure 3 and Figure 4 further illustrate the superior generalization ability of the proposed model. By aggregating information from multiple ViT outputs, the model successfully minimizes misclassifications, particularly in classes that other models, such as ResNet50 and ViT, tend to confuse. This improvement in handling subtle intra-class variations across multiple leaf images underscores the model’s effectiveness in capturing the fine-grained details necessary for accurate disease diagnoses.
Additional experiments focusing on an ROC-AUC analysis were conducted for multiple models across different datasets as shown in Figure 5. The proposed model achieves the highest averaged ROC-AUC score, demonstrating superior performance in multi-class classification. ViT, ResNet50, and VGG show strong but slightly lower ROC-AUC values.

6. Discussion

The experimental results demonstrate that the proposed Attention Score-Based Multi-ViT model consistently outperforms traditional models, such as ResNet50, MobileNetV2, and standalone Vision Transformers, in plant disease classification tasks. For instance, the proposed model achieved an accuracy of 99.49% and an F1 score of 0.9949 in grape leaf disease classification, significantly surpassing ResNet50, which only achieved an accuracy of 96.73%. However, our model demonstrates additional advantages by effectively handling spatially distributed symptoms through a multi-leaf analysis. The integration of multiple ViTs with a dynamic attention score mechanism addresses the limitations of single-leaf-based diagnoses. By aggregating diverse feature representations, the model captures subtle intra-class variations, enabling more precise classification. This capability is particularly evident in apple leaf disease classification, where overlapping symptoms among classes traditionally cause high misclassification rates in other models. The confusion matrices presented in Figure 2, Figure 3 and Figure 4 illustrate a significant reduction in misclassification rates compared to competing models. From a practical perspective, the proposed model’s ability to process multiple leaves simultaneously enhances its applicability in real-world scenarios. Field conditions often involve complex symptom distributions across various leaves, making single-leaf-based models less effective. By accounting for the spatial distribution of symptoms, the Multi-ViT approach offers a more reliable solution for diagnosing tree diseases in field environments. Nonetheless, certain limitations remain. For example, the model exhibited a slight performance dip in tomato leaf disease classification, likely due to the higher variability in symptom presentation. Additionally, our experiments with YOLOv8 indicate that object detection-based models can also achieve strong performance in plant disease classification. However, the proposed Multi-ViT model demonstrates unique advantages in multi-leaf-based scenarios, offering more effective classification for complex disease patterns. Future work may explore hybrid approaches that combine the strengths of detection-based methods with multi-leaf strategies to further improve diagnosis accuracy. Future work may focus on incorporating environmental and contextual data to further enhance robustness and generalization.

7. Conclusions

In conclusion, the proposed Attention Score-Based Multi-ViT model demonstrates significant advancements in plant disease classification, consistently outperforming traditional models like ViT, ResNet50, VGG, YOLOv8, and MobileNetV2 in terms of accuracy, mIoU, and F1 scores. The integration of multiple ViTs, coupled with a dynamic attention score mechanism, enhances the model’s ability to capture complex patterns and prioritize relevant image features, leading to fewer misclassifications and superior generalization across different datasets. This was particularly evident in the classification of apple and grape leaf diseases, where the model achieved exceptional performance. While the model experienced a slight performance dip in the tomato leaf disease dataset, it still surpassed competing models, highlighting its robustness in handling varied and complex disease symptoms. The results suggest that this Multi-ViT approach offers a scalable and efficient solution for real-world agricultural applications, enabling more accurate and reliable plant disease diagnoses.

Funding

This study was supported by research funding from Honam University, 2024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available in the “New Plant Diseases Dataset” at Kaggle, accessible through https://www.kaggle.com/datasets/vipoooool/new-plant-diseases-dataset.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Naikwadi, S.; Amoda, N. Advances in image processing for detection of plant diseases. Int. J. Appl. Innov. Eng. Manag. 2013, 2, 168–175. [Google Scholar]
  2. Singh, V.; Misra, A.K. Detection of plant leaf diseases using image segmentation and soft computing techniques. Inf. Process. Agric. 2017, 4, 41–49. [Google Scholar] [CrossRef]
  3. Bauer, S.D.; Korč, F.; Förstner, W. The potential of automatic methods of classification to identify leaf diseases from multispectral images. Precis. Agric. 2011, 12, 361–377. [Google Scholar] [CrossRef]
  4. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  5. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  6. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  7. Ratnasari, E.K.; Mentari, M.; Dewi, R.K.; Ginardi, R.H. Sugarcane leaf disease detection and severity estimation based on segmented spots image. In Proceedings of the International Conference on Information, Communication Technology and System (ICTS), Surabaya, Indonesia, 24 September 2014; pp. 93–98. [Google Scholar]
  8. Gavhale, K.R.; Gawande, U.; Hajari, K.O. Unhealthy region of citrus leaf detection using image processing techniques. In Proceedings of the International Conference for Convergence for Technology-2014, Pune, India, 6–8 April 2014. [Google Scholar]
  9. Zhang, Z.Y.; He, X.Y.; Sun, X.H.; Guo, L.M.; Wang, J.H.; Wang, F.S. Image recognition of maize leaf disease based on GA-SVM. Chem. Eng. Trans. 2015, 46, 199–204. [Google Scholar]
  10. Padol, P.B.; Yadav, A.A. SVM classifier based grape leaf disease detection. In Proceedings of the 2016 Conference on Advances in Signal Processing (CASP), Pune, India, 9–11 June 2016. [Google Scholar]
  11. Parikh, A.; Raval, M.S.; Parmar, C.; Chaudhary, S. Disease detection and severity estimation in cotton plant from unconstrained images. In Proceedings of the 2016 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Montreal, QC, Canada, 17–19 October 2016; pp. 594–601. [Google Scholar]
  12. Thilagavathi, M.; Abirami, S. Application of image processing in diagnosing guava leaf diseases. Int. J. Sci. Res. Manag. (IJSRM) 2017, 5, 5927–5933. [Google Scholar]
  13. Suresha, M.; Shreekanth, K.N.; Thirumalesh, B.V. Recognition of diseases in paddy leaves using knn classifier. In Proceedings of the 2017 2nd International Conference for Convergence in Technology (I2CT), Mumbai, India, 7–9 April 2017; pp. 663–666. [Google Scholar]
  14. Vaishnnave, M.P.; Devi, K.S.; Srinivasan, P.; Jothi GA, P. Detection and classification of groundnut leaf diseases using KNN classifier. In Proceedings of the 2019 IEEE International Conference on System, Computation, Automation and Networking (ICSCAN), Pondicherry, India, 29–30 March 2019; pp. 1–5. [Google Scholar]
  15. Ishak, S.; Rahiman MH, F.; Kanafiah SN, A.M.; Saad, H. Leaf disease classification using artificial neural network. J. Teknol. 2015, 77, 109–114. [Google Scholar] [CrossRef]
  16. Grinblat, G.L.; Uzal, L.C.; Larese, M.G.; Granitto, P.M. Deep learning for plant identification using vein morphological patterns. Comput. Electron. Agric. 2016, 127, 418–424. [Google Scholar] [CrossRef]
  17. Ghosal, S.; Blystone, D.; Singh, A.K.; Ganapathysubramanian, B.; Singh, A.; Sarkar, S. An explainable deep machine vision framework for plant stress phenotyping. Proc. Natl. Acad. Sci. USA 2018, 115, 4613–4618. [Google Scholar] [CrossRef]
  18. Militante, S.V.; Gerardo, B.D. Detecting sugarcane diseases through adaptive deep learning models of convolutional neural network. In Proceedings of the 2019 IEEE 6th International Conference on Engineering Technologies and Applied Sciences (ICETAS), Kuala Lumpur, Malaysia, 20–21 December 2019; pp. 1–5. [Google Scholar]
  19. Sethy, P.K.; Barpanda, N.K.; Rath, A.K.; Behera, S.K. Deep feature based rice leaf disease identification using support vector machine. Comput. Electron. Agric. 2020, 175, 105527. [Google Scholar] [CrossRef]
  20. Hassan, S.M.; Maji, A.K.; Jasiński, M.; Leonowicz, Z.; Jasińska, E. Identification of plant-leaf diseases using CNN and transfer-learning approach. Electronics 2021, 10, 1388. [Google Scholar] [CrossRef]
  21. Thakur, P.S.; Chaturvedi, S.; Khanna, P.; Sheorey, T.; Ojha, A. Vision transformer meets convolutional neural network for plant disease classification. Ecol. Inform. 2023, 77, 102245. [Google Scholar] [CrossRef]
  22. Dosovitskiy, A. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  23. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  24. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–20 June 2018; pp. 4510–4520. [Google Scholar]
  25. Reis, D.; Kupec, J.; Hong, J.; Daoudi, A. (). Real-time flying object detection with YOLOv8. arXiv 2023, arXiv:2305.09972. [Google Scholar]
Figure 1. Model architecture for Attention Score-Based Multi-ViT for plant disease classification.
Figure 1. Model architecture for Attention Score-Based Multi-ViT for plant disease classification.
Sensors 25 00270 g001
Figure 2. Confusion matrix for apple leaf disease classification across models. (a) Proposed; (b) VIT; (c) Resnet50; (d) VGG; (e) Mobilenet; (f) YOLOv8.
Figure 2. Confusion matrix for apple leaf disease classification across models. (a) Proposed; (b) VIT; (c) Resnet50; (d) VGG; (e) Mobilenet; (f) YOLOv8.
Sensors 25 00270 g002
Figure 3. Confusion matrix for grape leaf disease classification across models. (a) Proposed; (b) VIT; (c) Resnet50; (d) VGG; (e) Mobilenet; (f) YOLOv8.
Figure 3. Confusion matrix for grape leaf disease classification across models. (a) Proposed; (b) VIT; (c) Resnet50; (d) VGG; (e) Mobilenet; (f) YOLOv8.
Sensors 25 00270 g003
Figure 4. Confusion matrix for tomato leaf disease classification across models. (a) Proposed; (b) VIT; (c) Resnet50; (d) VGG; (e) Mobilenet; (f) YOLOv8.
Figure 4. Confusion matrix for tomato leaf disease classification across models. (a) Proposed; (b) VIT; (c) Resnet50; (d) VGG; (e) Mobilenet; (f) YOLOv8.
Sensors 25 00270 g004
Figure 5. Comparative ROC-AUC curves for apple, grape, and tomato leaf disease classification across multiple models. (a) Apple leaf disease; (b) grape leaf disease; (c) tomato leaf disease.
Figure 5. Comparative ROC-AUC curves for apple, grape, and tomato leaf disease classification across multiple models. (a) Apple leaf disease; (b) grape leaf disease; (c) tomato leaf disease.
Sensors 25 00270 g005
Table 1. Performance Metrics of Different Models for Apple Leaf Disease Classification.
Table 1. Performance Metrics of Different Models for Apple Leaf Disease Classification.
MethodProposedVIT [22]Resnet50 [5]VGG [23]Mobilenet [24]YOLOv8 [25]
F1 score0.99110.97530.96900.95200.98500.99
Accuracy99.116097.524296.912095.213698.50750.961
mIoU0.9822270.9542140.943770.8994570.9677280.981025
Table 2. Performance Metrics of Different Models for Grape Leaf Disease Classification.
Table 2. Performance Metrics of Different Models for Grape Leaf Disease Classification.
MethodProposedVITResnet50VGGMobilenetYOLOv8
F1 score0.99490.98220.96720.94960.98220.99
Accuracy99.4998.227196.732594.958598.227199.2
mIoU0.98100.93600.9370.9040.94250.9373
Table 3. Performance Metrics of Different Models for Tomato Leaf Disease Classification.
Table 3. Performance Metrics of Different Models for Tomato Leaf Disease Classification.
MethodProposedVITYOLOv8Resnet50VGGMobilenet
F1 score0.96720.79990.940.92210.94960.9283
Accuracy96.692480.196393.592.213794.958592.8899
mIoU0.9742 0.9647 0.9513 0.9728 0.9671 0.9565
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Baek, E.-T. Attention Score-Based Multi-Vision Transformer Technique for Plant Disease Classification. Sensors 2025, 25, 270. https://doi.org/10.3390/s25010270

AMA Style

Baek E-T. Attention Score-Based Multi-Vision Transformer Technique for Plant Disease Classification. Sensors. 2025; 25(1):270. https://doi.org/10.3390/s25010270

Chicago/Turabian Style

Baek, Eu-Tteum. 2025. "Attention Score-Based Multi-Vision Transformer Technique for Plant Disease Classification" Sensors 25, no. 1: 270. https://doi.org/10.3390/s25010270

APA Style

Baek, E.-T. (2025). Attention Score-Based Multi-Vision Transformer Technique for Plant Disease Classification. Sensors, 25(1), 270. https://doi.org/10.3390/s25010270

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop