Introduction

Nowadays, agriculture serves as a cornerstone for addressing the nutritional needs of an expanding population. Agriculture, fisheries, and forestry sectors contribute approximately 18% to the GDP. However, the share of agricultural output in the national GDP has been steadily declining over the years1. More than three billion people depend on rice as their primary source of sustenance. However, the term “rice” is quite general, as there are many different varieties cultivated across the globe2. As a Kharif crop, rice grows well in warm, humid climates and waterlogged conditions, which makes it vulnerable to various diseases. These diseases can be divided into two main categories according to their causes.

The first type is parasitic diseases, triggered by living organisms such as harmful organisms, pests, and weeds. Pathogens like fungi, bacteria, and viruses are responsible for a broad spectrum of infections in rice. Some of these diseases can significantly reduce crop yields. The second type is non-parasitic diseases, caused by environmental factors like extreme temperatures, radiation, nutrient deficiencies, or water-related issues. Such abiotic diseases include straight head, cold injury, white tip, panicle blight, alkalinity, and bronzing. The main problem is the need for regular plant monitoring. Additional factors could involve farmers’ inexperience, who may overlook important details or need more knowledge about diseases that impact plants and their seasonal patterns. These diseases can strike any time, but consistently tracking plant growth and development stages can help control their spread. Numerous pathogens, including viruses, bacteria, and fungi pose a substantial risk to rice production by causing illnesses like sheath rot, bacterial leaf blight, leaf folder, blast disease, sheath blight, brown leaf spot, and hispa3,4. Implementing corrective actions, such as applying fungicides in the early stages, can help prevent significant losses5. Without proper identification, applying pesticides becomes ineffective. This issue has inspired us to pursue research on identifying rice diseases prevalent in India. Rice plant infections may be categorized according to the origin of the infection or the specific plant part impacted. Symptoms may appear on the rice plant’s leaves, stems, panicles, or roots. These infections can lead to significant crop losses, with damage ranging from 30 to 100% of the yield6.

Detection of plant diseases at an early stage requires arduous effort7. Traditionally, farmers rely on manual methods, using their knowledge and visual observations to detect diseases in rice crops. However, this approach could be more efficient, time-consuming, and prone to errors8. It is even more crucial for experts to recognize the cumbersome disease and quantify it in the case of large paddy fields. Technological advancements, like ICT and IoT, have enabled the integration of real-time monitoring and disease detection, classification, and ramifications to a greater extent9,10. To assist this, ML and DL techniques have enabled a wide-open platform to detect, classify, and achieve accurate and timely diagnosis, especially by using DCNNs11,12. Though DCNN can handle large quantities of data, the model’s performance has been expanded even further using transfer learning architectures13. Feature extraction techniques help to identify various patterns, groups, and essential features that contain raw information and provide an understanding of disease detection and classification14. Existing methods based on C-Bi-LSTM suffer from less accuracy and suffers from unbalanced distribution of data15, RNN-LSTM16 suffers from low accuracy due to concealment of CNN methods, barouta feature selection algorithm17 suffers from high computation cost due to big data. Though different ML, DL, and optimization techniques are utilized, these techniques still suffer from vanishing gradients and accidental errors while training the model. Most optimization algorithms pose a limitation by converging to the local optima. Additionally, due to the high amount of data, the computational complexity leads to high computational time. This highlights the necessity of developing a novel automated rice foliage disease detection and categorization framework utilizing deep learning. Such a system would aid farmers by offering a cost-effective, lightweight computational model that delivers optimal performance.

Motivation

The motivation behind detecting multiple diseases on the same leaf of paddy plants using deep learning stems from the critical need to enhance agricultural disease management. It is identified that only 20–30% of farmers in India have the knowledge and access to disease detection and classification methods. Traditional methods need help to accurately identify and differentiate between multiple coexisting diseases, leading to misdiagnosis, ineffective treatments, and significant crop losses. This provokes the need to design a new automated paddy leaf illness detection and classification method powered by deep learning, which would provide optimized performance and better adoption, thereby incrementing the productivity of the paddy crop, benefitting the farmers.

Major contributions

The proposed automated hybrid system utilizes deep learning techniques for detecting and classifying paddy leaf diseases. It involves the use of the hybrid MobileNetV318, a transfer learning CNN model, and GKSO-SA19, followed by classification using the CatBoost classifier20. The proposed disease detection and classification model involves the following significant contributions:

  • Farmers lack knowledge and awareness of modern technologies to detect and recognize paddy leaf diseases. To assist them with a new automated hybrid MobileNetV3, a transfer learning CNN model along with GKSO-SA is utilized. This alleviates the need to wait for an expert, thereby allowing the farmers themselves to assess the corresponding paddy leaf diseases.

  • The proposed work involves the extraction of RoIs, data labelling, resizing, filtering, augmentation, and segmentation using adaptive thresholding and k-means clustering techniques for image pre-processing.

  • The features are then classified using the CatBoost classifier, which achieves the finest classification of all features, thereby liberating the selected features. The significant advantage of the CatBoost classifier is that it analyzes the image data with categorical labels, thereby providing a clear-cut analysis of the relationship between the features.

  • An analysis of performance was conducted on the Google Colab framework using Python, demonstrating superior results over existing methods for detecting and classifying paddy leaf diseases, with an accuracy of 98.52%, surpassing traditional algorithms.

Organization of the research work

The proposed work begins with Section. 1, which provides the introduction, motivation, and outlines the significant contributions towards the detection and categorization of paddy leaf diseases. This is followed by section. 2 "Related works" (related work), which reviews various studies conducted by researchers on paddy leaf disease detection and classification framework along with their solutions, benefits, and limitations. Section. 3 "Materials and methods" covers the materials and methods, offering a detailed discussion that leads into section. 4 "Proposed methodology", which presents the proposed methodology. Next, section. 5 "Experimentation" presents the experimental analysis. Section. 6 "Result analysis and discussion" discusses the results and discussion, and finally, section. 7 "Conclusion and future work" provides the conclusion and outlines direction for future work. Table 1 provides the chronicle of abbreviations utilized in the proposed work.

Table 1 Chronicles of abbreviations.

Related works

In agriculture, the application of image processing and pattern recognition tools to detect and diagnose plant diseases is growing. Computer vision systems, in particular, have played a key role in creating tools capable of accurately diagnosing and detecting disease symptoms in crops such as paddy. Recent developments in deep learning have empowered researchers to make significant strides in recognizing and detecting diseases and pests on plant leaves from images. These innovations assist farmers in identifying and managing diseases early, which helps to minimize crop losses and enhance overall productivity.

Bola et al.21 introduced a hybrid model built on traditional ML and DL techniques for identifying diseases of multiple crops, specifically in corn, rice, and wheat. This study compared the effectiveness of various methods to recognize illnesses in wheat, corn, and rice. As a result, the combination of DenseNet201 deep features with an SVM exhibited outstanding performance, reaching an exceptional accuracy of 87.23% with only 20.2 million parameters. The model was rigorously evaluated utilizing precision, recall, F1 score, and total accuracy metrics. It efficiently distinguishes between healthy and infected plants for each crop, attaining accuracy rates of 99.82% for corn, 98.75% for wheat, and 84.15% for rice.

Furthermore, integrating real-time data from crop fields with images can significantly improve the accuracy of disease classification and identification. Omer et al.22 suggested an algorithm for detecting diseases and pests on cucumber leaves using an enhanced YOLO v5 model, where the CSP module in the backbone was replaced with the CBAM. They developed a dataset consisting of 3,057 real-world images of cucumber leaf diseases. The modified method attained a superior appreciation accuracy of 80.10%, outperforming the original YOLO v5. Future work will refine the network model structure to improve cucumber leaf disease and pest detection performance. Dubey et al.23 designed a deep-learning model for the automated detection and classification of leaf conditions. Feature selection was initially carried out by combining SV-RFE and ARFA. The selected features were then used in an MBi-LSTM classifier to categorize the visuals as either Tungro, Bacterial Leaf Blight, Blast disease, or healthy. The proposed feature selection approach outperformed individual models, achieving a high accuracy of 97.16% based on various classification metrics compared to state-of-the-art techniques. However, this method is limited to diagnosing specific crop diseases. Elfatimi et al.24 projected a CNN method for detecting and preventing ten common tomato leaf diseases using DMCNN. They constructed a dataset of 11,000 images of tomato leaf diseases sourced from Kaggle. The recommended model attained a high accuracy of 99.1%, demonstrating its effectiveness in the early detection and prevention of plant leaf diseases. Future work aims to apply the model to other crops and explore transfer learning, although its practical implementation in precision agriculture has yet to be tested.

Wang et al.25 introduced a CNN model to classify seven common apple leaf diseases using an optimized, lightweight RegNet model. This model utilized a dataset of 2,609 images of apple leaf diseases sourced from Kaggle. The model demonstrated good generalization for Dataset1 and Dataset2, achieving testing accuracies of 93.85% and 99.23%, respectively. The model tested on Dataset 1 attained an accuracy of 93.85%, with 24 samples misclassified. Despite efforts, the model needed help to improve the testing accuracy. RegNet-trained models demonstrated superior robustness and accuracy, identifying various diseases in leaf images despite complex visual interference. Stephen et al.26 analyzed a new method utilizing 3D and 2D Convolutional Neural Networks for feature extraction in predicting rice leaf disease. They collected 5,200 images of rice leaf diseases to address the problem faced by farmers, who typically rely on visual examination, which is susceptible to errors, labor-intensive, and can result in significant yield loss due to misjudgment in detecting three varieties of paddy leaf diseases. The concept of rice leaf prediction was implemented using Google Colab with the IBS-optimized DGAN. The proposed IBSGAN architecture model accomplished a high accuracy of 98.7%, outperforming other established methods like XGBoost, transfer learning approaches, and support vector machines. The advantage of the IBS algorithm is that it primarily addresses instability and overfitting problems commonly encountered in DGAN.

Future studies will examine a broader range of leaf characteristics and address additional diseases in rice plants. Singh et al.27 discussed the automatic identification of fungal blast leaf sickness in paddy crops. The proposed approach utilized five deep transfer learning algorithms based on convolutional neural networks for binary classification, namely InceptionV1, AlexNet, Xception, LeNet, and VGG-16 models. The algorithm performance evaluation showed that AlexNet achieved the highest accuracy for binary classification, with an average of 98.7%. The evaluation parameters included accuracy, sensitivity, specificity, F1 score, FDR, and FOR. A dataset containing 6,300 images of rice leaves affected by fungal blast disease was gathered in real time. The model’s scope could be expanded to focus on more diseases in the same crop and could also be applied to detect diseases in other crops and fruit plants. Integrating it with IoT, computer vision, and robotics could improve its effectiveness. Dubey and Choubey28 explored a method for classifying rice leaf diseases, employing a highly efficient adaptive feature selection strategy.

This approach involves the creation of a model by the authors to identify four distinct diseases affecting the leaves of paddy plants: blast disease, bacterial leaf blight, Tungro, or healthy for image acquisition. The model consists of four stages: pre-processing, feature extraction, selection, and classification. With this approach, the input image was directed to the pre-processing stage, where it was converted to RGB color, and the noise was reduced using a median filter. The results showed that, compared to other state-of-the-art models, the proposed deep learning technique achieved an accuracy rate of 98.86%. Dogra et al.29 suggested a model for detecting rice leaf diseases using automated CNNs. Their approach specifically focused on identifying brown spot disease in the context of intelligent agriculture. The study contrasted the traditional, time-consuming manual process of disease identification with an automated system utilizing deep learning CNN models. The automated approach reached the highest accuracy. Four key metrics, sensitivity, specificity, precision, and F1-score, were evaluated, with VGG-19 achieving the best accuracy of 93.0%. Transfer learning methodologies were employed to improve prediction accuracy for more diverse datasets and additional rice leaf diseases, which enhanced accuracy and simplified the model training process. Bharanidharan et al.30 developed a machine-learning approach to categorize five types of paddy diseases: bacterial leaf blight, hispa, leaf folder, brown leaf spot, and rice blast. The model detects paddy leaf diseases using a filter-based feature transformation method, utilizing a dataset of 636 images. Each thermal image was used to extract seven statistical features and an additional seven Box-Cox transformed statistical attributes. Four machine learning techniques were tested: KNN classifier, Random Forest classifier, LDA classifier, and HGB classifier. By applying the suggested feature transformation method, the K-Nearest Neighbor classifier achieved the highest accuracy, reaching 90%.

Latif et al.31 developed a deep learning-based approach, modifying a VGG19 transfer learning method to accurately classify and diagnose six categories, such as healthy rice plant leaves and five diseases: bacterial leaf blight, brown spots, leaf blasts, leaf scalds, and narrow brown spots. The system was trained on a non-normalized augmented dataset, achieving a high accuracy of 96.08%. Future work will concentrate on developing an IoT system integrated with deep learning and powered by drones, to be tested under real-world, real-time conditions. Additionally, the aim is to improve deep learning approaches for identifying all rice leaf diseases and expand the study to cover other significant plant diseases in the agricultural sector. Sethy et al.32 presented a system that used 5,932 field images representing four categories of rice leaf diseases: blast, bacterial blight, tungro, and brown spot. The system incorporated an evaluation of the performance of 11 CNN models, utilizing a transfer learning technique along with deep feature extraction and SVM for classification. In this method, ResNet50 and SVM demonstrated the highest effectiveness as a classification model, attaining an F1-score of 0.9838 for deep feature-based identification of diseases in rice leaves. The scope of this research could be expanded by incorporating a broader range of rice diseases and further optimizing the CNN model for enhanced performance.

Additionally, future work will focus on developing an integrated application suitable for low-end devices. A limitation of this study is the potential need for further refinement to ensure the model’s robustness across various environments and device constraints. Debnath and Saha33 examined detecting rice leaf diseases using real-time image data, focusing on IoT-based smart farming with a CNN model for detecting and categorizing diseases. A primary advancement of their developed model lies in its capacity for early detection of Brown Spot disease in rice paddies, using a tailored image-processing tool for pre-processing and feature extraction. The model was trained on a dataset manually collected from rice fields and achieved a high accuracy of 97.701%. This approach could be extended to identify multiple disease types and assess the severity of crop infestations. Lamba et al.34 proposed the categorization of rice plant leaf diseases, including bacterial blight, leaf smut, and rice blast, based on machine learning approaches, which include GCLSTM networks, CNN, LSTM, and GAN. While the current research has achieved 97% accuracy in identifying these diseases, there still needs to be a gap in detecting disease severity levels. This study is restricted to disease identification without considering the varying degrees of severity for the three paddy diseases under investigation. Future research could focus on developing methods to classify the severity of these diseases.

Upadhyay and Gupta35 introduced a modified ResNeXt-based deep learning model for detecting fungal diseases across multiple fruit crops apple, custard apple, and guava collected from diverse regions. The approach outperformed prior methods such as Inception-v3 and ResNet, achieving 98.92% accuracy using a heterogeneous dataset of 14,408 images. Although the model excels in disease identification, it does not address the classification of disease severity levels. Future research should aim to incorporate severity estimation to support better decision-making in crop management. Upadhyay and Gupta36 presented a meta-learning-based approach for estimating mango crop maturity using few-shot learning and DenseNet-121 architecture. The framework integrates Faster R-CNN for mango segmentation, followed by feature extraction and classification using cosine similarity. Their method achieved 83.65% accuracy in a 5-shot classification setting, outperforming conventional models like MAML and D-CNN. Despite its effectiveness, the study focused solely on binary classification without assessing gradual maturity stages. Future research could explore multi-level maturity estimation to support more nuanced harvesting strategies. Upadhyay and Gupta37 proposed a CNN-based framework incorporating data augmentation and VGG19 for detecting potato leaf diseases such as early and late blight. The model achieved a classification accuracy of 99.2% on a dataset of 2,152 images. Multiple classifiers were evaluated, with logistic regression outperforming others when combined with VGG19-extracted features. While the approach is highly effective in identifying diseases, it focuses only on classification and does not address the varying severity levels of infection. Future studies could explore severity estimation to support more precise treatment planning. Upadhyay and Gupta38 developed an enhanced ResNeXt-based deep learning model for diagnosing fungal diseases in apple crops. Using a benchmark dataset of 9,395 images spanning four disease categories (scab, rot, rust, and healthy), the model incorporated preprocessing, segmentation with GrabCut, and a modified ResNeXt architecture with reduced kernel sizes. The approach achieved 98.94% accuracy, outperforming standard ResNet and Inception-v3. While the study demonstrated high precision in disease identification, it was limited to classification and did not address severity assessment. Future studies could expand to quantify disease severity and adapt the model to other crop types. Upadhyay and Gupta39 conducted a comprehensive survey on the use of artificial intelligence in agricultural crop disease detection, reviewing over a decade of literature. The study examined a wide range of AI models, including SVM, ANN, CNN, and hybrid approaches, across multiple crops such as rice, tomato, maize, and apple. While the survey highlighted significant advancements in classification accuracy some exceeding 99% it also noted a research gap in severity estimation and real-time implementation. Future work could focus on integrating AI techniques for early-stage severity assessment and deploying scalable, field-ready solutions. Table 2 presents a comparison of different studies on paddy leaf diseases.

Table 2 Comparative evaluation of different research works pertaining to paddy leaf disease.

Table 2 shows that the existing methods suffer from high computational costs, and complex models are heavy, leading to decreased efficiency. To alleviate this problem, the proposed model focused on lightweight computations that would effectively benefit the farmers in performing disease detection and classification.

Materials and methods

The proposed automated paddy leaf disease detection and classification model involves various stages: image acquisition, image labelling, image augmentation, image segmentation, image feature extraction, image feature selection, and image feature classification. The doctor paddy dataset40 is utilized to carry out the proposed work. The dataset contains diseases, namely bacterial leaf blight41, bacterial leaf streak42, bacterial panicle blight43, blast44, brown spot45, dead heart46, downy mildew47, hispa48, and tungro49 where the research work has been focused on. Figure 1 provides the working methodology of the proposed new automated deep learning model for paddy leaf disease detection and classification.

Fig. 1
figure 1

Proposed methodology.

The proposed work has been carried out by segregating the image data into 80:10:10 and 70:15:15. In the first phase, 80% of images are used for train the model, and the desired features are selected for classification by the classifier. The remaining untrained pictures of 10% are fed into the model during the testing and the remaining 10% for validation. In the second phase, 70% of images are used to train the model, and the desired features are obtained. They are fed into the classifier for classification. For testing, 15% of unknown images are tested, and the other 15% are utilized for validation. Each stage involves various processing methods, which are explained in detail as follows:

Image data acquisition Data collection is crucial in hoarding information from distinct sources50. Data being acquired enables businesses to gain meaningful insights utilizing ML models. The system will be trained and tested on thousands of images using the CV-based models. The AI model will utilize the dataset for classification and recognition51. Image Data Augmentation Image data enhancement is performed to prevent overfitting. Also, the issue of class imbalance or smaller datasets can be eradicated with the help of data augmentation. The proposed work involves data enhancement operations on images, such as flipping, cropping, rotation, geometric alterations, and colour enhancement52. Image Data Labeling Image data labelling involves tagging images for multi-class classification. Image data labelling is done for multi-class classification and bounding boxes53. Image Pre-processing The image collection should be pre-processed to eliminate noise, blurring, and low contrast has to be carried out54. Image Segmentation To improve the underlying deep learning model’s performance and computation complexity, the images are segmented. The proposed work involves various image segmentation approaches based on regions, edges, and clusters55. Image Feature Extraction Features are the pixels in the form of integer, float, or binary representation of numbers in an image. These pixels carry specialized information that is used for classification and recognition. The proposed work involves feature extraction based on color, text, and shape56,57. Image Feature Classification The categorization of image features is referred to as supervised classification, as the image data is labeled in the proposed study. The underlying deep learning model gets trained and tested with images, yielding the final classified output58.

Image acquisition

The proposed work utilizes the Paddy Doctor dataset of 13,866 diseased and normal paddy images obtained from the Kaggle repository. A high-end GPU machine has been employed to train and test the deep learning model to process vast images. DCNN has been utilized to carry out image classification and recognition since it reduces the problem of dimensionality and offers high accuracy and faster computations. The training and testing have been conducted by splitting and distributing the images into epochs to optimize the effectiveness of the deep learning model. The virtual machine was created using Microsoft Azure based on Windows 11 pro 24H2 and supported by NVIDIA Tesla’s V100 GPU and virtual CPUs with 128 GB RAM59,60. (i) Selection of RoIs The disease-affected regions are cropped at the infected areas of an image in freehand to extract the RoIs. Cropping of the infected areas of an image facilitates better focus, achieving high efficiency, thereby enabling the model to concentrate on irrelevant pixels61. (ii) Data labeling The proposed work encompasses Azure Machine Learning Studio, which provides various data labelling tools with ML-assisted data labelling capability. Data labels are created using ML models that run in the background to speed up the process. In the case of the DL, the transfer learning technique jump-starts the training process by leveraging a pre-trained model where the data gets labelled62,63. Figure 2. provides the diseases in the paddy doctor dataset chosen for the proposed methodology.

Fig. 2
figure 2

Diseases in paddy crops from the paddy doctor dataset.

Image pre-processing

Image pre-processing involves the following stages: image resizing, filtering, contrast enhancement, augmentation, segmentation, thresholding, edge detection, and color-based K-means clustering. (i) Image resizing The paddy images are adjusted to a uniform size, with a resolution of 1500 × 1500 pixels. Resizing the images helps minimize the memory storage requirements64. (ii) Image filtering Speckle noise is an inherent part of an image that is supposed to be removed for feature extraction. A wiener-frost filter is utilized to preserve the edge information and the feature retention65,66,67. The Wiener filter eliminates the noise with specks, where the speckling noise gets removed. Exponential weighting is utilized in a frost filter, which removes the additive noise. Thereby, the texture data of the image gets retained. (iii) Color Contrast Enhancement The primary purpose of performing contrast enhancement is to identify an image’s normal and diseased areas. Contrast enhancement takes the input as a grey scale and provides the output as a histogram-enhanced image68. Figure 3. provides the transformation of the image after performing image contrast enhancement operations. Figure 4. provides the transformation of the images during image pre-processing.

Fig. 3
figure 3

Image contrast enhancement operation.

Fig. 4
figure 4

Image pre-processing operation.

Image augmentation

Overfitting and underfitting significantly impact the performance of the underlying deep-learning technique51. Therefore, image augmentation has been performed to alleviate this. Although image augmentation expands the dataset size, it supports deep learning during the network training. The technique balances data by creating alternate versions of the same image52. Some of the image augmentation techniques employed in the proposed methodology include zooming, rotation, cropping, flipping, sheering, and enhancing the brightness of the pixels in an image69. Figure 5. provides the output of the various augmentation operations performed on a single image.

Fig. 5
figure 5

Augmentation operations.

Image segmentation

Image segmentation is the process of grouping different pixels that contain important information70. The primary objective of image segmentation is to understand the visual information carried by the pixels and to achieve uniformity71. In the proposed methodology, three distinct methods are performed for segmenting the images: adaptive thresholding72, Prewitt edge detection73, and color fuzzy k-means clustering74.

Adaptive thresholding

Thresholding has been carried out to differentiate distinct objects in an image. Adaptive thresholding identifies the smaller regions of an image and ascertains the threshold for each area. Figure 6. provides the output obtained after adaptive thresholding.

Fig. 6
figure 6

Adaptive thresholding operation.

Prewitt edge detection method

Prewitt edge detection is utilized to identify the edge that preserves essential information. Convolution kernels are employed to calculate the gradient magnitude of the pixels in the image75. Prewitt utilizes horizontal and vertical kernels along directions to compute gradient magnitude76. Figure 7. provides the output obtained after prewitt edge detection.

Fig. 7
figure 7

Prewitt edge detection operation.

Color fuzzy k-means clustering

Clustering involves grouping objects (pixels) in an image. It segregates the pixels that share similar or distinct properties and forms clusters. K-means Clustering works based on a group of randomly chosen centroids and finds an optimal computation for the placement of the centroids. The major disadvantage is that the effectiveness of clustering is directly proportional to the k-value, which is challenging77.

Figure 9 provides the output obtained after color fuzzy k-means clustering. Figure 8 shows the image clusters of distinct objects in violet color. The proposed color fuzzy k-means clustering technique yields good performance by segregating the pixels that carry information about both the foreground and background. The morphological activity over the image yields better segmentation precision with high computation speed78.

Fig. 8
figure 8

Color Fuzzy K-means clustering.

Proposed methodology

The proposed methodology adopts a structured approach to detect and classify paddy leaf diseases using a hybrid deep learning model. It combines efficient feature selection with advanced classification methods to enhance overall accuracy.

Proposed hybrid deep learning model with optimization algorithm

The proposed automated framework for paddy leaf disease detection and classification begins with preprocessing a Kaggle dataset comprising 13,866 images, covering nine disease categories and one healthy class. To enhance image quality and feature distinctiveness, images were initially resized to 1500 × 1500 pixels, denoised using the Wiener-Frost filter, and further processed through contrast enhancement and segmentation techniques such as adaptive thresholding, Prewitt edge detection, and fuzzy K-means clustering.

Feature extraction was performed using MobileNetV379, a lightweight deep convolutional neural network based on transfer learning. Due to its efficient architecture and suitability for edge devices, MobileNetV3 effectively extracts features with reduced computational cost. Input images were resized to 224 × 224 pixels, and the model generated 128-dimensional feature vectors using depthwise separable convolution layers with ReLU and h-swish activation functions.

To refine the extracted features, a hybrid feature selection technique combining Genghis Khan Shark Optimization (GKSO)80 and Simulated Annealing (SA)81 was applied. GKSO mimics intelligent foraging behaviors to explore the feature space locally, while SA enhances global exploration by probabilistically accepting suboptimal solutions, thereby avoiding local optima. This complementary strategy enables adaptive and efficient selection of informative features.

The optimized feature subset was subsequently classified using the CatBoost algorithm82, an efficient gradient boosting framework on decision trees. It was configured with 1000 iterations, a tree depth of 6, a learning rate of 0.1, Logloss as the objective function, and a verbosity interval of 100. The model was trained for both binary (healthy/diseased) and multi-class (individual diseases) classification tasks.

All model components were fine-tuned using a grid search strategy. Specifically, MobileNetV3 was optimized with a batch size of 32, a learning rate of 0.0001, and the Adam optimizer, while GKSO-SA was set with a population size of 30, 1000 iterations, and a cooling rate of 0.95. The integrated MobileNetV3 + GKSO-SA + CatBoost model achieved a classification accuracy of 98.52%, demonstrating superior performance compared to conventional machine learning and deep learning models83. The proposed hybrid deep learning model with an optimization algorithm, as described in the proposed methodology, is depicted in Fig. 9.

Fig. 9
figure 9

Proposed hybrid deep learning technique.

Feature extraction using MobileNetV3

To extract features for the proposed paddy disease classification, MobileNetV3, a deep learning algorithm based on transfer learning, has been used84. Leveraging the pre-trained weights on the Paddy Doctor dataset, the proposed model takes the image size of 224 × 224 and outputs a feature vector of size 128, which acts as an input to the feature selection process. The top layers of the mobilenetV3 have only been trained superseding by 1 × 1 point-wise convolution layers for feature extraction. The weights were adjusted during the fine-tuning of the network, consisting of two 2D convolution layers, 16 bottleneck layers, and an average pooling layer, depicted in Fig. 10.

Fig. 10
figure 10

Overview of MobileNetV3 architecture.

The learned features are saved as vectors, representing the clustered input images. The inverted residual block includes a depth-wise convolution layer, serving as the core component of the training model. The depth-wise separable convolution layer includes a squeeze-and-excitation block to extract essential features on a per-channel basis. The depth-wise convolution layer utilized for feature extraction is 3 × 3. It supersedes the 1 × 1 point-wise convolution, which reduces the computational overhead while training the network in case of surplus images. To transform the output into a linear combination, 1 × 1 point-wise convolution was applied to each channel in the inverted residual block. This resembles that of an MLP. Figure 11. provides various components utilized inside the depth-wise separable convolution layer.

Fig. 11
figure 11

Components of depth-wise separable convolution layer.

Two distinct activation functions, ReLu and h-swish85,86, are utilized, which significantly reduces the computational overhead incurred. The formula for the h-swish activation function is provided in Eq. (1) as follows.

$$h-swish=y*\sigma \left(y\right); \sigma \left(y\right)=\frac{ReLu 6\left(y+3\right)}{6}$$
(1)

From the Eq. (1) \(\sigma \left(y\right)\) denotes the piece-wise linear hard analog function. The major advantage of using MobileNetV3 is the reduction in the representation space.

Feature selection using GKSO-SA

GKSO is a population-based optimization algorithm composed of four stages: hunting, movement, foraging, and self-defense87. Figure 12. depicts the process involved in GKSO.

Fig. 12
figure 12

GKSO exploration & exploitation behaviour.

Mathematical formulation-hunting stage

Genghis khan type sharks conduct a broad space search near the unknown marine seashore. In the starting phase, the predator sharks travel in various directions till the position of the optimal food (prey) is located.

A random position near the prey is chosen by the shark which can be determined by computing the upper limits and lower limits. The superseding positions of the sharks are updated as per the following Eq. (2) as follows:

$${{Sh}_{i}}^{j+1}\left(t+1\right)={{Sh}_{i}}^{j}+\frac{{Lowerb}_{j}+rand*({Upperb}_{j}-{Lowerb}_{j})}{iter};\text{ i}=\text{1,2},3\dots \text{N};\text{ j}=\text{1,2},3\dots .\text{D};\text{ iter}=\text{1,2},3,\dots \text{T}$$
(2)

Movement stage

Once the prey is found the shark based on its olfactory senses tries to get closer enough to catch the prey. This movement of sharks can be defined as per the following Eq. (3) as follows:

$${{\widehat{Sh}}_{i}}^{j}\left(t+1\right)=\rho *\left({{sh}_{best}}^{j}\left(t\right)-{{sh}_{i}}^{j}\left(t\right)\right)$$
(3)
$$\rho =\mu {I}^{r}$$
(4)

From the Eq. (4), the value \(\mu\) is mainly responsible for the convergence rate of the algorithm which is a positive constant. To find the optimal solution the value of \(\mu\) has to be set as 1.5. The parameter \({I}^{r}\) denotes the Intensity attribute which is directly proportional to the capability of the self-sharks and \(r\) is the abstract integer with the value ranging from 0 to 1. This value is directly proportional to the convergence rate of the sharks towards the odor.

When the value of \(r=0\); no prey has been detected and the exploration phase of the shark continues. When \(r=1\); it signifies the movement stage by absorbing the odor information of the self-sharks. If found the prey then it will be taken as the optimal solution. Sometimes when the value of \(r=1\); relatively the value of \(\rho\) also soars high. This implies that the algorithm supersedes the existing optimal solution and takes the new solution as the optimal solution thereby seizing exploration. The algorithm retains back the previous value of \(r\) which denotes the optimal position of the prey and further updates of the remaining solutions is computed. The superseding of the position of the sharks is computed as per the Eq. (5) defined as follows:

$${{Sh}_{i}}^{j}\left(t+1\right)=\frac{{{\widehat{Sh}}_{i}}^{j}\left(t+1\right)+{{Sh}_{i-1}}^{j}\left(t\right)}{2}$$
(5)

Foraging stage

While hunting, the Genghis Khan-type sharks make a sudden shift in their travel exhibiting a parabolic-shaped attack which implies a cooperative hunting strategy. The foraging behaviour exhibited by the sharks is defined as per the Eq. (6) as follows:

$${{Sh}_{i}}^{j}\left(t+1\right)={{Sh}_{best}}^{j}\left(t\right)+rand* \left({{Sh}_{best}}^{j}\left(t\right)-{{Sh}_{i}}^{j}\left(t\right)\right)+\alpha * {\omega }^{2}*\left({{Sh}_{best}}^{j}\left(t\right)-{{Sh}_{i}}^{j}\left(t\right)\right)$$
(6)

The value \(\alpha\) is taken as an abstract integer that lies between 1 to -1. The step size of the shark is impacted by the value of \(\omega\) which is conducive to global exploration then falls to the local optima gradually. The value of \(\omega \left(0\right)=1.0\) during the initialization stage. The formula to compute this factor \(\omega\) is defined as follows:

$$\omega =2*\left\{1-{\left(\frac{t}{T}\right)}^\frac{1}{4}+\left|\cup \left(t+1\right)\right|*\left[{\left(\frac{t}{T}\right)}^\frac{1}{4}-{\left(\frac{t}{T}\right)}^{3}\right]\right\}$$
(7)

The value \(\left|\cup \left(t+1\right)\right|\) at time t + 1 can be computed by using the Eq. (8) as follows:

$$\left|\cup \left(t+1\right)\right|=1-2{\omega }^{4}(t)$$
(8)

Self-defense stage

In case of any attack by the other marine creatures’ sharks lighten their body and perform an escaping pattern. This self-defense behaviour can be calculated using Eq. (9) as follows:

$$\left\{\begin{array}{c}{{Sh}_{i}}^{j}\left(t+1\right)={{Sh}_{i}}^{j}\left(t\right)+{k}_{1}\left({b}_{1}{{Sh}_{best}}^{j}\left(t\right)-{b}_{2}{{Sh}_{k}}^{j}\left(t\right)\right)+ {k}_{2}\cap \left({\complement }_{3}{{Sh2}_{i}}^{j}\left(t\right)-{{Sh1}_{i}}^{j}\left(t\right)\right)+\frac{{\cap }_{2}}{2}({{Sh}_{u1}}^{j}\left(t\right)-{{Sh}_{u2}}^{j}\left(t\right)); when {b}_{1}\le 0.5\\ {{Sh}_{i}}^{j}\left(t+1\right)={{Sh}_{i}}^{j}\left(t\right)+{k}_{1}\left({b}_{1}{{Sh}_{best}}^{j}\left(t\right)-{b}_{2}{{Sh}_{k}}^{j}\left(t\right)\right)+ {k}_{2}\cap \left({b}_{3}{{Sh2}_{i}}^{j}\left(t\right)-{{Sh1}_{i}}^{j}\left(t\right)\right)+\frac{{\cap }_{2}}{2}({{Sh}_{u1}}^{j}\left(t\right)-{{Sh}_{u2}}^{j}\left(t\right))\end{array}\right.$$
(9)

where \(-1<{k}_{1}<1\) and \({k}_{2}\) pursues random distribution with a mean value of 0 and a standard deviation of 1. The computation of the random abstract attributes \({b}_{1}\), \({b}_{2}\), \({b}_{3}\) are defined in the manner described below:

$$\left\{\begin{array}{c}{b}_{1}={L}_{1}*2*rand+\left(1-{L}_{1}\right)\\ {b}_{2}={L}_{1}*rand+\left(1-{L}_{1}\right)\\ {b}_{3}={L}_{1}*rand+\left(1-{L}_{1}\right)\end{array}\right.$$
(10)

where \({L}_{1}\in \left[\text{0,1}\right]\) and \(\cap\) is an adaptive coefficient that defines the relation between \(\gamma\) and \(\delta\).

$$\gamma =\left|\delta *\text{sin}\left(\frac{3\pi }{2}+\text{sin}\frac{3\pi \delta }{2}\right)\right|$$
(11)
$$\delta ={\delta }_{min}+\left({\delta }_{max}-{\delta }_{min}\right)*{\left({\left(1-\frac{iter}{Thresh}\right)}^{3}\right)}^{2}$$
(12)
$$\cap = \gamma *2*(rand-1)$$
(13)

The value of \({\delta }_{min}\)=0.2 and \({\delta }_{max}=1\). The formula to compute \({{Sh1}_{i}}^{j}\left(t\right)\) and \({{Sh2}_{i}}^{j}\left(t\right)\) are defined as follows:

$${{Sh1}_{i}}^{j}\left(t\right)={Lowerb}_{j}+rand*\left({upperb}_{j}-{Lowerb}_{j}\right)$$
(14)
$${{Sh2}_{i}}^{j}\left(t\right)={Lowerb}_{j}+rand*\left({upperb}_{j}-{Lowerb}_{j}\right)$$
(15)

The formula to find the value of

$${{Sh}_{k}}^{j}\left(t\right)={L}_{2}*\left({{Sh}_{\omega }}^{j}\left(t\right)-{{Sh}_{\tau }}^{j}\left(t\right)\right)+{{Sh}_{\tau }}^{j}\left(t\right)$$
(16)

\({{Sh}_{\omega }}^{j}\left(t\right)\) denotes the solutions of two random at time t (\(\omega =\text{1,2},\dots N)\). Similarly, \({l}_{1},{l}_{2}\in \left[\text{0,1}\right]\) and \({{Sh}_{\tau }}^{j}\left(t\right)\) Implies the set of solutions generated randomly at time t. \({{Sh1}_{i}}^{j}\left(t\right)\) and \({{Sh2}_{i}}^{j}\left(t\right)\) are the selected random solutions. This provides the optimal solution by decomposing towards the optimal search space. This is similar to that of a neighbourhood search. This is a distinguishing feature or a property of GKSO which performs global exploration very efficiently.

Simulated annealing

Simulated annealing evolved by mimicking the behaviour of the metal being heated. It works based on the metal iron heating and cooling strategy. Let the temperature at the time \({t}_{0}\) represent the initial heat, and the temperature at \({t}_{f}\) represent the final temperature. The cooling rate is denoted as \(\varphi\). A random set of solutions is generated. Which then gradually superseded by the upcoming solution prior to the previous solution. When the target temperature is met, then the algorithm gets terminated.

$${\text{Temp}} = {\text{Temp}}*\varphi ; \, \varphi \in \left[ {0,1} \right]$$
(17)

To alleviate the problem of local convergence, probabilities based on the worst solution can be accepted. The current temperature can be computed as.

\(e\left(\frac{-\Delta }{Temp}\right)\le rand\) (18).

The factor \(\Delta\) is the fitness function which is a difference between the current optimal to that of the previous optimal solution. The proposed hybridized GKSO-SA involves the steps depicted using Fig. 13. Algorithm 1 provides the working procedure of the hybridized GKSO-SA.

Fig. 13
figure 13

Hybridized GKSO-SA.

Algorithm 1
figure a

Hybrid GKSO-SA.

Feature classification using CatBoost classifier

CatBoost classifiers possess advantages such as high classification performance, the capability to handle categorical data, and a robust regularization technique88. CatBoost is a potent ensemble learning-based predictive model that utilizes multiple decision trees (DT). The classifier generates a subsequent number of trees, which performs error corrections made by the previous tree89. Since the model maintains a symmetric tree organization, the output will not be affected by the order in which the features are given as input. The significant advantage of the CatBoost classifier is its performance by seizing the association between the categorical variable and the target variable90. The output obtained after classifying paddy diseased features will be either numerical (0 or 1) or binary (yes or no). A smoothing logloss function will be utilized for testing and validation91.

Experimentation

This section presents the experimental setup, hardware and software configurations, dataset preparation, and evaluation metrics used to valid the proposed model.

Implementation set-up

The proposed methodology was developed using Python93 and implemented on Google Colab92, leveraging libraries like Pandas, Keras, NumPy, and TensorFlow. The experiments were conducted using NVIDIA DGX GPU servers equipped with 512 GB of RAM and 8 high-speed Tesla V100 graphics proceeding units (GPUs), each with a capacity of 32 GB.

Datasets used

Table 3. Distribution of diseases pertaining to paddy. The dataset utilized in the proposed methodology involves nine distinct kinds of paddy plant diseases such as blast, downy mildew, bacterial panicle blight, bacterial leaf blight, tungro, dead heart, bacterial leaf streak, hispa, and brown-spot obtained from the doctor paddy dataset. Paddy doctor dataset has been obtained from Kaggle94. The resulting dataset gets segregated into a set of training, testing, and validation sets in the ratio of 70:15:15.

Table 3 Presents the distribution of illnesses associated with paddy crops.

Evaluation metrics

The effectiveness of the proposed hybrid model for classifying paddy leaf disease is evaluated using the following performance indicators:

  1. i.

    True Positives (TruePs) The feature is classified as true positives when the classifier successfully labels it.

  2. ii.

    True Negatives (TrueNs) True negatives are those features that fall under negative classification if the classifier deems them negative.

  3. iii.

    False Positives (FalsePs) When the classifier incorrectly classifies then the feature belongs to True Negatives.

  4. iv.

    False Negatives (FalseNs) False negatives are produced when the classifier classifies positive features as negative.

  5. v.

    Precision This metric indicates the number of correctly identified positive samples, commonly known as the precision score.

    $$Precision \left(P\right)=\frac{TruePs}{TruePs+FalsePs}$$
    (19)

vi) Recall This metric denotes the ratio of correctly identified positive samples to the total number of actual positive samples.

$$Recall \left(R\right)=\frac{TruePs}{FalseNs+TruePs}$$
(20)

vii) Accuracy This represents the proportion of correctly classified samples out of the total classifications.

$$Accuracy (Accy)=\frac{TruePs+TrueNs}{TruePs+TrueNs+FalsePs+FalseNs}$$
(21)

viii) F1-measure This value represents the balance between precision and recall, calculated as their harmonic mean

$$F1-Score=2*\frac{Precision*Recall}{Precision+Recall}$$
(22)

Result analysis and discussion

The methodology has been implemented and analyzed for its performance in different machine learning and deep learning models, as well as existing systems, using metrics such as accuracy, precision, recall, and F1-score.

Comparative analysis with various machine learning classification models

The proposed methodology has been classified against distinct ML classifiers like SVM, Random Forest, XGBoost, KNN, Naive Bayes, and CatBoost. The anticipated feature set was intended to enhance classifier performance, though it did increase the overall processing time required for the machine learning pipeline to complete. Table 4 displays the performance results of classifying the proposed methodology using different machine learning classifiers.

Table 4 Presents a comparison of various machine learning models.

The proposed model, along with the deep features obtained from the MobileNetV3 + GKSO-SA + CatBoost classifier and other ML classifiers, as depicted in Fig. 14, demonstrates significant performance improvements. Support Vector Machine (SVM) reached an accuracy of 86.72%, with a precision of 83.15%, a recall of 87.07%, and an F1-score of 86.04%, indicating balanced performance across these metrics. Random Forest performed slightly lower, attaining an accuracy of 80.71%, precision of 81.01%, recall of 76.47%, and an F1-score of 79.81%, suggesting its suitability when precision is more important than recall. XGBoost delivered an accuracy of 83.31% and showed similar precision and recall values (82.93%), resulting in an F1-score of 81.93%, highlighting its consistency across precision and recall. KNN yielded an accuracy of 81.34%, the highest precision among these classifiers (84.21%), with recall at 81.21% and an F1-score of 80.43%, making it a reasonable choice when higher precision is required. Naïve Bayes demonstrated the lowest performance, with an accuracy of 79.24%, precision and recall at 77.47% and an F1-score of 76.12%, indicating limitations in handling complex datasets. CatBoost outperformed all other models, achieving an impressive accuracy of 92.6%, precision of 89.12%, recall of 93.21%, and an F1-score of 92.13%, making it an excellent choice for situations where both high accuracy and recall are essential.

Fig. 14
figure 14

Shows a comparative analysis of various machine learning methods.

Comparative analysis with various deep learning classification models

The proposed methodology has been classified against different DL techniques such as Bi-GRU, LSTM, Bi-LSTM, CNN-Bi-GRU, CNN-LSTM, CNN-Bi-LSTM, CNN-GRU, and CNN. Table 5 displays the performance results of standard deep learning classifiers, comparing various deep learning models.

Table 5 Comparison of different deep learning models.

Figure 15 presents a comparative analysis of various deep learning methods, where CNN achieved a high accuracy of 96.22%, with precision at 97.99%, recall at 96.1%, and an F1-score of 94.22%. This model displays a balanced and robust performance across all metrics, showing its ability to capture complex data patterns effectively. LSTM recorded an accuracy of 92.99%, with its precision, recall, and F1-score values slightly lower than CNN, at 91.29%, 90.75%, and 89.26%, respectively. While it performs well, LSTM may be more suited for tasks where moderate accuracy is acceptable, as it doesn’t quite reach the top performance observed in CNN and Bi-LSTM. Bi-LSTM stands out as the highest-performing model, recording an accuracy of 97.55%, precision of 96.42%, recall of 95.28%, and F1-score of 94.66%. The bidirectional characteristic improves its efficacy, rendering it appropriate for situations necessitating optimal accuracy and recall. Bi-GRU achieved an accuracy of 90.81% with a precision of 92.54%, a recall of 91.88%, and an F1-score of 92.11%. Although it performs well in precision and recall, its accuracy is slightly lower than CNN and Bi-LSTM. CNN-LSTM attained an accuracy of 91.25%, precision of 91.85%, recall at 90.15%, and an F1-score of 91.92%. This model offers a balanced performance, though it does not reach the highest accuracy scores seen in Bi-LSTM and CNN, making it a moderate choice. CNN-Bi-LSTM demonstrated an accuracy of 95.18%, with precision, recall, and F1-score values of 94.48%, 93.21%, and 92.12%, respectively. This combination shows strong performance and is especially beneficial for tasks that benefit from both CNN’s feature extraction and Bi-LSTM’s bidirectional processing. CNN-GRU achieved an accuracy of 93.95%, precision at 93.45%, recall at 92.45%, and an F1-score of 91.21%. While this model performs well, it doesn’t match the highest accuracy of Bi-LSTM or CNN, but it is still a practical choice in terms of balanced performance. CNN-Bi-GRU had an accuracy of 94.25%, with precision, recall, and F1-score values of 94.81%, 93.73%, and 90.75%, respectively. This model demonstrates vital precision and recall, although its F1-score is somewhat lower compared to other top-performing models.

Fig. 15
figure 15

Comparative analysis of various deep learning methods.

Comparative analysis of GKSO-SA versus other optimizers (PSO, GA, ACO)

This section provides a performance comparison between the proposed GKSO-SA hybrid optimization method and traditional metaheuristic algorithms, including Particle Swarm Optimization (PSO), Genetic Algorithm (GA), and Ant Colony Optimization (ACO), with emphasis on their effectiveness in feature selection and classification accuracy. Table 6 presents the performance results from a comparative analysis of the proposed GKSO-SA method against other optimization algorithms.

Table 6 Comparison of other optimization models.

Comparative analysis with various existing methodologies

The proposed methodology has classified the existing methods, such as Mbi-LSTM23, CNN-VGG29, DCNN31, IoT-CNN33, and GCN34. Table 7 shows the performance results from a comparative analysis of various existing methodologies.

Table 7 Comparative examination of different established methodologies.

Figure 16 presents the performance evaluation of the proposed model in comparison to the previous approach. A comparative analysis was conducted between the proposed model, which integrates MobileNetV3 with GKSO-SA (Genghis Khan Shark Optimizer with Simulated Annealing) and CatBoost, and several existing methodologies. The evaluation was based on metrics such as accuracy, precision, recall, and F1-score. Mbi-LSTM achieved an accuracy of 97.16%, with a precision of 96.85%, a recall of 97.25%, and an F1-score of 95.99%. This model demonstrates strong overall performance but needs to improve accuracy and F1-score compared to the proposed model. CNN-VGG reported an accuracy of 93%, precision of 92.85%, recall of 91.55%, and an F1-score of 90.54%. While this model performs well, its lower accuracy and F1 score highlight its limitations compared to more advanced or hybrid approaches. DCNN (Deep Convolutional Neural Network) showed an accuracy of 96.08%, with a precision of 95.52%, a recall of 94.45%, and an F1-score of 95.75%. DCNN performs effectively, though its accuracy and F1-score remain slightly lower than the proposed model’s. IoT-CNN achieved an accuracy of 97.7%, a precision of 96.13%, a recall of 95.55%, and an F1-score of 96.55%. This model demonstrates strong performance, particularly in the F1-score, but the proposed model surpasses it in accuracy. GCL recorded an accuracy of 97%, precision of 96.55%, recall of 95.25%, and an F1-score of 94.32%. GCN’s performance is solid, especially in precision, but it is still slightly outperformed by the proposed method in accuracy and F1-score. The proposed model (MobileNetV3 + GKSO-SA + CatBoost) stands out with the highest accuracy of 98.52%, precision of 95.45%, recall of 96.96%, and an F1-score of 97.54%. The combination of MobileNetV3’s lightweight architecture with the optimization capabilities of GKSO-SA and the robust classification power of CatBoost contributes to this model’s superior performance.

Fig. 16
figure 16

Performance evaluation of the proposed model in comparison to the previous approach.

Conclusion and future work

The research emphasizes the urgent need for efficient disease management strategies in rice farming, given the impact of plant diseases on both crop yield and quality. With increasing demand due to population growth, early and accurate detection of rice diseases is essential for timely intervention. The proposed automated deep learning model successfully enhances disease identification and categorization on paddy leaves by utilizing advanced image processing, feature extraction, and selection techniques. By utilizing MobileNetv3 for feature extraction and hybridizing Genghis Khan Shark Optimization (GKSO) with Simulated Annealing (SA) for optimal feature selection, the model attains a robust accuracy of 98.52% in classifying diseases. This highlights the model’s efficacy in improving agricultural productivity through deep learning technologies. Future work will focus on advancing this model by integrating Explainable AI (XAI) methods to enhance the transparency and interpretability of disease detection, enabling farmers to better comprehend the model’s predictions better.

Further enhancements may involve Artificial Specific Intelligence (ASI) techniques to create highly specialized models tailored for rice disease detection, ensuring even higher classification accuracy and efficiency. Exploring Artificial General Intelligence (AGI) could also expand the model’s adaptability, allowing it to autonomously generalize across various crop types and disease classifications. Together, these advancements could provide a more robust and adaptive tool for crop disease management.