Introduction

Recently, there has been a noticeable trend towards the convergence of computer vision and artificial intelligence within plant pathology. This convergence between the two domains is becoming more prominent and increasing the scope of development of new technology in the early diagnosis and monitoring of plant diseases1,2. Plant pathology aims to investigate the plant diseases induced by pathogens and environmental factors3,4. Particularly, using computer vision to detect crop diseases helps farmers take steps at the right moment for associated treatments. Apart from the crops, other types of trees are also considered for pathological research, depending on their medicinal and economic significance.

Currently, researchers are eager to identify diseases at the earliest possible stage. This is because early detection of diseases can help prevent problems such as minimizing yield losses5,6, preventing disease spread in large-scale spread7,8, managing specific pathogens and emerging diseases9,10, and so on. Furthermore, early disease identification is essential for sustainable and precision agriculture as it enables accurate and timely treatments that reduce environmental impact and enhance resource utilization11,12. Current methods for early detection reduce the need for chemical pesticides, encouraging environmentally sensitive disease management and substantially enhancing sustainable agricultural practices13. Severity measurement provides a quantitative assessment of how much a disease impacts a plant or its parts. This provides more information than identifying the existence or absence of a disease. An adequate understanding of severity assists in evaluating the intensity of the illness in individual plants or within a population. It can serve as a crucial component of disease monitoring and management. An appropriate severity-measuring approach may accurately assess overall plant health, monitor disease progression, and enable intelligent decision-making for precise and targeted treatment. In the automatic severity estimation approach, the accurate localization of the disease region on the leaf is crucial. Furthermore, accurately classifying the lesion stage is essential for exact severity assessment. The proposed work addresses the first task by optimizing parameters for simple linear iterative clustering (SLIC) and the second task by classifying superpixels using the optimized feature set. Researchers have developed several strategies for illness detection and severity estimation in recent years, explored in the section “Related works on plant disease detection”. A limited number of research studies have investigated the SLIC for severity estimation. The SLIC parameters have been optimized to provide more precise results, followed by feature selection, superpixel classification, and, ultimately, severity score computation in the proposed work.

Chlorosis is among the most frequently encountered diseases affecting green leaves, leading to the yellowing of leaves14. The main reason for chlorosis is a viral infection. At the beginning of yellowing, no significant leaf structure and geometry changes are observed. Chlorophyll decreases cause due to chnage of the structure and function of the chloroplast. Chlorophyll is essential for photosynthesis and gives a green color to the leaf. Iron and manganese deficiency results in leaf yellowing (chlorosis)15. Chlorosis is also caused by excess potassium, magnesium, and phosphorus16. Chlorosis is a visible symptom that might indicate several underlying concerns, including nutrient deficits, infections, pests, environmental stresses, or physiological disorders. Plant pathologists can accurately diagnose and identify the underlying causes of plant health problems by comprehending chlorosis patterns and related factors. Furthermore, it has the potential to greatly influence both the quantity and quality of crop production. Plant pathologists can enhance agricultural productivity by analysing chlorosis patterns and identifying their underlying causes. Implementing effective techniques for alleviating nutritional deficiencies, diseases, and other stress factors is made possible because of this, which ultimately results in a reduction in yield losses. In addition, chlorosis is a symptom that can be caused by a wide variety of illnesses and pests. Developing an integrated disease prevention and surveillance control strategy necessitates thoroughly comprehending the relationship between chlorosis and insect or disease infestations. Plant pathologists must accurately identify and measure chlorosis in order to develop the appropriate strategy for automated decision-making in smart agriculture, as evidenced by the previous discussion. The principal objective of our investigation is to develop a framework for accurately quantifying chlorosis, thereby enabling the possibility of early prediction. Therefore, we developed a system that enables automated localization. The localization of the diseased region is enhanced by an optimization layer that enables the automated adjustment of system parameters based on disease characteristics to some extent. Moreover, precise classification of the superpixels enables the system to identify the severity stages. Ultimately, it may serve as a tool for high-precision disease detection, facilitating early identification in precision agriculture.

In the framework, the superpixel method is employed on plant leaves to produce superpixels that cluster the different yellowing stages. An evolutionary optimization method is utilized to improve the quality of superpixels and adjust parameter values based on leaf conditions. Subsequently, color features are derived from four different types of superpixels. A novel feature selection strategy based on multi-swarm Cuckoo search has been implemented to identify crucial features from the extensive feature set.

Finally, a model for estimating severity based on superpixels has been introduced to assess the overall severity score accurately. The following points distinguish the proposed method from other existing models.

  1. 1.

    A machine learning-based precise chlorosis quantification model for plant chlorosis monitoring has been proposed.

  2. 2.

    The Evolutionary superpixelling method has been introduced for accurate disease localization. Furthermore, the evolutionary approach helps to automatically select the superpixel’s parameter value according to the leaf’s disease stage.

  3. 3.

    The suggested approach is a comprehensive model for chlorosis disease detection, lesion area localization, disease stages, and severity score calculation.

  4. 4.

    Additionally, a novel feature selection method based on a multi-swarm Cuckoo search has been deployed to optimize the selection of features from a vast collection, thus enhancing the efficiency of the classification model.

  5. 5.

    The chlorosis severity index of the diseased leaf has been proposed and calculated to understand the effect of chlorosis on the leaf.

  6. 6.

    Extensive experiments have been conducted to validate the proposed model with different evaluation parameters.

The structure of the paper is as follows. An overview of the possible use of computer vision in plant pathology is provided in the section “Introduction”. The paper discusses recent advancements in plant disease diagnosis technologies in the section “Related works on plant disease detection”. The required techniques, assessment metrics,and details on the dataset are found in the section “Materials and methods”. The section “Proposed precise chlorosis severity assessment system” comprehensively describes the proposed precise chlorosis assessment system. The section “Results and discussion” illustrates the exhaustive analysis of the results obtained from the various stages of the proposed system. The conclusions have been summarized in the section “Conclusion & future work”.

Related works on plant disease detection

Over the years, many approaches have been developed to detect and classify an infection on plant leaves automatically. Compared to conventional processing methods, the wavelet transform results improve image resolution and better preserve edge information. Discrete Wavelet Transform (DWT) is one of the most widely used methods in leaf disease identification agriculture. Pankaja et al.17 presented a leaf identification model which consists of statistical information from HSV color moments and local features. Texture information and shape features are extracted by DWT and Chebyshev Moments, respectively. All three features are then combined to produce the final feature vector. The proposed model took thirty classes from the Flavia dataset, totalling 270 leaf images. This proposed feature extraction scheme improves the system’s accuracy to \(96.29\%\). Islam et al.18 presents the method which decomposes the input disease image into horizontal, vertical, and diagonal sub-bands using the multi-resolution analysis and Discrete Wavelet Transform (DWT). The random subspace Method was used to classify data by a combination of linear classifiers. The method has an accuracy rate of \(95\%\) in identifying four major rice plant diseases. Tampinongkol et al.19 proposed a method to identify the source of diseases from Jabon leaf image in the nursery. Segmentation has been done by reducing the RGB color cylinders to isolate the disease entity from the background. Finally, using DWT in the 3-level decomposition process, the disease type can be identified. SVM classification results show that the disease groups distinguish from each other with accuracy \(84.67\%\). Dhingra et al.20 have proposed a computer vision-based leaf disease detection method by combining DWT with different Gabor filter orientations to find the features. The later stage employs a histogram binning pattern. A Gaussian classifier makes the final prediction on the extracted features. They did a comparative analysis on five leaf data sets and proved the supremacy of the method. On various data sets, the proposed system got results with an accuracy of \(98.45\%\), \(96.45\%\), \(99.85\%\), \(99.83\%\), and \(99.75\%\). Superpixeling is a clustering technique that groups adjacent pixels to form homogeneous regions, seeing the similarity of color or texture information21. A superpixel-based method for identifying leaf disease has been proposed in recent years. Zhang et al.22 introduced a hybrid clustering-based segmentation approach for plant disease identification. In this work, the simple linear iterative clustering (SLIC) divides the entire color leaf image into several compact and nearly uniform superpixels. Then, the pixels defining the lesion are segmented from each superpixel by expectation (objective) maximization. The experimental results show that the method is effective compared to other methods. Another hybrid approach proposed by Zhang et al.23 for detecting cucumber diseases uses superpixels. The method consists of two stages; where in the first stage, the diseased leaf image is partitioned into several compact regions using the SLIC algorithm. The logarithmic frequency PHOG features are extracted and used for disease detection in the second stage. A fusion-based approach presented by Khan et al.24 segments the lesion area of the diseased leaf, which suffers from uneven illumination. In this method, the input image is transformed into a color-balanced image and subsequently the SLIC method is applied. Singh25 proposed PSO-based lesion area segmentation of six types of sunflower disease. The texture features are extracted from each lesion category and classified with a minimum distance classifier. Finally, this method achieved the average classification of \(98\%\). An adaptive pixel intensity-based thresholding approach was developed by Sengar et al.26 to segment the lesion area of cherry leaves. The method achieved a very high accuracy of \(99.0\%\). Pandey et al.27 presents an automatic and non-destructive detection of the healthy, mild, and severe categories of Vigna-mungo leaf. This method segments the leaf images and extract features thereafter. Finally, the disease is predicted using a support vector machine (SVM). Moreover, it can predict the category of the healthiness of the leaf with an accuracy of \(95.69\%\). Chakraborty et al.28 evaluate the effectiveness of various deep learning models, including VGG16, VGG19, MobileNet, and ResNet50, on the PlantVillage Dataset to create an automated system for detecting late and early blight infections in potato leaves. Compared to other models, VGG16 exhibits the highest accuracy of \(92.69\%\), as evidenced by the experiment. Additionally, they improve the efficacy of VGG16 by adjusting its parameters. Compared to healthy potato leaves, the improved model achieved a classification accuracy of \(97.89\%\) in distinguishing between late and early blight disease. Phan et al.29 proposes an automated method to detect diseased regions on corn leaves using deep learning and image segmentation. By applying Simple Linear Iterative Clustering (SLIC) to segment corn leaf images into super-pixels, and using pre-trained CNN models (like VGG16, ResNet50, DenseNet121, etc.), the system classifies regions as healthy or affected by specific diseases (e.g., NLB, GLS, rust). The DenseNet121 model achieved the highest accuracy of \(97.77\%\) on the CD&S dataset with optimal SLIC and train-test settings. The approach proves effective for replacing manual scouting with automated, image-based disease detection and has been deployed via web and mobile applications. Abisha et al.30 presents a novel approach for detecting and categorising diseased brinjal leaves using Deep Convolutional Neural Networks (DCNN). The first phase was the successive use of image enhancement and segmentation techniques on the original brinjal leaf image. Subsequently, the discrete Shearlet transform is utilised to extract the characteristics of the segmented region afflicted by the illness. The Deep Convolutional Neural Network (DCNN) attained a mean accuracy of \(93.30\%\) in accurately categorising leaf diseases. Singla et al.31 have created a web-based application that utilises a deep learning model. The objective is to detect plant leaf diseases by analysing a image of a leaf and then alerting farmers through message. They examined the different deep learning models using the data from PlantVillage. MobileNet attained a classification accuracy of \(97.35\%\) in the multi-classification challenge. Bedi et al.32 presented a computationally lightweight method for the early diagnosis of plant diseases. The system comprises a Convolutional Autoencoder (CAE) and Few-Shot Learning (FSL), whereby the CAE is trained to reconstruct leaf images with minimal loss, and its pre-trained layers are utilized to develop classification and segmentation models trained with FSL. The technique offers a scalable solution for diagnosing plant diseases and estimating their severity while requiring minimal annotation effort.

Table 1 gives a summary of the related works.

Table 1 A summary of the recent works related to disease detection.

Materials and methods

This section briefly explains the SLIC method, Colour GLCM features, principal component analysis, evolutionary feature selection, the data used in the experiment, and other relevant entities associated with this experiment.

Simple linear iterative clustering

Superpixel methods are quickly becoming famous for a variety of computer vision applications. It decreases image redundancy and reduces the complexity of subsequent image-processing tasks. Among the various superpixel techniques, the work reported in34 proposed imple linear iterative clustering (SLIC) as an effective superpixel generation method. The method uses three-dimensional color space and a two-dimensional image plane, forming a five-dimensional space. The inputs to the algorithm are input image (\(i_m\)), number of superpixels \(\left( k\right)\), weighting factor \(\left( m\right)\), and radius \(\left( r\right)\). The details of the parameters, k, m and r are discussed in the section “Performance analysis of evolutionary superpixel” and a suitable optimization of these parameters has been presented.

Algorithm 1
Algorithm 1
Full size image

SLIC-Algorithm

Color-GLCM: texture features

The gray-level co-occurrence matrix (GLCM) is used widely for texture analysis. In this work, a color-based texture analysis has been used based on the color co-occurrence matrix (CCM) to detect disease and subsequently for severity measurement of a leaf. The CCM is calculated using four distance values, \(d=1\), 2, 4, and 8 and thirteen (13) directions considering the 3D color channels35,36. The twelve (12) features used by us are known as Haralick features37 extended in 3D GLCM (color-GLCM), and they are given in the equations given below (Eqs. 112). They are also reported in35,36.

$$\begin{aligned} Energy= & \sum _{i=1}^{L}\sum _{j=1}^{L}M_{d}^{\phi }\left( i,j\right) ^2 \end{aligned}$$
(1)
$$\begin{aligned} Entropy= & \sum _{i=1}^{L}\sum _{j=1}^{L}M_{d}^{\phi }\left( i,j\right) log_{2}\left( M_{d}^{\phi }\left( i,j\right) \right) \end{aligned}$$
(2)
$$\begin{aligned} Correlation= & \sum _{i=1}^{L}\sum _{j=1}^{L}M_{d}^{\phi }\left( i,j\right) \frac{\left( i-\mu _{x}\right) \left( i-\mu _{y}\right) }{\sigma _{x}\sigma _{y}} \end{aligned}$$
(3)
$$\begin{aligned} Contrast= & \sum _{n=1}^{L}\left( n^2 \sum _{i=1}^{L}\sum _{j=1}^{L}M_{d}^{\phi }\left( i,j\right) \right) , n=\left| i-j \right| \end{aligned}$$
(4)
$$\begin{aligned} Inverse \ difference \ moment= & \frac{1}{1-\left( i-j \right) ^2 }\sum _{i=1}^{L}\sum _{j=1}^{L}M_{d}^{\phi }\left( i,j\right) \end{aligned}$$
(5)
$$\begin{aligned} Variance= & \sum _{i=1,j=1}^{L}\left( 1-\mu \right) ^2 M_{d}^{\phi }\left( i,j\right) \end{aligned}$$
(6)
$$\begin{aligned} Sum average= & \sum _{i=1}^{2L-1}\left( ip_{x+y}\left( i \right) \right) \end{aligned}$$
(7)
$$\begin{aligned} Dissimilarity= & \sum _{i=1}^{L}\sum _{j=1}^{L}\left( i-j\right) M_{d}^{\phi }\left( i,j\right) \end{aligned}$$
(8)
$$\begin{aligned} Cluster \ shade= & \sum _{i=1}^{L}\sum _{j=1}^{L}\left( i+j-\mu _{x}+\mu _{y}\right) ^3 M_{d}^{\phi }\left( i,j\right) \end{aligned}$$
(9)
$$\begin{aligned} Cluster \ prominence= & \sum _{i=1}^{L}\sum _{j=1}^{L}\left( i+j-\mu _{x}+\mu _{y}\right) ^4 M_{d}^{\phi }\left( i,j\right) \end{aligned}$$
(10)
$$\begin{aligned} Maximum \ probability= & \sum _{i=1}^{L}\sum _{j=1}^{L}\max _{i,j}\left\{ M_{d}^{\phi }\left( i,j\right) \right\} \end{aligned}$$
(11)
$$\begin{aligned} Difference \ variance= & \sum _{i=1}^{L}\sum _{j=1}^{L}i^{2}p_{x-y}\left( i\right) \end{aligned}$$
(12)

Here, L is the number of distinct gray levels in the image window, M(ij) represents the (ij)-th entry in the color-GLCM (3D-GLCM considering three channels R, G and B), and \(p_{x}(i)=\sum _{i=1}^{L} M_{d}^\phi\), \(p_{y}(j)=\sum _{i=1}^{L} M_{d}^\phi\) denotes the i-th and j-th entry in the marginal-probability distribution matrix; \(\sigma _x\), \(\sigma _y\), \(\mu _x\), \(\mu _y\) denotes the means and standard deviations of the partial probability density functions \(p_x\) and \(p_y\) respectively and expressed as given in Eqs. 13-16.

$$\begin{aligned} \sigma _{x}= & \left( \sum _{i=1}^{L} \sum _{j=1}^{L}\left( i-\mu _{x}\right) ^{2}\right) ^{1 / 2} \end{aligned}$$
(13)
$$\begin{aligned} \sigma _{y}= & \left( \sum _{i=1}^{L} \sum _{j=1}^{L}\left( j-\mu _{y}\right) ^{2}\right) ^{1 / 2} \end{aligned}$$
(14)
$$\begin{aligned} \mu _{x}= & \sum _{i=1}^{L} \sum _{j=1}^{L} i M_{d}^\phi (i, j) \end{aligned}$$
(15)
$$\begin{aligned} \mu _{y}= & \sum _{i=1}^{L} \sum _{j=1}^{L} j M_{d}^\phi (i, j) \end{aligned}$$
(16)

Principal component analysis

In general, Principal Component Analysis (PCA) is a technique used to reduce the dimensionality of a dataset without significant loss of information38. It transforms the original set of possibly correlated features into a new set of uncorrelated variables called principal components39. These components are ordered by the amount of variance they capture from the data, allowing for effective dimensionality reduction. PCA can thus identify the most informative directions in the feature space, which can then be used as input to supervised or unsupervised learning algorithms. In the literature, PCA-based feature extraction has been widely employed to improve the performance of classification algorithms40.

Table 2 Parameters’ settings for various classifiers used in the work.

Evolutionary feature selection

Evolutionary Feature selection strategy has advantages over traditional feature selection methods when the feature vector dimension is very large41. It treats the feature selection problem as a subset selection problem. The evolutionary approach quickly finds the optimal or near-optimal feature set by avoiding the exhaustive search on a large feature set41. Recent research on evolutionary feature selection will be covered here. Saraç et al.42 used an ant colony optimization (ACO) technique to choose the best features for the automatic classification of web pages. They reported that the ACO-based strategy finds better features over information gain and chi-square techniques. Yadav et al.43 introduced a filter-based feature selection method that utilizes particle swarm optimization (PSO) to find the optimal feature set to improve classifier performance in the named entity recognition problem of biomedical natural language processing. Marie-Sainte et al.44 applied the firefly algorithm to determine the best feature for classifying Arabic text. The experiment was conducted on an OSAC dataset, and the approach achieved a precision value of \(99.40\%\). Alzaqebah et al.45 proposed a modified Cuckoo search for feature (genes) selection from the cancer gene expression datasets. They used the concept of memory with the traditional Cuckoo search algorithm to consider the selected features of each iteration. Finally, microarray datasets have been used to compare the method with other algorithms. Khan et al.46 introduced a new feature selection algorithm based on genetic algorithm (GA) called diversification of Population (DP) in GA, and named DPGA. They utilized this method to identify the most advantageous features from the Uniform variant of the LTrP (ULTrP) feature set extracted from the microstructural images. Finally, classifying a seven-class microstructural image dataset using a reduced feature set confirms its superiority over other contemporary methods. In our work, a new multi-swarm-based feature selection strategy has been proposed that will be discussed in the section “Multi-swarm cuckoo search for feature selection”.

Supervised machine learning

Supervised learning involves an input variable X and an output variable Y, and it uses an algorithm to learn the mapping from the input set to the output set. The goal is to approximate the mapping function for an unknown input data \(x'\), such that we can predict the corresponding output \(y'\). The algorithm learns using the training dataset, and when the algorithm achieves an acceptable level of accuracy in performance, it stops learning. A brief introduction of the four methods used in our experimentation is given in the following subsections. The parameters’ settings of all the classifiers used by us in this work are shown in Table 2.

Decision tree

The decision tree (DT) obtains the knowledge as a tree, which can also be expressed as a set of discrete rules47. The critical aspect of the decision tree classifier is its ability to use different feature subsets. Moreover, decision rules should be used at different steps of classification. The decision tree classifier is an efficient and robust classifier for complex decision-making problems. Each internal node of the decision tree implies the test on an attribute, and each branch node denotes the test outcome. Each leaf node of the decision tree implies the class label. It performs the classification by splitting branches of the tree. The classification algorithms based on the decision tree can be found in the literature in different forms48.

K-nearest neighbors

The k-Nearest-Neighbours (KNN)classifier is a simple and effective classifier based on one lazy learning method. It has been used as a non-parametric technique in statistical estimation and pattern recognition. The basic concept of the KNN classifier is to store the available instances and classify the testing instances as per the similarity that is based on distance49. For a specific test sample and a training set, the distances between the test sample and the other samples in the training set can be calculated. The training set closest to the test sample has the shortest distance. As a result, the test sample is classified based on the classification of its nearest neighbor50.

Support vector machine

The Support Vector Machine (SVM) theory was introduced by51. A hyperplane is built to classify data in appropriate classes in realistic time51. The data points closest to the hyperplane are referred to as support vectors. The performance of the SVM classifier depends on maximizing the distance between class data points on the hyperplane52. SVM uses a different mathematical function called the kernel function. Choosing the proper kernel function is vital to obtain the maximum performance53.

Multilayer perceptron (MLP)

Multilayer perceptron (MLP) classifier consists of three layers: input, output, and hidden. The input layer receives the signal for further processing, and classification is done by the output layer54. The hidden layer between input and output works as the central computation unit of multilayer perceptron. The perceptron generates a single output from multiple real-valued inputs using a linear combination of weighted inputs. The output is forwarded through some nonlinear activation functions. It uses the back propagation learning algorithms to train the network55. This classifier can solve most problems that are not linearly separable56.

Experimental setup

The necessary parameters’ settings are shown in Table 2. The default settings for all classifiers were used as provided by the respective tools (e.g. WEKA, MATLAB) to ensure reproducibility and alignment with standard baseline practices. This study aims to assess the performance of the proposed model independent of significant hyperparameter tuning.

Pongamia pinnata dataset

The photos of Pongamia pinnata were taken in a controlled setting between March and May 2019 by the Madhav Institute of Technology & Science in India57,58. The acquisition procedure was fully equipped with Wi-Fi connectivity. The leaf images were taken with a Nikon D5300 camera equipped with a built-in performance timing feature for capturing shots. The images were captured with a \(18-55\) mm lens in JPG format, with RGB color representation, 24-bit depth, 1000-ISO sensitivity, and without a flash. For experimentation, a set of 550 images has been used in this work. Out of 550 images, 275 are healthy images and 275 images contain chlorosis patches of various severity.

Performance evaluation metrics

The different classification evaluation parameters have been discussed here. This experiment has considered a four-class classification problem with categories of superpixels labelled as Type-1, Type-2, Type-3, and Type-4. The terms True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) are defined by focusing on one type at a time. For example, if we consider Type-1 as a True Positive (TP), the model correctly predicts a superpixel as Type-1 when it is Type-1. A False Positive (FP) occurs when the model predicts Type 1, but the actual class is either Type 2, Type 3, or Type 4. A False Negative (FN) happens when the actual class is Type-1, but the model incorrectly predicts it as Type-2, Type-3, or Type-4. Finally, a True Negative (TN) means the model correctly predicts an image as not Type-1, i.e., it correctly identifies an image as Type-2, Type-3, or Type-4 when it is truly one of those. This same consideration is repeated separately for Type-2, Type-3, and Type-4 to fully evaluate the model’s performance across all classes.

Accuracy. Classification accuracy is a quantitative measure that evaluates the effectiveness of a classification model by calculating the ratio of correct predictions to the total number of predictions made.

$$\begin{aligned} Accuracy=\frac{TP+TN}{TP+FN+FP+TN} \end{aligned}$$
(17)

Precision. Precision evaluates the fraction of correctly classified instances or samples among the ones classified as positives.

$$\begin{aligned} Precision=\frac{TP}{TP+FP} \end{aligned}$$
(18)

Recall. In an imbalanced classification problem with more than two classes, recall is calculated as the sum of true positives across all classes divided by true positives and false negatives across all classes.

$$\begin{aligned} Recall=\frac{TP}{TP+FN} \end{aligned}$$
(19)

Sensitivity. Sensitivity (true positive rate) refers to the proportion of those who have the condition truly that received a positive test result.

$$\begin{aligned} Sensitivity=\frac{TP}{TP+FN} \end{aligned}$$
(20)

Specificity. Specificity (true negative rate) refers to the proportion of those which are truly negative and received a negative test result.

$$\begin{aligned} Specificity=\frac{TN}{TN+FP} \end{aligned}$$
(21)

F1-Score. The F1-score or F-measure calculates the harmonic mean of precision and recall, providing an equitable assessment of the model’s performance. A high F-score signifies that the model exhibits precision (making few errors) and strong recall (identifying the most relevant instances).

$$\begin{aligned} F1 \ Score= & 2 \times \frac{Recall \times Precision}{Recall + Precision} \end{aligned}$$
(22)
$$\begin{aligned} Error= & 1-Accuracy \end{aligned}$$
(23)

Kappa. (\(\kappa\)) It provides a measure of the improvement in performance of the classifier compared to a classifier that randomly guesses based on the frequency of each class. Cohen’s Kappa coefficient is always less than or equal to 1. Values less than or equal to zero indicate that the classifier is ineffective.

Matthews correlation coefficient (MCC). The Matthews correlation coefficient is a more dependable statistical measure that produces a high score only when the prediction performs well in all four categories of the confusion matrix. It is responsible for the dataset’s proportion of positive and negative elements.

Proposed precise chlorosis severity assessment system

The steps that comprise the proposed framework are illustrated in Fig. 1. The stages of the PQCSAF are discussed in the following sections.

Preprocessing

All leaf images are taken in a controlled environment with a consistent dark-colored background. A preprocessing step eliminates the backdrop and retains only the leaf area as the region of interest. The preprocessing stage has three main steps: global thresholding, mask construction, and leaf extraction. The Green channel \(\hat{G}\) is extracted from the original leaf with background color image \(\hat{I}\) and Otsu’s thresholding59 is applied to estimate \(\hat{T_{h}}\) for the elimination of the background. The \(\hat{G}_{binary}\) image is generated using image binarization method based on the \(\hat{T_{h}}\) where the pixel value which is greater or equal than the \(\hat{T_{h}}\) are considered as 1 otherwise 0.

Fig. 1
Fig. 1
Full size image

The flow diagram of the proposed framework.

Evolutionary superpixel

The primary goal of superpixeling is the delineation of diverse color regions based on the level of chlorosis manifested in the leaf. Consequently, the parameters of the initial SLIC algorithm have been fine-tuned through an evolutionary methodology. This evolutionary strategy aids in the automatic selection of suitable parameter values in accordance with the textural patterns of the diseased leaf by reducing the overall standard deviation in color distribution. In this study, four commonly utilized bio-inspired algorithms, namely Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Ant Colony Optimization (ACO), and Cuckoo Search (CS) have been employed to facilitate evolutionary superpixels. Among these algorithms, the approach based on Cuckoo search demonstrates superior performance compared to the others. The implementation details of the Cuckoo search-based strategy have been detailed in the current section, and the analysis of the outcomes is provided in the section “Performance analysis of evolutionary superpixel”. The steps for optimizing superpixels using the Cuckoo search method are outlined in the algorithm 2. Here, the cuckoo is represented in the form of \(ck_{n}=\left\{ k,m,r\right\}\) and initialized within a given range as mentioned Table 3 at Line-1. The Cuckoo search algorithm is a metaheuristic algorithm developed by Yang and Deb in 200960. The algorithm mimics the brood parasitic behaviour of specific cuckoo species. Cuckoo species like Ani and Guira cuckoos follow an aggressive reproduction strategy. It lays eggs in the host bird’s nest and destroys the host bird’s egg to increase the hatching probability of its eggs. It visits one nest to another by L\(\acute{e}\)vy flights. The recent review of the Cuckoo search algorithm by Guerrero-Luis et al.61 studied the effectiveness of the Cuckoo search algorithm in diverse domains, including clustering, scheduling, parameter estimation, image analysis, and other various areas. The SLIC algorithm has three parameters: number of superpixels \(\left( k\right)\), weighting factor \(\left( m\right)\), and radius \(\left( r\right)\)34. The number of superpixels should be chosen so that pixel variance is low within superpixels. The weighting factor, m, determines how color and spatial information are balanced. A high k-value signifies the superpixel to have a more regular and smoother form. If one region is less than r, it will be united with a neighboring region. The textural pattern of a diseased leaf varies depending on the stage of the disease. As a result, it is critical to provide suitable values for all three factors to obtain a compact superpixel capable of calculating the lesion area and stage precisely. Here, the Cuckoo search algorithm has been utilized to optimize the value of k, m, and r. Here, the summation of the standard deviation of color components of all superpixels has been considered as the objective function as mentioned in Eq. 24. Initial nests are randomly generated within the ranges of k, m, and r. The range of all there parameters are given in Table 3. From each solution, superpixels are generated using SLIC method. After the generation of the superpixel-based image \(i_{spx}\) and evaluated by the objective function, all solutions are ranked based on fitness. The best cuckoo is calculated from the sorted solutions. New set solution is generated according to the best cuckoo with l\(\acute{e}\)vy flight as given in Eqs. 25 and 26. A fraction of the nests are randomly deleted with discovering probability \(p_{a}\) to bring diversification in the population. All the modified solutions are again used to generate and evaluate superpixel-based images. Finally, the new best cuckoo is compared with the previous one and updated if the new one is better than the previous one. It iterates up to maximum iteration \(\left( M_{t}\right)\). The Cuckoo search-based optimized parameters finding is outlined in Algorithm 2. Sample healthy leaves and various types of unhealthy leaves are shown in Fig. 4. The optimized value of k, m and r for each of the samples shown in Fig. 4 are given in Table 4.

$$\begin{aligned} f= & \sum _{i=1}^{k}\sum _{j=1}^{c}i_{spx}{\left( i,j\right) }\end{aligned}$$
(24)
$$\begin{aligned} x_{i+1}= & x_{i}+\alpha \oplus L\acute{e}vy\left( \lambda \right) \end{aligned}$$
(25)
$$\begin{aligned} L\acute{e}vy~\left( u\right)= & t^{-\lambda } \quad \text {where} \quad \left( 1<\lambda \le 3\right) \end{aligned}$$
(26)
Algorithm 2
Algorithm 2
Full size image

Cuckoo-Search-Optimized-SLIC.

Table 3 Parameters’ settings for the cuckoo search used in the proposed method.

Multi-swarm cuckoo search for feature selection

The original Cuckoo search algorithm was a single swarm. In this work, a multi-swarm Cuckoo search has been proposed for feature selection. The total feature set is partitioned into four groups with equal length. Initially, the four subswarms are created to select the feature for four groups, which means each subswarm will take care of each feature group. After a fixed iteration, the master swarm will be constructed from the subswarm. From each subswarm, a fixed number of the cuckoo is represented by a real number between 0 and 1. The general architecture of the proposed multi-swarm Cuckoo search strategy has been shown in Fig. 2, and detailed steps of the algorithm are also mentioned in the Algorithm 3. The following steps are involved in the algorithm.

Initialization. In this phase, the four subswarm are randomly initialized by 0 or 1 with length k. 1 indicates that the particular feature is selected, and 0 indicates the feature is not included in the selected set. So, all subswarms are randomly initialized. However, the master swarm is still not initialized. It is initialised after the g number of iterations because after each subswarm will be executed g in parallel. After the g iterations end, the first \(C^{r5}_sub\) solutions are selected to form the master swarm.

Communication Strategy. In the proposed architecture, two types of communication strategies have been taken. After a fixed interval of iteration, the subswarms return a set of best-performing cuckoos that are performing best up to that iteration in their subswarm. The arrow from the subswarm mentions this to the master swarm in Fig. 2. The symbol \(\bullet\) represents a candidate solution within both the subswarm and the master swarm. After obtaining such a subset, the master swarm is formed and begins executing a fixed iteration. The master swarm is responsible for identifying the set of features that can achieve high accuracy in classifying the four types of superpixels. Information exchange among individual subswarms is also permitted in the multi-swarm framework. Periodically, the subswarms share their few solutions, which helps the algorithm maintain a balance between exploration and exploitation while identifying optimal or near-optimal feature subsets. Here, the term “information exchange” refers to the transfer of binary-encoded feature subsets, where each binary string represents a selected feature set. This periodic communication is illustrated by arrows between the subswarms in Fig. 2. According to the traditional Cuckoo Search rules, there is a probability that the host bird will discover some of the cuckoo nests in each iteration, enabling each subswarm to explore its respective sub-search space independently. Similarly, the master swarm employs the same strategy to explore the combined feature space effectively.

Fig. 2
Fig. 2
Full size image

Multi swarm cuckoo search architecture.

Algorithm 3
Algorithm 3
Full size image

Multi-swarm parallel binary Cuckoo search for feature selection.

$$\begin{aligned} S\left( x^{k}_{i}\left( t \right) \right)= & \frac{1}{1+e^{-{x^{k}_{i}\left( t \right) }}}\end{aligned}$$
(27)
$$\begin{aligned} S^{k}_{t+1}= & {\left\{ \begin{array}{ll} 1, & \text {if } S\left( x^{k}_{i}\left( t \right) \right) > \sigma .\\ 0, & \text {otherwise}. \end{array}\right. }\end{aligned}$$
(28)
$$\begin{aligned} Fitness= & (1-accuracy) \end{aligned}$$
(29)

Quantification of chlorosis through severity index

The chlorosis severity score, \(\zeta\), as proposed by us, is the weighted average of the scores concerning all the superpixels in the population as shown in Eq. 30. Here, \(k^{1}\) denotes the count of the Type-1 superpixels coming from the leaf. Similarly, \(k^{2}\), \(k^{3}\), and \(k^{4}\) denote the counts of Type-2, Type-3, and Type-4 superpixels, respectively, from the leaf. Now, each Type of superpixels is assigned a weight \(\omega _{j} \in \left\{ 1, 2, 3, 4\right\}\), where \(j \in \left\{ 1, 2, 3, 4\right\}\). The severity index calculation does not consider a superpixel with a size of less than seventy (70) pixels. The severity index \(\zeta\) will lie in the range [1, 4]. The value of \(\zeta\) towards 1 indicates less or negligible chlorosis effect, whereas a higher value of \(\zeta\) indicates the presence of a higher degree of chlorosis. We can use Eq. 30 to collectively measure the chlorosis severity for a random group of plant leaves. Instead of individual leaves, a severity score can be measured using Eq. 31 where \(N_{sp}\) denotes the total number of sample leaves taken from the field.

$$\begin{aligned} \zeta= & \frac{\left( \omega _{1} \times k^{1}+\omega _{2}\times k^{2}+ \omega _{3}\times k^{3}+\omega _{4} \times k^{4} \right) }{\left( k^{1} + k^{2} + k^{3} + k^{4}\right) }\end{aligned}$$
(30)
$$\begin{aligned} \zeta= & \frac{\left( \omega _{1}\sum _{i=1}^{N_{sp}}k_{i}^{1}+\omega _{2}\sum _{i=1}^{N_{sp}}k_{i}^{2}+\omega _{3}\sum _{i=1}^{N_{sp}}k_{i}^{3}+\omega _{4}\sum _{i=1}^{N_{sp}}k_{i}^{4} \right) }{\left( \sum _{i=1}^{N_{sp}}k_{i}^{1}+\sum _{i=1}^{N_{sp}}k_{i}^{2}+\sum _{i=1}^{N_{sp}}k_{i}^{3}+\sum _{i=1}^{N_{sp}}k_{i}^{4} \right) } \end{aligned}$$
(31)
Table 4 Optimized parameters applicable for SLIC for some of the samples (representative examples) as decided by Cuckoo search.
Fig. 3
Fig. 3
Full size image

Preprocessing of images.

Results and discussion

This section provides a comprehensive analysis of the experimental findings for each stage of the PQCSAF.

Leaf area extraction

The leaf dataset that is utilized here has a fixed background. Hence, a preprocessing step has been applied to remove the background and keep only the leaf area as the region of interest. The preprocessing stage has three main steps: global thresholding, mask construction, and leaf extraction. The picture shown in Fig. 3a is obtained from the dataset, while the mask displayed in Fig. 3b is constructed using the previously discussed thresholding approach. Fig. 3c represents the ultimate extracted leaf area employed for the superpixeling process in the subsequent phase.

The classification problem in this work has been designed as a multi-class classification problem. Classification models were created using DT, KNN, SVM, and MLP for the work. The models were tested with a k-fold cross-validation strategy. We considered \(k=40\) for the experiment.

Performance analysis of evolutionary superpixel

The parameters of the SLIC algorithm have been optimized using an evolutionary technique and applied to the extracted leaf area. The CS-optimized SLIC outputs of Sample-1 to 8 are illustrated in Fig. 4. The visual results show that the superpixels should be adaptable to correctly capture the lesion depending on the affected region and yellowing. Indeed, the outputs of Fig. 4 are produced by the optimized parameter values in Table 4. The efficacy of the CS in optimizing the superpixel parameters has been compared to that of the other three metaheuristic algorithms: Genetic Algorithm (GA), Particle Swarm Optimization (PSO), and Ant Colony Optimization (ACO). Figure 5a displays the Convergence curve of GA, PSO, ACO, and CS algorithms for optimizing parameters for sample-6. The performance of the CS algorithm surpasses that of the other three methods. Figure 5b shows the convergence curve for all the samples 1 to 8. Based on the graph, it is clear that when the leaf has a minimal color variation (either a small affected area or less variation in yellowing), the convergence curve does not show significant improvement with each iteration. Conversely, in the opposite scenario, the convergence graph significantly improves as the iterations progress.

Fig. 4
Fig. 4
Full size image

Superpixels by Cuckoo search optimized SLIC.

Fig. 5
Fig. 5
Full size image

Convergence plots (#iterations vs fitness value) of optimized superpixels a (for the Sample-06 shown in Fig. 4) for various methods and b (for all the samples shown in Fig. 4) for Cuckoo search.

Performance analysis of multi-swarm cuckoo search for feature selection

The efficacy of the multi-swarm Cuckoo search for feature selection has been discussed in this section. The Table 5 provides a comparative analysis of five algorithms (GA, PSO, ACO, CS, and MSCS) in terms of their performance in “Fitness” and “Feature” over the iterations. It has been noted that CS achieves the greatest fitness value of 0.0046, while MSCS attains the value of 0.0095. Therefore, MSCS does not surpass the CS in terms of fitness value, nor does it surpass the GA, PSO, or ACO. However, MSCS has the lowest standard deviation of 0.00037, indicating it produces very consistent results in terms of fitness. MSCS has obtained 73 features, the lowest compared to other algorithms. The standard deviation of MSCS is 9, the second smallest value. Figure 6a and b illustrate the convergence curve and the number of selected features for each method, respectively. It is clear from the discussion that CS has the highest fitness performance, but its feature efficiency is only modest. The performance of MSCS is balanced and characterized by the utmost efficacy in feature selection, very low variability in results, and reasonable fitness values.

Table 5 Performance of MSCS in feature selection with other different algorithms.
Fig. 6
Fig. 6
Full size image

a Convergence curve; b No of selected features.

Performance analysis of classification on selected features from affected superpixels

An initial analysis is conducted on a collection of 408 original features derived from four superpixel categories. The DT, KNN, SVM, and MLP classifiers achieved average classification accuracy of \(87.10\%\), \(91.00\%\), \(94.50\%\), and \(96.80\%\) correspondingly for the category of superpixels, as indicated in Table 6. The MLP has demonstrated superior performance to the other three classifiers when evaluated using multiple metrics on the original feature set. The evaluation measures are outlined in Table 6. The remaining part of this section will investigate classifier behavior following the implementation of five evolutionary feature selection methodologies.

The classifiers DT, KNN, SVM, and MLP are trained and tested using the features obtained by GA, PSO, ACO, CS, and MSCS. In addition, the classification is conducted on the components generated by the PCA when applied to the original feature set. The classification result is assessed using the following performance metrics: Accuracy, Error, Sensitivity, Specificity, and Precision. The average performance of the DT classifier on all selected feature sets for classifying four categories of superpixels is shown in Fig. 7a. DT obtains \(87.8\%\) accuracy based on the ACO selected feature set, indicating it is the most effective in correctly classifying instances. At the same time, the lowest accuracy obtained by DT is based on PCA components. Hence, the error rate is lowest in the case of ACO and highest in PCA. The DT obtains \(87.8\%\), \(95.93\%\), and \(87.87\%\) for Sensitivity, Specificity, and Precision, respectively, based on ACO-obtained features. However, PCA-based performances by DT on the previous three parameters are again the lowest compared to the other methods. Therefore, the ACO selected features set performs very well with the DT classifier. At the same time, DT obtains the lowest performance with PCA. GA, PSO, CS, and MSCS perform well but do not outperform ACO. GA and CS perform slightly better than PSO and MSCS in most metrics. Figure 7b displays the mean performance of the KNN classifier across all chosen feature sets for categorizing four types of superpixels. The KNN classifier achieves a maximum accuracy of \(91.3\%\) when utilizing GA based feature selection. On the other hand, the lowest accuracy of \(88.9\%\) is obtained when employing PCA. The performance of KNN using different feature selection methods falls within the range of the lowest and highest accuracy achieved. That suggests the level of accuracy is relatively consistent. The error rate associated with GA-based methods is the lowest, at \(8.7\%\), while PCA-based methods have the highest error rate, at \(11.1\%\). The GA-based and ACO-based models exhibit the highest sensitivity, with values of \(91.3\%\) and \(91.1\%\), respectively. The sensitivity of PCA-based methods is the lowest, at \(88.9\%\). The ACO-based technique achieves the maximum specificity, with a value of \(97.03\%\), while the PCA-based method has the lowest specificity, with a value of \(96.3\%\). The remaining technique similarly exhibits a high level of specificity, falling within the performance range of ACO and PCA. When using PCA, the precision of the KNN classifier is lowest at \(89.42 \%\) and maximum at \(91.43\%\) when using a GA-based technique. The GA is the most successful approach for the KNN classifier, achieving the best accuracy and lowest error rates. Additionally, it demonstrates good sensitivity, specificity, and precision. PCA exhibits lower performance in this context. Furthermore, ACO and PSO are also formidable competitors. To categorize four different kinds of superpixels, Fig. 7c shows the average performance of the SVM classifier across all selected feature sets. The obtains the highest accuracy at \(95.0\%\) MSCS-based method and the lowest at \(93.1\%\) with PCA. Error is also minimal with the MSCS-based method. SVM also achieves a high sensitivity of \(95.0\%\), a specificity of \(98.33\%\), and a precision of \(94.99\%\). Similarly to the MSCS-based method, SVM’s efficacy in sensitivity, specificity, and precision is comparable to that of the other evolutionary feature selection techniques. MSCS-based feature selection is the most dependable method for the SVM classifier in this particular condition. MSCS may be the preferred option to maximize the performance of their SVM classifier; however, GA, PSO, and ACO also exhibit comparable performance to MSCS. The average performance of the MLP classifier in categorizing four categories of superpixels across all selected feature sets is illustrated in Fig. 7d. MLP classifier has obtained the highest accuracy of \(97.60\%\) and minimum error of \(2.4\%\) using the MSCS-based method, whereas using the ACO-based method, MLP has obtained the lowest accuracy of \(93.7\%\) and highest error of \(6.3\%\). Using the MSCS-based technique, the MLP classifier regularly achieves the greatest sensitivity value of \(97.6\%\), the highest specificity of \(99.2\%\), and the best precision of \(97.62\%\). The MLP classifier, using the ACO-based technique, obtains the lowest performance in sensitivity of \(93.7\%\), specificity of \(97.9\%\), and precision of \(93.69\%\). Additionally, other methodologies are strong competitors in terms of a few parameters. The above classification results analysis, which includes all feature selection strategies, indicates that the MLP with an MSCS-based strategy consistently performs well under average conditions in all five parameters: accuracy, error, sensitivity, specificity, and precision. A more comprehensive analysis of the performance of MLP classification has been shown in Fig. 8. The performance of MLP with five evolutionary algorithms-based feature selection strategies for the average case is illustrated in Fig. 8a. The five metrics have been considered: number of features, accuracy, precision, recall, F1 score, and MCC. The MSCS algorithm utilizes the minimum number of features, precisely 73, whereas the GA algorithm employs the maximum number of features, 206. The quantity of features significantly influences the classifier’s performance. The MSCS-based technique has a maximum accuracy of \(97.6\%\), while the CS method achieves an accuracy of \(95.6\%\). The accuracy of ACO is \(93.7\%\), which is the lowest of the options. The average classification accuracy with MSCS is notable, using minimal features, indicating the effectiveness of feature selection. MSCS exhibits the highest precision (\(97.62\%\)), suggesting that it generates minimal false positives. The precision of ACO is the lowest at \(93.69\%\). Recall signifies the actual positives correctly identified instances. MSCS achieved the maximum recall (\(97.6\%\)), while ACO achieved the lowest (\(93.7\%\)). The harmonic mean of precision and recall is the F1 score, which balances the two. MSCS has the highest F1 score of 0.9759, suggesting robust stability between precision and recall. ACO has the lowest F1 score of 0.9367. MCC is a metric that evaluates the quality of classifications in the context of true and false classifications. The MSCS-based method has the highest MCC of 0.97, which generates the most dependable classification results. The lowest MCC is 0.92 for the ACO-based strategy. The performance of MLP for the average case and individual superpixel category is illustrated in Fig. 8b. The MLP classifier demonstrates high accuracy in all categories, with the highest accuracy observed in Type - 3 (\(99.2\%\)) and the lowest in Type - 2 (\(94.8\%\)). The performance is robust, as evidenced by the average accuracy of \(97.6\%\). The error rates are inversely proportional to the accuracy rates. Type 3 has the lowest error rate at \(0.8\%\), while Type-2 has the highest at \(5.2\%\). The average error rate is a mere \(2.4\%\). The sensitivity values demonstrate the classifier’s high effectiveness in distinguishing positive instances, which is consistent with the accuracy values. The sensitivity values are highest for Type-3 (\(99.2\%\)), lowest for Type-2 (\(94.8\%\)), and average (\(97.6\%\)). The classifier performs exceptionally well in reliably identifying negative instances, with consistently high specificity values over \(98\%\) for all types. Type-1 has the maximum level of specificity, measuring at \(99.87\%\), while the Average demonstrates a specificity of \(99.2\%\). The precision values for all types are consistently high, showing that the MLP classifier has a very low rate of false positive errors. Type-1 exhibits the highest level of precision, at \(99.60\%\). Conversely, Type-3 has the lowest precision, at \(95.75\%\). The average precision value is \(97.62\%\). Furthermore, the comprehensive and detailed information on classifier performances for average and individual categorization with different feature selection strategies for assessing metrics is provided in Table 7. Overall, MSCS-selected features achieve the highest accuracy (\(97.6\%\)), precision (\(97.62\%\)), recall (\(97.6\%\)), F1 score (0.9760), and MCC (0.97) compared to other approaches. This performance implies that MSCS selects features effectively, leading to better classification performance with only 73 features in the average case and for each category. Moreover, MSCS is the most effective and efficient feature selection technique for the MLP classifier in this superpixel categorization task.

Fig. 7
Fig. 7
Full size image

Average classification performance of different classifiers on selected features by various feature selection strategy.

Fig. 8
Fig. 8
Full size image

Specific performance of MLP classifier with various parameters.

Table 6 Performance of different classifiers on complete dataset of 1000 samples of superpixels without feature selection.
Table 7 Performance of different classifiers based on the selected features by various bio-inspired algorithms on complete dataset of 1000 samples of superpixels.

Ablation study with four classifiers applied to original and MSCS feature sets

A comparative investigation of classification performance was performed across four classifiers: DT, KNN, SVM, and MLP, utilizing both the original and MSCS-selected feature sets. The initial feature set had 408 features; however, the MSCS approach reduced this to only 73 features. The evaluated performance criteria comprise accuracy, sensitivity, specificity, and precision. The original feature set for the DT classifier marginally surpasses the MSCS-selected features, exhibiting a minor decline in all measures when employing MSCS. On the other hand, KNN demonstrates consistent performance across all feature sets, with just a slight decrease in sensitivity and precision when utilizing MSCS, which suggests robustness to feature reduction. The SVM demonstrates a marginal enhancement in accuracy, sensitivity, and specificity when utilizing MSCS-selected features, indicating that the model advantages from a more concise and relevant feature set. MLP consistently attains superior performance across all measures, with enhancements observed when employing MSCS-selected features. Accuracy increased from \(96.8\%\) to \(97.6\%\), sensitivity from \(96.8\%\) to \(97.6\%\), specificity from \(98.93\%\) to \(99.2\%\), and precision from \(96.8\%\) to \(97.62\%\). The performance comparison graphs between the original and MSCS-selected feature set for each classifier are illustrated in Fig. 9, with (a) denoting DT, (b) KNN, (c) SVM, and (d) MLP. It indicates that the MSCS method not only reduces computational complexity but also enhances classification performance.

Fig. 9
Fig. 9
Full size image

Performance comparison of different classifiers on original and MSCS-selected feature sets.

Severity score generation

The four categories of superpixels, classified by their degree of severity, are illustrated in Fig. 10. Finally, the severity score is computed using Eq. 30. Figure 11 shows twelve diseased images and their disease severity scores. The computation of the severity score \(\zeta\) using Eq. 30 for the examples shown in Fig. 11 is explained in detail in Table 8. Furthermore, Fig. 12 displays the scores of the initial twenty-five sample leaf pictures on a scale ranging from 1 to 4. The overall severity score \(\zeta\) for all 25 sample leaves, as seen in Fig. 12, has been computed using Eq. 31 and yielded a value of 2.0185.

Fig. 10
Fig. 10
Full size image

Superpixel category: a, e Type-1, b, f Type-2, c, g Type-3, d), h: Type-4.

Fig. 11
Fig. 11
Full size image

Severity scores of some sample leaves.

Table 8 Details of severity score \(\zeta\) calculation of Fig. 11 samples.
Fig. 12
Fig. 12
Full size image

Severity scores of twenty-five sample leaf images.

The above investigation is conducted using a dataset that is of high quality and readily accessible. Due to the current unavailability of a substantial dataset for the experiment, the study is grounded in conventional machine-learning techniques. A suitable deep-learning model that functions successfully with a minimal number of datasets may be considered for this problem. The method achieves a high degree of precision with the standard model; it will likely achieve even greater precision with deep learning models. When the strategy is combined with deep learning using deep features, it may be possible to identify more severity grades (more than four).

Comparative analysis of optimized SLIC and other segmentation methods

In this section the proposed approach is analyzed with other methods.

Comparison with global thresholding based methods

The efficacies of the two methodologies proposed by Golzari Oskouei et al. (2021)62 and Song et al. (2024)63 are given herein. Golzari Oskouei et al. (2021) presented the Cluster-weight and Group-local Feature-weight learning in the Fuzzy C-Means (CGFFCM) method for color image segmentation. They developed an enhanced fuzzy c-means (FCM) clustering technique for color image segmentation. This study demonstrates that CGFFCM applies automated cluster weighting and group-local feature weighting to address FCM’s sensitivity to initial cluster selection and uniform feature weighting. Song et al. (2024) introduced a Modified Snake Optimizer (MSO) to enhance color image segmentation for crop disease leaf, utilizing Kapur’s entropy as the objective function. This study presents dynamic parameter modification, enhanced location updates, Lvy flight for evading local optima, and a balancing method to increase global and local search efficacy. Figure 13 displays the sample images with their respective ground truth images. The segmentation outcomes of the two approaches are illustrated in Fig. 13. The visual analysis indicates that Golzari Oskouei et al. (2021)62 outperform Song et al. (2024)63. Furthermore, both approaches can segment the lesion region to a considerable extent. This signifies the adequate performance of both methods in segmenting diseased plant images. Further study is conducted based on the various segmentation quality parameters in Table 9. The parameter values also show good segmentation quality, notably for the approaches proposed by Golzari Oskouei et al. (2021)62.

Fig. 13
Fig. 13
Full size image

First column: four sample images; second column: segmentation ground truth; third column: segmentation using Song et al. (2024); fourth column: segmentation using Golzari Oskouei et al. (2021).

Table 9 Performance comparison of MSO63 and CGFFCM62 on segmentation quality metrics.

Based on the preceding segmentation results in Fig. 13, a severity score can be calculated by determining the Euclidean distances between clusters representing diseased areas and healthy leaves. The distinct clusters in the segmentation output signify the varying levels of severity. Thus, the severity score (\(\zeta\)) is computed according to Eqs. 32– 34. In this context, \(\vert S_{i} \vert\) denotes the number of pixels in cluster \(S_i\), u signifies the total number of clusters, and \(\vert L_{area} \vert\) indicates the total leaf area (in pixels). The set H represents a set of five healthy leaf samples (shown in Fig. 14). The distance between \(S_{i}\) and H is denoted by the Euclidean distance (\(\delta _{i}\)). Here, we have considered the mean of five distances (concerning five healthy samples) as \(\delta _{i}\). Finally, \(\zeta\) is calculated using Eq. 34, which incorporates the individual leaf severity \(d_{n}\) and the maximum leaf severity \(d_{max}\).

$$\begin{aligned} d_{n}= & \sum _{i=1}^{u} \delta _{i} \times \frac{|S_{i}|}{|L_{area}|}\end{aligned}$$
(32)
$$\begin{aligned} \delta _{i}= & distance(H,S_{i}) \end{aligned}$$
(33)
$$\begin{aligned} \zeta= & \frac{d_{n}}{d_{max}}\times 100 \% \end{aligned}$$
(34)

While this method may be straightforward, it has two limitations: firstly, determining the optimal number of color threshold levels, which can vary according to the severity stages as the color changes, and secondly, the challenge of precise detection, where minor color variations may not accurately reflect changes in severity stages. The detection of tiny color patches is crucial for early disease identification. Segmentation quality for the four samples shown in Fig. 13 are presented in Table 9 and the estimated severities (of the four leaf samples) using these segmentation approaches are shown in Table 10. These leaf image samples show the challenges related to optimal segmentation. The corresponding Jaccard indices show this. Another sample is shown in Fig. 15, which shows that some segmentation strategies may not detect the disease area properly (light-shaded leaf areas are not detected as a separate segment in this example). In this context, we believed that capturing lesion regions with appropriate training-based classification using superpixels obtained from the lesions may enhance understanding of disease severity.

Fig. 14
Fig. 14
Full size image

Samples of healthy leaf images.

Fig. 15
Fig. 15
Full size image

Sample leaf image (with clear colored segments) vs Corresponding segmentation output using Golzari Oskouei et al. (2021).

Table 10 Euclidean distance based severity estimation from the segments.

Histogram matching based comparison

A further study has been conducted that focuses on histogram matching-based severity estimation. Here, the histograms are prepared from individual color channels of the healthy reference image and diseased leaves. The normalized histograms of healthy image and diseased image are presented \(H_{h}^c(i)\) and \(H_{d}^c(i)\) at intensity level \(i \in {\{0,...,255\}}\), for color channel \(c \in {\{R,G,B\}}\). In this context, three parameters, Chi-square distance, histogram intersection, and correlation coefficient, have been considered for histogram matching. The Chi-Square Distance quantifies the difference between the two histograms; lower values indicate more similarity; histogram intersection measures the overlap between the source and reference histograms; higher values indicate greater similarity in colour distribution and correlation coefficient: evaluates the linear relationship between the histograms; values closer to 1 indicate a substantial similarity. The calculation of the Chi-Square distance, histogram intersection, and correlation coefficient are provided in the Eqs. 3540 respectively. The \(cov(H_{d}^c, H_{h}^c)\) denotes covariance between the histograms of the diseased and healthy images for channel c. The \(\sigma _{H_{d}^c}\) indicates the standard deviation of the histograms of the diseased images for channel c.

$$\begin{aligned} \chi ^2_c= & \sum _{i=0}^{255} \frac{\left( H_{d}^c(i) - H_{h}^c(i)\right) ^2}{H_{h}^c(i) + \epsilon } \end{aligned}$$
(35)
$$\begin{aligned} \overline{\chi ^2}= & \frac{1}{3} \sum _{c \in \{R, G, B\}} \chi ^2_c \end{aligned}$$
(36)
$$\begin{aligned} \text {HI}^c= & \sum _{i=0}^{255} \min \left( H_{d}^c(i), H_{h}^c(i)\right) \end{aligned}$$
(37)
$$\begin{aligned} \overline{\text {HI}}= & \frac{1}{3} \sum _{c \in \{R, G, B\}} \text {HI}^c \end{aligned}$$
(38)
$$\begin{aligned} \rho _c= & \frac{\text {cov}(H_{d}^c, H_{h}^c)}{\sigma _{H_{d}^c} \cdot \sigma _{H_{h}^c}} \end{aligned}$$
(39)
$$\begin{aligned} \overline{\rho }= & \frac{1}{3} \sum _{c \in \{R, G, B\}} \rho _c \end{aligned}$$
(40)

After determining the distance of each channel, the k-means clustering technique is used to categorize them into visually identifiable three groups according to the disease impact. Four samples from each group are shown in Fig. 16. Visually, it is evident that the images in Fig. 16a–d depict leaves with minimal damage (Grade-1), while those in (e)–(h) exhibit more significant color distortion (Grade-2). The images in (i)–(l) show even more pronounced color changes (Grade-3). Table 11 presents the mean values of the parameters, Chi-square distance, histogram intersection, and correlation coefficient for individual channels for the images shown in Fig. 16. It is worth mentioning that when we use color-profile-distance-based grouping or categorization of the leaves, the image shown in Fig. 16i is grouped into Grade-3. However, this does not match with human perception. Our proposed superpixel-based estimation of severity for the leaf is 1.77 on a scale of 1 (lowest) to 4 (highest), which relates to the human perception with more perfection (see Fig. 17).

Fig. 16
Fig. 16
Full size image

Three grade of severity based on histogram matching: Grade-1: (a)–(d), Grade-2: (e)–(h), Grade-3: (i)–(l).

Fig. 17
Fig. 17
Full size image

Leaf image shown in Fig. 16i with severity 1.77.

Table 11 Comparison of metrics across severity grades.

Comparative study with varying number of superpixels

Figure 18 displays the original sample leaf photos with the results obtained with different numbers of superpixels. The number of superpixels varies with the values of 30, 60, 90, 120, and 150. The outcomes of the superpixels obtained using Cuckoo search optimization are also presented. The results indicate that increased superpixels lead to more precise lesion area detection. The optimal superpixels may identify the lesion more precisely than the superpixel range of 30 to 150. In sample-1 of Fig. 18, when the number of superpixels is 30, there exists a significant number of superpixels showing color variation over the optimal case.

Table 12 provides a comparative analysis of SLIC superpixel segmentation across four example images at varying superpixel counts, utilizing three parameters: average compactness, average superpixel size, and average color variation. An optimal number of superpixels is determined for each sample image, signifying the best trade-off balance among these criteria. Compactness often decreases with increased superpixels, signifying more uniformly formed superpixels. Reduced compactness values in the optimum rows indicate that precise boundary adherence, essential for identifying irregular lesion forms, is most effectively attained with relatively high superpixel counts. The average size of the superpixels decreases as the number of superpixels increases. The optimum results show reduced superpixel sizes, which may enable more accurate differentiation between diseased and healthy regions. The variation in color decreases as the number of superpixels increases, indicating that higher segmentation levels provide more homogeneous regions. The optimum result shows minimal color variation, which is crucial for identifying areas impacted by minor color differences. This level of detail is crucial for detecting the early stage of leaf disease. The optimal values presented in the table provide a balanced segmentation outcome that provides a precise evaluation of disease severity. This system provides a more accurate, automated assessment of crop disease severity by ensuring precise localization and segmentation.

Fig. 18
Fig. 18
Full size image

Output for varying the number of superpixels: row-1: original Sample \(1-4\), row-2: 30, row-3: 60, row-4: 90, row-5: 120, row-6: 150, row-7: 300, row-8: optimum result (357, 416, 351 and 380).

Table 12 Comparison of SLIC superpixel results across different samples and superpixel counts.

We observed that as the number of superpixels (k) increases (from 30 to 150), the average superpixel size decreases, color variation is often reduced, and compactness improves, as shown in Table 12. When k is fixed at 300, and the optimization is limited to two parameters (m and r), further improvements are observed in color variation for most instances. We also observe improvements in compactness for many instances. Hence, determining an appropriate value of k for an arbitrary leaf image is challenging; however, the results suggest that increasing k may enhance the metrics. However, if we designate k as too large, it results in over-segmentation. Here, minimal color variations indicate enhanced consistency within each superpixel, which is essential for precisely capturing local image characteristics. Furthermore, decreased superpixel size allows for improved granularity, although decreased compactness values signify more uniform and coherent superpixel configurations, which promote improved alignment with lesion borders and assist in accurate segmentation. As shown in Table 12, optimal performance across all three parameters (k, m, and r) consistently yields better results in terms of color variation and compactness. Consequently, the individual superpixels more closely conform to the actual lesion structure, enabling a more accurate and robust representation of disease-affected areas that leads to more precise severity estimation.

The execution time for the experiments was measured with the image size of \(1000 \times 700\) pixels, and after experimenting with 10 sample images, the average execution time has been reported. The method based on Golzari Oskouei et al. (2021) took an average of 2.5587 minutes to run, while the method based on Song et al. (2024) took about 11.1637 minutes. For the SLIC algorithm, the average execution times with different numbers of superpixels were: 0.0867 mins (30 superpixels), 0.0917 mins (60), 0.0908 mins (90), 0.1003 mins (120), and 0.1051 mins (150). For the proposed method, the optimization step took an average of 21.5637 minutes, while the final execution time using the optimal SLIC parameters was 0.1565 minutes. It is important to note that the time for the feature selection step is not included here, as it is performed only once for the entire dataset.

Conclusion & future work

In summary, the main objective of this research is to develop a chlorosis detection framework with high precision. PQCSAF measures the degree of chlorosis of the disease leaf. The color co-occurrence matrix-based features were utilized with the newly proposed multi swarm cuckoo search based feature selection strategy. It utilizes DT, KNN, SVM, and MLP-based supervised models for classification. Various performance measures validate all the experimental results. It is seen that MLP has obtained the maximum accuracy of \(97.60\%\). The experiment demonstrates that utilizing superpixel-based lesion localization on the leaf results in the development of disease severity scores with a high level of accuracy. Furthermore, using a multi-swarm-based evolutionary feature selection technique enhances the system’s ability to adapt to textural changes on the leaf caused by different diseases. The proposed approach has been implemented, and the chlorosis that impacts the leaves of Pongamia pinnata has been assessed. This dataset has been selected due to the extensive variation of color caused by the disease. In addition, datasets with other plant leaves and disease conditions may be considered for further analysis to adjudge the method’s applicability in general. The model is robust, and it may be applied to monitoring crops as well as forests.

The research may be extended to include more types of leaf diseases by using multiple objective functions that account for disease-specific information. The model’s performance may be further analyzed through hyperparameter adjustment and other classifiers. The optimization pipeline is subject to additional enhancement. Further investigations could look into the potential benefits of implementing the proposed framework directly on AI-enabled edge devices. This can broaden the research scope to develop an effective device for the Fog computing layer in on-field agricultural applications with reduced reliance on the Internet.