Abstract
Sustainable agriculture is an approach that involves adopting and developing agricultural practices to increase efficiency and preserve resources, both environmentally and economically. Jute is one of the primary sources of income grown in many countries. At this stage, increasing efficiency in jute production and protecting it from pests is essential. Detecting jute pests at an early stage will not only improve crop yield but also provide more income. In this paper, an artificial intelligence-based model was suggested to detect jute pests at an early stage. In this developed model, two different pre-trained models were used for feature extraction. To improve the performance of the developed model, the features obtained using the DarkNet-53 and DenseNet-201 models were combined. After this stage, the metaheuristic Mountain Gazelle Optimizer (MGO) was used, allowing the developed model to work faster and achieve more successful results. Feature selection was carried out using MGO; thus, more successful results were obtained with fewer, more compelling features. The proposed model was compared with six different models and five different classifiers accepted in the literature. In the developed model, 17 different jute pests were detected with 96.779% accuracy. The accuracy value achieved in the developed model is promising in successfully detecting jute pests.
Similar content being viewed by others
Introduction
The continuation of life in nature largely depends on insects. Insects pollinate plants, ensuring the continuation of plant diversity1. In this way, sustainable agriculture enables living beings to obtain food and maintain a high quality of life. The goals of sustainable agriculture include various elements, such as nutrition in Fig. 1, protecting the future, protecting quality food, and ensuring food safety. In addition, many economic, cultural, and social improvements are among the targets2.
Ensuring and protecting plant diversity is a crucial aspect of sustainable agriculture. This is facilitated by establishing technological infrastructures, such as plant protection and early disease diagnosis systems, which utilize advanced technologies like the Internet of Things (IoT)3. While beneficial insects contribute to plant diversity, harmful insects also pose a threat4. These destructive insects must be combated through agriculture to minimize their damage. Eliminating harmful insects through such control measures preserves plant diversity, quality, and sustainability and prevents economic losses. Figure 2 shows images of the jute plant as it is grown and its image transformed into a product after processing.
Now that the effects of harmful insects have been observed, it is necessary to determine how to combat them effectively and take action promptly. The fight against pests is being continued with internationally accepted methods such as integrated pest management (IPM) and area-wide (AW) pest management5. Many different methods can combat insects that fall into the pest category, such as biological, chemical, pesticide use, heat application, or cooling. The critical point during the struggle phase is memorizing the correct method. An incorrectly determined method may even cause an increase in harmful insects, let alone getting rid of them. While some insects increase in hot environments, others survive and reproduce by adapting to cold environments. In addition, a pesticide used for drug-resistant pests can create a more comfortable breeding ground instead of destroying them. To prevent such situations, the insect species to be combated must be determined correctly. This correct detection phase is the first step and the most critical point in the fight against harmful insects. Then, deciding which method is most suitable for pest control and starting the fight begins. An incorrectly determined control method may fail to destroy the harmful insect species and cause additional damage to the plant species.
Considering the harmful effects of insects, the importance of accurate and rapid detection becomes evident. Two different methods can be followed to detect harmful insects. The first step is to get expert support and determine the type of support. This method may not be accessible for every region. Disadvantages, such as a lack of experts and remote areas, may limit the usability of this method for general use. The second method is recognition using artificial intelligence systems. Considering the success of artificial intelligence systems, especially in image processing, it is possible to identify species with insect images.
Identification of harmful insect species, which has been transformed into a problem that can be solved by artificial intelligence, will provide an essential step towards ensuring biodiversity and sustainable agriculture. At this point, the issues that are intended to be solved step by step with this study can be listed as follows;
-
Collection of images of harmful insects
-
Separation into categories by pre-processing
-
Creating the appropriate artificial intelligence model
-
Extracting features
-
Implementation of optimization
-
Classification
-
Deciding on the type of pest and its performance.
Artificial intelligence-based methods are frequently used to classify plant diseases. Classification of diseases on tomato6, tea7, rice8, and apricot9 leaves are just a few of them. However, more studies are needed to detect harmful insects in plants. Therefore, this study we conducted is original regarding the subject and the methods used. Artificial intelligence offers significant potential in data analysis and pattern recognition techniques, especially in solving complex problems like detecting harmful insects. Such methods have become an essential tool for increasing productivity in agriculture by creating accurate diagnosis and early warning systems.
Related work
When research on sustainable agriculture is examined, it is revealed that effects such as infertile soils, environmental pollution, quality deterioration, and decreased profitability endanger sustainable agriculture10. Just as there are studies on pest detection, there are also studies on plant diseases11. Ebrahimi et al.12 proposed an image-processing-based model for automatic pest detection. They determined the threshold value for pixel segmentation by using the histogram graph created with the Otsu method to remove the background from the image obtained after applying the gamma operator. From the new images obtained, hue (H), saturation (S), and intensity (I) components were obtained using HSI (hue, saturation, intensity) color decomposition as features. They made a classification using SVM to detect the pest.
In recent years, deep learning algorithms have been frequently used in studies to identify pest species. Trufelea et al.13 determined the interspecific similarity and intraspecific variability of four Phyllocephalidae insect species with the SSD deep learning method, the extended convolution layer of VGG-16. As a performance indicator, they reached an intersection over the union (IoU) rate of 70.2%.
Maican et al.14 used transfer learning to detect pests with the MobilNet-SSD-v2 they used. For pest detection on corn plants, they selected five insect species (Coleoptera), four of which are harmful (Anoxia, Diabrotica, Opatrum, and Zabrus) and the beneficial ones (Coccinella sp.). A mAP of 0.892 performance was achieved with MobileNet-SSD-v2-Lite. have achieved value.
Meena et al.15, in their study on the detection of diseases in plants and the detection of harmful insects with ready-made deep learning models, studied 38 different diseases, nine different harmful insect species, and four different weed species. They achieved disease detection with 99.06% accuracy with DensNet, weed detection with 97.56% accuracy by fine-tuning the Hyperparameter Search 2D layer, and 88.75% accuracy in pest detection with MobileNet.
Yang et al.16 created a dataset to detect pests in tea plants. Then, they designed a model with the Yolov7-small deep learning model. They correctly detected the pests on the tea plant with a 93.2% accuracy rate with the model they tested with a dataset of 782 images.
Vedhamuru et al.17 proposed a Feature Pyramid Dilation Residual Convolution neural network (FDPRC) to detect pests in tomato plants. After training and testing the model, they detected pests with a 93.46% accuracy rate.
In addition to using images, some academic studies classify insects by analyzing their vocalizations. One of these studies is Dhanaraj et al. It was carried out using the deep learning technique proposed by18 using high accuracy with Multi-Layer Perceptron (MLP) models. In addition to their proposed model, the dataset they tested with models such as DenseNet, VGG16, YOLO, and ResNet50 reached a superior accuracy rate compared to other models, with a 99.78% accuracy rate with MLP.
Talukder et al. created the jute pest dataset and conducted their studies using the relevant dataset. The jute plant is essential because it is the most used plant fiber after cotton and is an important source of agricultural income in many countries19. Like other agricultural products, jute can be attacked by pests, and as a result, economic losses may occur. This paper suggests a model for detecting and classifying pests harmful to the jute plant.
Contribution and novelty
This study used a 17-class dataset to detect jute pests. For faster model development, instead of training a model from scratch, features obtained from six different architectures accepted in the literature were classified using five different classifiers. Among these architectures, DarkNet-53 and DenseNet-201 achieved the highest performance, and these two models served as the base for the proposed model. The features obtained in DarkNet53 and DenseNet201 models are concatenated to use different features of the images together. In this way, the developed model achieved more successful results because various features of the same image were combined. At the same time, only feature extraction was performed with DarkNet53 and DenseNet201; these models were not trained. This allowed this phase to be completed quickly. However, these combined features have the potential to contain redundant features, and due to their high dimensionality, efficient feature selection is required to improve classification performance. For feature selection, this study uses Mountain Gazelle Optimizer (MGO), a novel meta-heuristic optimization technique proposed by Abdollahzadeh et al.20. Inspired by the social behavior and group dynamics of mountain gazelles, MGO is employed. Meta-heuristics can provide successful and efficient solutions when they balance the exploration and exploitation phases, which are crucial for efficiently navigating the search space and converging on optimal solutions. The four different mechanisms in the base design of the MGO provide a strong balance between the exploration and exploitation phases. To verify the performance of MGO, Abdollahzadeh et al.20 conducted experiments on a total of 52 benchmark functions, including seven unimodal, 16 multimodal, 29 CEC2017 functions, and seven real-world engineering optimization problems, in their original study. Then, they compared the results obtained from MGO with nine widely used metaheuristic optimization algorithms, namely Particle Swarm Optimization (PSO), Grey Wolf Optimizer (GWO), Farmland Fertility Algorithm (FFA), Tunicate Swarm Algorithm (TSA), Gravitational Search Algorithm (GSA), Moth-Flame Optimization Algorithm (MFO), Sine Cosine Algorithm (SCA), Multi-Verse Optimizer (MVO) and Whale Optimization Algorithm (WOA), and showed that MGO gives more successful results than the others. The advantages mentioned above and the fact that MGO has not yet been applied to agricultural image classification problems, especially for pest detection, are the primary motivations for choosing MGO in this study. The study eliminated unnecessary features by using MGO for feature selection, and an accuracy rate of 96.779% was achieved. The study reached the specified accuracy value without performing the data augmentation step. Not performing the data augmentation step allowed our developed model to run faster. When the studies in the literature were examined, more successful results were achieved in the model we proposed without performing the data augmentation step.
Organization of paper
The second part of this proposed study on classifying pests harmful to jute plants includes information about the dataset, CNN architectures, and classifiers used. In the third section, metaheuristic MGO and the developed method are discussed. The fourth part includes information about the performance values calculated after applying the model to the images. The fifth part gives the performance results obtained from applying the model to the images in detail. The final section assesses the proposed model’s scientific contribution, results, and potential for future studies.
Backgrounds
This section examines the dataset used in the paper and pre-trained models and classifiers used in the study to detect jute pests. The summary flow chart of the paper is given in Fig. 3.
Jute pest dataset, deep models and classifiers
Jute plants, like many plants, can be damaged by insects. Fighting against these insects is essential regarding the jute plant’s production of quality fiber and its economic contribution. Therefore, detecting these pests will increase the quantity and quality of the product. Using web resources, Talukder et al.19,21 brought together the insects that damage the jute plant. They completed the dataset by removing possible light, environment, perspective, and image distortions in the collected pest images. Since the number of images in the test folder was minimal in this study, the images in the train and test folders were combined. As a result of these operations, the dataset consists of a total of 6828 images. Then, 20% of this data is reserved for testing. There are 17 different pest classes in the created dataset. The images used in the study are in .jpg format and are of various sizes. Since pre-trained models automatically resized the images, original-size images were used. The species in the dataset are given in Fig. 4.
Figure 5 visually shows the total number of images and sample pest types for all classes in the dataset.
Six deep models accepted in the literature were used to compare the performance of our hybrid model developed for detecting Jute pests. In addition, the two models that showed the highest performance among these models were used as the basis for the model we created. ShuffleNet22, MobileNetV223, Vgg1924, EfficientNetb025, Darknet5326 and DenseNet20127 are the models used in the study. Five different classifiers were used to classify the features obtained in the relevant deep and developed models. At this stage, Support Vector Machine (SVM)28, Fine Tree (FT)29, Linear Discriminant (LD)30, k-nearest Neighbors (KNN)31, and Ensemble Subspace KNN (ES)32 these are classifiers used in the classification process.
The mountain gazelle optimizer (MGO)
Abdollahzadeh et al.20 introduced the Mountain Gazelle Optimizer (MGO) algorithm, a novel metaheuristic optimization technique derived from mountain gazelles’ social behavior and group dynamics. The optimization process in MGO consists of four main stages: Territorial Solitary Males, Maternity Herds, Bachelor Male Herds, and Migration to Search for Food. Implementing these four primary stages concurrently executes the algorithm’s exploitation and exploration phases.
Within the parameters of the search space, MGO begins with an initial population generated at random. A gazelle is a candidate solution, and every member of the original population is a candidate solution. Each gazelle (X(i)) in the population has the potential to join one of the three groups mentioned in the main stages above. An adult male gazelle represents the global optimum solution. The four main stages of the algorithm are presented in detail below.
Territorial solitary males (TSM)
Male gazelles establish solitary territories when they reach adulthood. These territories are defined as in the following equations:
In Eq. 1, maleg is defined as a vector that contains the position of the best global candidate solution. Rndi1 and Rndi2 are random integers with values of 1 or 2. BH is the coefficient vector for the young male herd, calculated by Eq. 2. F is calculated using Eq. 3. Cr is a coefficient vector used to increase the algorithm’s searchability, as calculated by Eq. 4.20
In Eq. 2, Xrnda denotes a randomly selected candidate solution in the population at interval rnda. Mpr represents the randomly generated average population (N/3). Here, N is the population number of gazelles. Also, rnd1 and rnd2 are random numbers between 0 and 1. In Eq. 3, N1 represents a random number from a standard distribution. MaxIter is the maximum number of iterations and Iter is the number of current iterations. Exp is the exponential function. In Eq. 4, rnd3 and rnd4 are randomly selected numbers between 0 and 1. N2, N3, and N4 are random numbers in the normal range and dimensions of the problem. a is calculated as in the following equation20:
Maternity herds (MH)
Maternal herds play a crucial role in the life cycle of mountain gazelles as they give birth to strong male gazelles. Male gazelles may also contribute to the possession of females and the birth of gazelles. This is formulated using the following equation20:
In Eq. 6, BH is calculated as in Eq. 2. C2r and C3r are randomly chosen co-efficient vectors calculated independently by Eq. 4. Rndi3 and Rndi4 are random integers with values of 1 or 2. Xrnd is a randomly selected gazelle from the population.
Bachelor male herds (BMH)
As male gazelles mature, they seek to establish their territories and search for female gazelles, which leads to conflicts with other male gazelles. This behavior can be expressed mathematically as follows20:
In Eq. 7, X(t) represents the position of the gazelle at the current iteration. Rndi5 and Rndi6 are random integers with values of 1 or 2. maleg is defined as a vector that contains the position of the best global candidate solution. BH and Cr are calculated by Eqs. 2 and 4, respectively. The following equation calculates D:
In Eq. 8, X(t) represents the position of the gazelle at the current iteration. maleg is defined as a vector that contains the position of the best global candidate solution. rnd5 is a random number between 0 and 1.
Migration to search for food (MSF)
Mountain gazelles migrate to find food sources. This is expressed by the following equation20:
In Eq. 9, U and L are the search space’s upper and lower bounds, respectively. rnd6 is a random number between 0 and 1.
The main stages of the algorithm, TSM, MH, BMH, and MSF, are applied to all gazelles to generate new generations. In each iteration, gazelles are sorted according to their fitness values, and gazelles with promising values are preserved in the population, while gazelles with poor values are eliminated. The algorithm’s flowchart is presented in Fig. 6.
Proposed model for automatic detection of jute pests
This study proposes a CNN and MGO-based model for detecting Jute Pests. Feature extraction was performed with six different models to be used as a base in the proposed model. The features obtained from six different models were classified into five different classifiers. Since the highest accuracy values were obtained from DarkNet-53 and DenseNet-201 models, DarkNet-53 and DenseNet-201 models were used as a base in the study. Then, feature selection was performed with MGO, and the SVM classifier, which achieved the highest performance among five different classifiers, was used as the base in the proposed model. The flow chart of the developed model is presented in Fig. 7.
First, the features of the images in the dataset are obtained from the “conv53” layer of DarkNet53 and the “fc1000” layer of DenseNet201. There are 6828 images in the dataset; for each image, both CNN models give 1000 features. In other words, two different 6828 × 1000 feature matrices are obtained from both CNN models. These two matrices are then combined to obtain a single 6828 × 2000 feature matrix.
Feature selection is then performed with MGO. The MGO population is initially created with N gazelles (candidate solutions). Each candidate solution is a randomly generated vector of 2000 dimensions between 0 and 1, derived from the number of features obtained from CNN. A feature will be chosen for classification if its value in the candidate solution is higher than 0.5. This feature will be disregarded otherwise. The SVM classifier is then used to determine each candidate solution’s fitness value in the MGO. The classifier considers features that match dimensions larger than 0.5 in the candidate solution. The accuracy value at the end of the classification process determines the fitness value of the candidate solution. Once the maximum number of iterations has been achieved, the candidate solution with the highest fitness value is considered the issue solution.
Performance metrics
While analyzing the classification results of the developed model, performance metrics are calculated, and the obtained results are interpreted to decide whether the model is suitable or not. Thanks to the calculated metrics, the correct prediction rates of the model and the incorrect classification rate and deficiencies are determined. While the calculations are made for each class and the rates are evaluated separately, a general evaluation is made by calculating the general rates. While calculating the performance values, True-Positive (TP), True-Negative (TN), False-Positive (FP), and False-Negative (FN) values are used. These values are given in the confusion metric and are calculated from the columns corresponding to the true and predicted columns.
The formulas in Table 1 are used to calculate the basic criteria for Accuracy, Precision, recall, and F1-Score. The calculations are first made separately for each class, and then the overall accuracy rate, Overall Precision (Precision—macro Mean and Micro Mean), Overall Sensitivity (Recall—macro Mean and Micro Mean), and Overall F1-Score values are calculated.
Application results
The experimental results in this study, which aimed to detect jute pests automatically, were acquired in the Matlab environment. The same dataset and metrics were used to compare models and classifiers under the same conditions. Twenty percent of the Jute Pets dataset’s jute pest images were used for testing, and the remaining eighty percent were kept aside for training. In the first stage, six distinct pre-trained models were used to create feature maps of the images in the dataset to compare the model’s performance we suggested. Five distinct classifiers were used to categorize these acquired attributes.
In the second step, the feature maps obtained in the DarkNet53 and DenseNet201 models were combined, and this combined feature map was classified similarly to five different classifiers. The reason for combining the features obtained in the DarkNet53 and DenseNet201 models in this step is to get the most successful results in these two architectures and to use these architectures as the basis for the suggested model. The flow chart of the results obtained in this study is given in Fig. 8.
Results were obtained in five different classifiers using feature extraction methods
In this section, feature extraction is performed using popular deep models frequently used in the literature. Feature extraction in MobileNetV2 architecture was obtained from Logits layers, node_202 in ShuffleNet architecture, conv53 in DarkNet53 architecture, efficientnet-b0|model|head|dense|MatMul in EfficientNetb0 architecture, fc8 in VGG19 architecture, and FC1000 in DenseNet201 architecture. In each architecture, 1000 features were extracted, and these features were classified into five different classifiers. The relevant results are given in Table 2.
When Table 1 is examined, the features extracted with CNN architectures in classifiers were classified. The most successful results were obtained after combining the features obtained with DenseNet201 + DarkNet53. The most successful classifiers in this step were LD and SVM.
The comparative graph of the performance values obtained for all models is given in Fig. 9.
Figure 10 shows the confusion matrix obtained when the feature maps of DenseNet201 + DarkNet53 architectures are given to the LD classifier.
Figure 11 shows the confusion matrix obtained when the feature maps of DenseNet201 + DarkNet53 architectures are given to the SVM classifier.
When Figs. 10 and 11 are examined, in Fig. 10, out of 1366 test images, 1286 were predicted correctly, while 80 were mispredicted. In Fig. 11, out of 1366 test images, 1283 were predicted correctly, while 83 were mispredicted.
Results of the proposed model
In the method we proposed for automatically detecting Jute malware, feature extraction was first performed using DenseNet201 and DarkNet53 models. To improve the performance of our suggested model, we merged feature maps from two distinct architectures. This combined various aspects of the same image. The suggested model’s feature selection was then performed using MGO, and SVM was employed as a classifier. The initial population number of MGO was set as 10, the maximum number of iterations was set as 100, and 10 independent experiments were performed. Table 3 shows the mean accuracy, max accuracy, and standard deviation values obtained in 10 independent experiments.
Figure 12 gives the confusion matrix of the best result obtained in 10 independent experiments.
When Fig. 12 is examined, 1322 of the 1366 test images were predicted correctly, while 44 were mispredicted. The performance values of the proposed model calculated separately for each class are given in Table 4. Separate calculations were made for 17 classes in the table.
The overall performance values calculated for the proposed model are given in Table 5.
The study’s limitation is that the proposed model is tested using a single public dataset. Another area for improvement is that no other dataset on the subject exists in the literature. Our goals in the future include creating a more detailed dataset by including experts in the field and developing a mobile application on the subject.
Conclusion
Artificial intelligence-based methods have recently gained prominence in many fields. Therefore, this study used it to detect jute pests automatically. Since jute is a significant income source, the product’s quality and quantity must be high. In addition, it is tough to detect jute pests visually, and it may be too late to take some precautions. A hybrid structure is used in the proposed model to detect Jute pests automatically. Pre-trained models were used instead of training a model from scratch so that the proposed model could work quickly. In addition, feature merging was performed to improve the proposed model’s results. The metaheuristic Mountain Gazelle algorithm was preferred for our model to work faster. When the optimized feature map was classified in the SVM classifier, an accuracy value of 96.779% was achieved. This value shows that our model can be used to detect jute pests.
Data availability
A publicly available dataset was used in the article. https://archive.ics.uci.edu/dataset/920/jute+pest+dataset.
References
Isbell, F. et al. Benefits of increasing plant diversity in sustainable agroecosystems. J. Ecol. 105(4), 871–879. https://doi.org/10.1111/1365-2745.12789 (2017).
Velten, S., Leventon, J., Jager, N. & Newig, J. What is sustainable agriculture? A systematic review. Sustainability (Switzerland) 7(6), 7833–7865. https://doi.org/10.3390/su7067833 (2015).
Buja, I. et al. Advances in plant disease detection and monitoring: From traditional assays to in-field diagnostics. Sensors 21(6), 1–22. https://doi.org/10.3390/s21062129 (2021).
Calvo-Agudo, M. et al. IPM-recommended insecticides harm beneficial insects through contaminated honeydew. Environ. Pollut. https://doi.org/10.1016/j.envpol.2020.115581 (2020).
Hendrichs, J., Kenmore, P., Robinson, A. S., & Vreysen, M. J. B. Area-wide integrated pest management (AW-IPM): Principles, practice and prospects. Area-Wide Control of Insect Pests: From Research to Field Implementation, pp. 3–33, https://doi.org/10.1007/978-1-4020-6059-5_1 (2007).
Cengil, E. & Çınar, A. Practice, and undefined 2022, Hybrid convolutional neural network based classification of bacterial, viral, and fungal diseases on tomato leaf images, Wiley Online LibraryE Cengil, A ÇınarConcurrency and Computation: Practice and Experience, 2022•Wiley Online Library, vol. 34, no. 4, Feb. https://doi.org/10.1002/cpe.6617 (2021)
Yücel, N. & Yıldırım, M. Classification of tea leaves diseases by developed CNN, feature fusion, and classifier based model. Int. J. Appl. Math. Electr. Comput. 11(1), 30–36 (2023).
Yucel, N. & Yildirim, M. Automated deep feature fusion based approach for the classification of multiclass rice diseases. Iran J. Comput. Sci. 7(1), 131–138 (2024).
Kang, X. et al. Assessing the severity of cotton Verticillium wilt disease from in situ canopy images and spectra using convolutional neural networks, Elsevier X Kang, C Huang, L Zhang, M Yang, Z Zhang, X Lyu The Crop Journal, 2023 Elsevier, Accessed: Apr. 23, 2024. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S221451412200263X.
Reganold, J. P., Papendick, R. I. & Parr, J. F. The rewards are both environmental and financial. Sci. Am. Div. Nat Am. Inc. 262(6), 112–121 (1990).
Ahila, A., Prema, V., Ayyasamy, S. & Sivasubramanian, M. An enhanced deep learning model for high-speed classification of plant diseases with bioinspired algorithm. J. Supercomp. 80(3), 3713–3737. https://doi.org/10.1007/s11227-023-05622-4 (2024).
Ebrahimi, M. A., Khoshtaghaza, M. H., Minaei, S. & Jamshidi, B. Vision-based pest detection based on SVM classification method. Comput. Electron. Agric. 137, 52–58. https://doi.org/10.1016/j.compag.2017.03.016 (2017).
Trufelea, R., Dimoiu, M., Ichim, L. & Popescu, D. Detection of harmful insects for orchard using convolutional neural networks. UPB Sci. Bull., Series C: Electr. Eng. Comp. Sci. 83(4), 85–96 (2021).
Maican, E., Iosif, A. & Maican, S. Precision corn pest detection: Two-step transfer learning for beetles (Coleoptera) with MobileNet-SSD. Agriculture (Switzerland) https://doi.org/10.3390/agriculture13122287 (2023).
Meena, S. D., Susank, M., Guttula, T., Chandana, S. H. & Sheela, J. Crop yield improvement with weeds, pest and disease detection. Proced. Comput. Sci. 218(2022), 2369–2382. https://doi.org/10.1016/j.procs.2023.01.212 (2022).
Yang, Z., Feng, H., Ruan, Y. & Weng, X. Tea tree pest detection algorithm based on improved Yolov7-Tiny. Agriculture (Switzerland) 13(5), 1–22. https://doi.org/10.3390/agriculture13051031 (2023).
Vedhamuru, N., Malmathanraj, R. & Palanisamy, P. Features of pyramid dilation rate with residual connected convolution neural network for pest classification. SIViP 18(1), 715–722. https://doi.org/10.1007/s11760-023-02712-x (2024).
Dhanaraj, R. K., & Ali, M. A. Deep Learning-Enabled Pest Detection System Using Sound Analytics in the Internet of Agricultural Things. p. 123, https://doi.org/10.3390/ecsa-10-16205 (2024).
Talukder, M. S. H. et al. JutePestDetect: An intelligent approach for jute pest identification using fine-tuned transfer learning. Smart Agricult. Technol. 5(12), 100279. https://doi.org/10.1016/j.atech.2023.100279 (2023).
Abdollahzadeh, B., Gharehchopogh, F. S., Khodadadi, N., & Mirjalili, S. Mountain Gazelle Optimizer: A new Nature-inspired Metaheuristic Algorithm for Global Optimization Problems, Advances in Engineering Software, vol. 174. https://doi.org/10.1016/j.advengsoft.2022.103282 (2022).
URL: https://archive.ics.uci.edu/dataset/920/jute+pest+dataset (2024).
Zhang, X., Zhou, X., Lin, M., & J. Sun. ShuffleNet: An extremely efficient convolutional neural network for mobile devices. in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 6848–6856, https://doi.org/10.1109/CVPR.2018.00716 (2018).
Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L. C. MobileNetV2: Inverted residuals and linear bottlenecks. in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 4510–4520, https://doi.org/10.1109/CVPR.2018.00474 (2018).
Mascarenhas, S. & Agarwal, M. A comparison between VGG16, VGG19 and ResNet50 architecture frameworks for Image Classification. Proceed. IEEE Int. Conf. Disrupt. Technol. Multi-Discipl. Res. Appl., CENTCON 2021, 96–99. https://doi.org/10.1109/CENTCON52345.2021.9687944 (2021).
Patel, C. H., Undaviya, D., Dave, H., Degadwala, S., & Vyas, D. EfficientNetB0 for Brain Stroke Classification on Computed Tomography Scan. in Proceedings of the 2nd International Conference on Applied Artificial Intelligence and Computing, ICAAIC 2023, pp. 713–718, https://doi.org/10.1109/ICAAIC56838.2023.10141195 (2023)
Varnima, K. E., & Ramachandran, C. Real-time gender identification from face images using you only look once (yolo). in Proceedings of the 4th International Conference on Trends in Electronics and Informatics, ICOEI 2020, pp. 1074–1077, Jun. https://doi.org/10.1109/ICOEI48184.2020.9142989. (2020)
Jaiswal, A., Gianchandani, N., Singh, D., Kumar, V. & Kaur, M. Classification of the COVID-19 infected patients using DenseNet201 based deep transfer learning. J. Biomol. Struct. Dyn. 39(15), 5682–5689 (2021).
Hearst, M. A., Dumais, S. T., Osuna, E., Platt, J. & Scholkopf, B. Support vector machines. IEEE Intell. Syst. Appl. 13(4), 18–28 (1998).
Azar, A. T. & El-Metwally, S. M. Decision tree classifiers for automated medical diagnosis. Neural Comput. Appl. 23(7–8), 2387–2403. https://doi.org/10.1007/S00521-012-1196-7 (2013).
Tharwat, A. E., Gaber, A., Ibrahim, T. & Hassanien, A. Linear discriminant analysis: A detailed tutorial. AIC 30(2), 169–190 (2017).
Peterson, L. E. K-nearest neighbor. Scholarpedia 4(2), 1883 (2009).
Zhang, Y., Cao, G., Wang, B. & Li, X. A novel ensemble method for k-nearest neighbor. Patt. Recogn. 85, 13–25 (2019).
Acknowledgements
This work was supported by the National Research Foundation of Korea(NRF) grant funded by the Korea government(MSIT) (No. RS-2023-00218176) and the Soonchunhyang University Research Fund. This work was supported through Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R719), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.
Author information
Authors and Affiliations
Contributions
Conceptualization, Software, Methodology, Validation, Original Writeup Draft: Soner Kiziloluk, Mücahit Karaduman, Serpil Aslan, Muhammed Yildirim, Muhammad Attique Khan. Project Administration, Methodology, Supervision, Funding, Review Original Draft, and Visualization: Muhammad Attique Khan, Fatimah, Yunyoung Nam.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Kiziloluk, S., Karaduman, M., Aslan, S. et al. A novel feature fusion and mountain gazelle optimizer based framework for the recognition of jute pests in sustainable agriculture. Sci Rep 15, 18148 (2025). https://doi.org/10.1038/s41598-025-00642-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-025-00642-x