Introduction

Breast cancer is the most frequently detected type of cancer in females worldwide. Studies indicate that 2 in 5 women are prone to breast cancer throughout their lifetime. Breast cancer1 has been the second leading cause of death in women since 2013.Similarly, multiple studies have shown that women have a greater chance of surviving from breast cancer if the disease is detected and treated early2,3. The Nitric Oxide generated by cancer cells is responsible for the local temperature rise, which has been documented in several investigations4,5. The vasodilation process caused by nitric oxide causes breast tissue to heat up over its typical temperature because it interrupts the regular blood vessel flow in the area. An increase in skin temperature has been linked to deep breast malignant tumors. The relationship between an object’s energy output and its temperature has been described by Stefan Boltzmann’s Law6. As a result, the distribution of body temperature can be determined by observing infrared radiation.

Hence screening methods are essential in the detection and diagnosis of early stages of cancer. Ultrasound, Positron Emission Tomography (PET), Computed Tomography (CT), Magnetic Resonance Imaging (MRI), and Mammography are some of the medical imaging techniques that have been heavily used to detect early signs of breast cancer. Ultrasound7 is another important breast screening method to reduce a woman’s lifetime radiation exposure. However, due to noise levels and the technician’s expertise, this method has some limitations. This technique fails to detect micro-calcifications and deep breast tissue. Despite the stressful and painful process involved in getting a mammogram, it has become the most extensively utilized diagnostic imaging technology for breast cancer screening8,9. Mammography uses X-ray radiation, which is harmful to human tissue. Furthermore, mammography has accuracy issues with small tumors, with a 4%–34% percent false negative ratio. Another diagnostic method commonly used for breast cancer screening is magnetic resonance imaging (MRI). MRI has a number of disadvantages compared to mammography and ultrasound, such as a higher probability of false positives and a longer data gathering time10.

Physiological or vascular alterations in the breasts are seen in thermographic11 pictures. The use of thermal imaging for monitoring purposes is risk-free since it is nonionizing, noninvasive, painless, passive, and real-time. This technique is safe for pregnant and nursing women since it does not employ ionizing radiation. It’s especially helpful for young women since their tissues are thick, making X-ray screening more challenging and potentially dangerous. Authors12,13suggest that CADs might benefit doctors in making diagnoses by highlighting anomalies or diseases in medical imaging.

However, the segmentation method has been used to many problems with great success. Identifying aberrant areas or categorizing the elements in digital pictures requires the use of image segmentation14. The fundamental objective of segmentation is to ease the analysis process by dividing an image into discrete parts according to predetermined criteria such as texture, color, or brightness. After segments have been extracted, visualization, detection, identification, and quantitative analysis are often performed. Thresholding has also been widely employed in the last decade to aid clinicians in the diagnosis process through the medical image processing automation. The complexity of structures with similar characteristics, low contrast, noise conditions, and unclear boundaries typical of medical images15,16,17 make interpretation and analysis of image details difficult even though many works on automatic and semiautomatic segmentation have been done.

Medical digital images need to be segmented for a variety of purposes, including pre- and post-operative evaluation, diagnosis, and therapy planning. Despite thermography screening’s limited clinical utility, enhancing its application and automated detection is crucial to bringing it to widespread clinical acceptability18,19.

An eight-step semi-automatic procedure for segmenting thermal images was proposed by the authors. According to the authors, the inframammary fold and edge detection issues account for the vast majority of the malfunctions. After having a human expert segment the photos by hand, the authors employed a fuzzy rule-based system to identify the segmented regions20. A hybrid version of the snake optimization (SO) algorithm combined with the opposite-based learning (OBL) mechanism. Then, it is used to segment CT liver images.The imaged thresholding technique is used for this purpose and employed SO-OBL to identify the best, or optimal, thresholding values for gray levels in real cases with liver disease images21 .

In22, an algorithm was introduced that uses automated thresholding and boundary detection to partition thermal breast images. Despite encouraging findings, it’s possible that the area of interest being discovered may not encompass all of the upper breasts. Therefore, there is need for further study into the creation of an automated approach that may aid in the diagnosis of breast cancer using thermograms and has a cheap computing cost and high degree of resilience. Thus, reliable thresholding of breast thermograms could be used to quickly assess clinical diagnoses. The swarm algorithm is used to analyze breast thermograms in this paper.

Utilization of swarm intelligence algorithms in the selection of multi-level thresholding segmentation of medical images is not a new concept. Several authors implemented various optimization algorithms to find the optimal threshold values for brain MRI image segmentation23,24,25,26 based on different objective functions such as Otsu’s between class variance and Kapur’s entropy etc. A multi-level image segmentation technique using HRO-based Renyi’s entropy algorithm for microscopic images of cement is proposed. Taking Renyi’s entropy as the objective function, HRO is used to search the optimal threshold values and compared with other optimization algorithms27 .

The authors used a novel swarm technique called Dragonfly Algorithm (DA)28 based on the Otsu method and Kapur entropy as objective functions on the DA in the designed methodology. The method presented uses the energy curve29,30,31 instead of the histogram to solve this issue. The energy function (EF) first described in32,33 is used to derive the energy curve. The EF determines the power of each pixel’s intensity level by considering its position. Energy curves offer advantages in terms of noise reduction, feature detection, spatial information representation, and integration with other image processing techniques. These advantages make energy curves particularly useful in applications requiring detailed and smooth analysis of intensity variations, such as medical imaging and advanced image processing tasks. However, histograms remain valuable for their simplicity and effectiveness in summarizing the global intensity distribution and for tasks like thresholding.

“The No-Free Lunch Theorem” states that there is no single modern optimization strategy that can effectively tackle all real-world optimization issues. Furthermore, since every real-world issue has a unique characteristic that can be compared to the actions of one specific optimization method. For example, in image processing edge and lane detection problems can be used to describe ant systems’ behavior toward pheromone deposits. This analogy can be applied to the route planning problem in robotics, which can likewise be represented as an agent following the scent of an unknown substance. An agent’s olfactory organ (nose) is filled with smell molecules that move through the air and into the specialized cells (receptors) that deliver the message to the brain for proper interpretation. This paper gives a detailed implementation and validation of a new metaheuristic algorithm termed the chaotic smell agent optimization with adaptive control strategy (ChBSAOACS) in light of these interactions between an agent and its surrounding environment.

From the literature, it is observed that SAO showed superior performance over the other algorithms on CEC benchmark functions and many authors have utilized Otsu and Kapur’s objective approaches for multilevel image thresholding using various optimization algorithms based on energy curve. An improved version of SAO34 algorithm has been developed and implemented for the optimal multilevel thresholding of breast thermographic images based on energy curve.

The important contributions of this paper are:

  • Chaotic Bidirectional Smell Agent Optimization with Adaptive Control Strategy is developed and implemented to multi-level thresholding for the first time.

  • To find the best thresholds for different breast thermographic images, ChBSAOACS algorithm maximizes the between class-variance and entropy described by Otsu and Kapur functions.

  • Qualitative and quantitative examination for eight different breast thermographic images using proposed approach.

  • A detailed validation of ChBSAOACS with other optimization algorithms proved the efficacy of the proposed approach.

Methodology

This section provides an overview of the optimization method for olfactory agents and shows how it may be implemented in multilevel thresholding scenario. In place of the image histogram, the chosen swarm and evolutionary algorithm learn the search space defined by the Otsu and Kapur methods, both of which are based on the energy curve.

Energy curve

Histogram-based multilevel thresholding is a widely employed method due to its convenience and effectiveness. An alternative to histogram-based segmentation is energy curve-based image segmentation35. The energy curve method, in contrast to the histogram method, maintains peaks and valleys in its plotted curve. In order to separate the objects in an image, segmentation algorithms must first find threshold values in the energy curve’s midpoint. Each valley is located between two nearby modes, and each mode stands in for a different feature of the image.

$$\begin{aligned}E_g=-\sum _{i=1}^{M}\sum _{i=1}^{N}\sum _{pq\in N_{ij}^{2}}b_{ij}b_{pq}+C\end{aligned}$$
(1)

Consider \(E_g \)is an EC of the image I. where \(I=\left( g_{ij},1\le i\le M,\;\;1\le i\le N\right\} \). \(g_{ij} \) is the imaging value of I at pixel location \(\left( i,j\right) \) defined under the interval \(\left( 0,\;L-1\right] \;where\;L=256 \). The energy of I is computed at grey level \(g\left( 0\le g\le L-1\right) b_{ij} \) are the elements of a two-dimensional matrix \(B_g.B_g=\left( b_{ij},1\le i\le M,\;\;1\le i\le N,\;b_{ij}=1,\;if\;g_{ij}>g;else\;b_{ij}=-1\;\right\} \). To ensure \(E_g>0 \) constant C is added to Eq. 1.Special correlation between adjacent pixels of an image I is \(N_{ij}^{d}=\left( \left( i+u,\;j+v\right) ,\left( u,v\right) \in N^{d}\right\} . \) In this paper d is considered a second-order system i.e., \(\left( u,v\right) \in \left( \left( \pm \;1,0\right) ,\left( 0,\pm 1\right) ,\left( 1,\pm 1\right) ,\left( -1,\pm 1\right) \right\} . \)

Otsu’s between class variance

Otsu’s method36 takes a bi-modal histogram and separates foreground and background pixels by computing the best threshold that minimizes the two classes’ combined spread (intra-class variance). For the multi-level thresholding expansion of the Otsu method, the term “Multi Otsu” was coined. Otsu presented a nonparametric thresholding technique that uses the maximum variance of the distinct classes as a criterion for segmenting the image.

nt thresholds are required to divide the original image into \(nt+1 \) classes for the multi-level technique. As a result, \(th\;=\;\left( {th}_1,\;{th}_2,\dots .{th}_{nt}\right\} \)encodes the collection of thresholds used for image segmentation. In this way, the grey value \(E_i \) of each pixel generates a probability \({PE}_i=\frac{E_i}{NP} \) where \({\sum _{i=1}^{NP}{{PE}_i}}=\;1\; \)and NP is the total number of pixels in the image.In order to calculate the variance \(\sigma ^{2} \) and the means \(\mu _k \) for each of the produced classes, the following formulae are used:

$$\begin{aligned}\sigma ^{2}={\sum _{k=1}^{nt}{\sigma _k}}={\sum _{k=1}^{nt}{\omega _k}}\left( \mu _k-\mu _T\right) \end{aligned}$$
(2)
$$\begin{aligned}&\mu _k=\sum _{i=th_k}^{th_{k+1}-1}\frac{kPE_i}{\omega _i\left( th_i\right) }\\ & where,\;\omega _k=\sum _{i=th_k}^{th_{k+1}-1}PE_i\end{aligned}$$
(3)

By optimizing the following equation, the best thresholds may be selected:

$$\begin{aligned} & f\left( th\right) =max\left( \sigma ^{2}\left( th\right) \right) ,\;0\le th_i\le L-1, \\& where\;\;i=1,2,\dots nt\;and\;L=256\end{aligned}$$
(4)

Kapur’s entropy

Kapur37 presented another nonparametric method for determining the best threshold values. It is based on the image histogram’s probability distribution and entropy. The approach seeks to determine the ideal ’th’ in terms of overall entropy. Entropy is a measure of an image’s separability and compactness between classes. In this view, entropy is maximized when the optimal ‘th’ values correctly separate the classes using Eq. 6.

$$\begin{aligned}f\left( th\right) =max\left( {\sum _k^{nt}{H_k}}\right) \end{aligned}$$
(5)
$$\begin{aligned}H_k={\sum _{i={th}_k}^{{th}_{k+1}-1}{\frac{PE_i}{\omega _k}}}{{ln}{\left( \frac{PE_i}{\omega _k}\right) }}\end{aligned}$$
(6)

Smell agent optimization (SAO)

It is through the sense of smell that most people get their first impressions of the world. The majority of living creatures can detect the presence of toxic chemicals in their surroundings utilizing their sense of smell. It’s natural to think about creating SAO with the help of human olfaction. However, as it turned out, olfaction is used by the majority of biological agents for the same main function in the process of hunting, mating, and avoiding danger. This caused a widespread smell agent to be developed as a result of an algorithm for maximizing efficiency38.

Sense of smell (olfaction)

Olfactory nerves, which are located in the nose, are responsible for allowing us to detect the scent of a material. Chemicals with molecular weights smaller than 300 Dalton are present in the smell, which diffuses from a source unevenly. Most agents need to be able to detect and distinguish these scent molecules. Among olfactory-capable agents, the process of olfaction is nearly identical. The ability to follow the scent plume is a crucial part of identifying remote odors. Several creatures, including humans, rely on this skill for survival. A person’s ability to identify potentially dangerous items such as iron tablets, household cleaners, and button batteries is based on their ability to detect the odor.

Steps of the SAO algorithm

The SAO has structured around three unique modes. Using the processes outlined in the previous paragraph, these modes are derived. Detecting odor molecules and determining whether or not to look for the source is the initial phase of operation for the agent. In \({2\text{nd}}\) mode, the agent uses the \({1\text{st}}\) mode’s choice to follow the smell molecules in search of the smell’s source. Using the third option prevents the agent from becoming confined within a small area if its trail is lost.

Sniffing mode

Since scent molecules tend to scatter in the direction of the agent, we begin the method by assigning them a randomly generated initial position. Using Eq. 7, we can find the total number of thresholds ‘d’ in the search space given the total number of smell molecules, ‘N’.

$$\begin{aligned} x_N^{d}=x_{min,d}+\left( x_{max,d}-x_{min,d}\right) *rand \end{aligned}$$
(7)

To find its optimal position in the search space, the agent uses Eq. 7 to build a position vector. \(x_{max,d} \) and \(x_{min,d} \) are maximum and minimum limits for the decision variable and rand () is a random number between 0 and 1 in case of SAO. A cubic map has been used to initialize the position vector based on Eq. 18 in the case of ChBSAOACS. It is determined that the molecules of odor will diffuse from the source/origin at an initial velocity determined by Eq. 8.

$$\begin{aligned}v_N^{d}=v_{min,N}+\left( v_{max,N}-v_{min,N}\right) *rand()\end{aligned}$$
(8)

Each molecule of a smell is a potential solution in the search space that may be used to solve a problem. These potential solutions (smell molecules) may be located using the location vector in Eq. 7 and the velocities of the molecules in Eq. 8. Since the smell molecules disperse in a Brownian fashion, their velocities may be updated using Eq. 12.

$$\begin{aligned}x_N^{d}\left( t+1\right) =x_N^{d}\left( t\right) +v_N^{d}\left( t+1\right) *\triangle t\end{aligned}$$
(9)

Since it is assumed that the agent optimizes in increments of one step at a time, t is set to 1. For instance, until the last iteration is achieved, the algorithms keep increasing the iteration by 1 until they reach the end of the process. Therefore, the new position of the odor molecules is given by Eq. 10.

$$\begin{aligned}x_N^{d}\left( t+1\right) =x_N^{d}\left( t\right) +v_N^{d}\left( t+1\right) \end{aligned}$$
(10)

Each smell molecule evaporates at a different rate, shifting its location in the search space at a rate proportional to its diffusion velocity. Since the molecules of smell travel in a path-dependent way to the agent’s position, Eq. 10.

$$\begin{aligned}v_N^{d}\left( t+1\right) =v_N^{d}\left( t\right) +v\end{aligned}$$
(11)
$$\begin{aligned}v=r_1\sqrt{\frac{3KT}{m}}\end{aligned}$$
(12)

In Eq. 12,\(\;v \) denotes the component of the velocity that is updated throughout time. Temperature and mass of smell particles, T and m, respectively; affect molecules’ momentum in the sense of smell. The normalizing constant for the sense of smell is denoted by the sign k. The technique is robust regardless of the starting parameters used for fragrance molecules, including temperature T and mass m. The ideas of mass and temperature (m and T) are derived from the ideal theory of gas. The values of m and T have been empirically determined to be 0.175 and 0.825, respectively, to facilitate efficient use of time. Eq. 10 tests the viability of the rearranged smell molecule’s potential binding areas. This means that we can now find the \(x_{best}^{d} \) agent (best fitness position) and exit sniffing mode.

Trailing mode

When an agent is tasked with locating the source of an unpleasant odor, this mode mimics the actions of the agent. If the agent is searching for a smell source and it detects a new place with a higher concentration of scent molecules than its present position, the agent will move to that location using Eq. 13.

$$\begin{aligned} x_N^{d}\left( t+1\right) &=x_N^{d}\left( t\right) +r_2*olf*\left( x_{gbest}^{d}\left( t\right) -x_N^{d}\left( t\right) \right) \\&\quad -r_3*olf*\left( x_{worst}^{d}\left( t\right) -x_N^{d}\left( t\right) \right)\end{aligned}$$
(13)

\(r_2 \) and \(r_3 \) are random numbers in the (0,1) range. Relatively speaking, both \(r_2 \) and \(r_3 \) penalizes the impact of olf on \(x_{gbest}^{d}\left( t\right) \) and \(x_{worst}^{d}\left( t\right) \).

Sniffing mode provides a way for the agent to decide its current position \(x_{gbest}^{d}\left( t\right) \) agent and the point with the worst fitness \(x_{worst}^{d}\left( t\right) \). This allows the agent to track the scent’s route more effectively. The relationship between exploration and exploitation in the algorithm is seen in Eq. 13. Olfaction ability is largely influenced by the size of one’s olfactory lobes, as well as one’s psychological and physical state, therefore the value of olf should be carefully chosen. Local searches benefit from low values of olf, while global searches benefit from high values of olf, since the former shows more strength in SAO’s ability to detect scents on a global search.

Random mode

Due to the discrete nature of smell molecules, their concentrations and intensities can change over time if they are spread out over a vast area relative to the search space. Because of this, trailing becomes more difficult and the scent is lost, making it difficult for an agent to follow. The agent may be unable to continue trailing at this moment, resulting in a local minimum. When this occurs, the agent enters a random mode, as follows:

$$\begin{aligned}x_N^{d}\left( t+1\right) =x_N^{d}\left( t\right) +r_4*SL\end{aligned}$$
(14)

Random number \(r_4 \) penalizes the value of SL by stochastically increasing the steplength of SL. If the agent loses its trail or the trailing mode is unable to get the greatest fitness or identify the smell source, the agent will take a random step using Eq. 14.

This means that the agent will constantly want to stay where the scent molecules are most concentrated throughout the route. The agent uses its knowledge of its present position, \(x_{gbest}^{d}\left( t\right) \) agent, and its worst positions, \(x_{worst}^{d}\left( t\right) \) agent, throughout the trailing process described in Eq. 14. Although the agent can only look in the feasible region because of the optimization problem. In the flowchart shown in Figure 1, ChBSAOACS’s comprehensive implementation is explained.

Fig. 1
figure 1

Implementation flowchart for ChBSAOACS algorithm.

The proposed ChBSAOACS

In this section, an improved version of the SAO algorithm is proposed, and the one-dimensional cubic map is introduced in SAO initialization and an adaptive (non-linear) parameter control strategy is also performed.

Chaotic Map

Chaos occurs frequently in nonlinear systems. The formula for the fundamental cubic map39 is as follows:

$$\begin{aligned}Z_{n+1}={\propto Z}_n^{3}-{\beta Z}_n\end{aligned}$$
(15)

When the chaos factors, denoted by \(\alpha \) and \(\beta \), \(\beta \) in the range [2.3,3], the cubic map is chaotic. The interval of the cubic map is (− 2, 2) for \(\alpha \) = 1, while the interval of the series is (− 1, 1) for \(\alpha \) = 4. The cubic map is also possible as:

$$\begin{aligned}Z_{n+1}={\rho Z_n(1-Z}_n^{2})\end{aligned}$$
(16)

Where the \(\rho \) denotes a tunable variable. In Eq. 18, the cubic map sequence is in (0, 1), and the ergodicity of the produced chaotic variable \(Z_n \)improves when \(\rho \) = 2.595.

During the search phase, the chaotic map can arbitrarily disperse the population of olfactory agents throughout the interval [0, 1]. In the proposed technique, the agents’ positions are first determined using a cubic map, with z (0) set to 0.315 to guarantee that the initialized interval lies within the range (0, 1).

Concept of bidirectional search

The bidirectional adjustment can enhance the algorithms’ searching performance by enabling the parameters to conduct local searches in both forward and backward directions given in Eq. 17.

$$\begin{aligned}\begin{array}{l}if\;[f(x+SL)\;<f(x)]\;\rightarrow \;x=x+SL\\ else\;[f(x-SL)\;<f(x)]\;\rightarrow x=x-SL\\ else\;no\;change\;in\;the\;present\;position\end{array}\end{aligned}$$
(17)

Where x, f(x) and S represents the current position of butterfly, objective fitness function value and step length respectively.

Adaptive control strategy

From Eq. 13, we may deduce that the olf contributes significantly to ChBSAOACS’s optimal solution finding performance. The value of olf should be carefully chosen since olfactory capability is mostly determined by olfactory lobe size, agent psychology and health40. So, whereas a low value of olf suggests weak olfaction, which favors local searching, a high value of olf shows greater olfaction, which prefers global searching of SAO. However, in standard SAO, olf = 0.75, and setting a to a constant does not allow for a good compromise between global and local search. As a result, we suggest the following method of adaptive parameter control:

$$\begin{aligned}olf\left( t\right) ={olf}_{min}-\left( {olf}_{max}-{olf}_{min}\right) .{{sin}{\left( \frac{\pi }{\mu }{\left( \frac{t}{maxit}\right) }^{2}\right) }}\end{aligned}$$
(18)

Parameter olf has the initial value and final value, \(\mu \) tuning parameter, and maxit is maximum number of iterations. The parameters used in this study were \(\mu \;=\;2 \) maxit = 200/800, \({olf}_{min}=0.1 \) and\({olf}_{max}=0.9 \). It’s important to remember that as the parameter gets bigger, the improvement technique starts to backfire.

Performance evaluation of ChBSAOACS

To demonstrate the effectiveness of the proposed ChBSAOACS in locating the global solution, it is applied to a large number of complicated test functions; sample test results for unimodal and multimodal functions from CEC14 and CEC17 are shown in Table 1. High Conditioned Elliptic, Discuss, Rosen Brocks, Weierstrass, Griewanks, Rastrigins’, Zakharov, Happy Cat, Ackley’s, power Functions, etc. pass the CEC14 tests, as do the planned ChBSAOACS. The performance of the SAO, DA, KH, GA, and PSO algorithms are compared with the test results. Table 1 displays the findings, which demonstrate that the ChBSAOACS performed better than the other algorithms in determining the global best fitness values for the designated test functions. Figure 2 depicts the plots of convergence for a few chosen functions. As can be seen in Figure 2, ChBSAOACS has outperformed the other algorithms when it comes to finding the optimum global fitness values.

Table 1 Performance evaluation of optimization algorithms over benchmark functions.
Fig. 2
figure 2

Convergence characteristics of selected test functions from CEC14 and CEC17.

Experimental evaluation and discussion

In order to compare the performance of several optimization techniques for multi-level image thresholding, a standard collection of breast thermogram images has been employed. The initial step in solving any optimization issue is to set the parameters of the method to be used. Prior to execution the SAO and ChBSAOACS algorithms, values are assigned based on a sensitivity study of the parameters (parameter tuning), as indicated in Table 2. Energy curve-based Otsu and Kapur techniques are used to evaluate the performance of ChBSAOACS using a collection of eight widely used breast thermogram images from literature. Figure 3 gives information about the test images under consideration and their corresponding histogram and energy curve. The outcomes are compared and analyzed with other methods based on Dragonfly Algorithm (DA), Kill Herd Algorithm (KHA), Genetic Algorithm (GA) and Particle Swarm Optimization (PSO) algorithms24,25,41 from the literature.

Table 2 Algorithm parameters and assigned values.
Fig. 3
figure 3

Test images and their corresponding histogram and energy curve.

Results of ChBSAOACS-based algorithms for multilevel image thresholding are evaluated using post-execution evaluations of picture quality such as Root Mean Square Error (RMSE), Peak Signal to Noise Ratio (PSNR), Structural Similarity Index (SSIM)42, Feature Similarity Index (FSIM)and43,44. Wilcoxon’s rank-sum test is a prominent non-parametric statistical analysis method for evaluating the performance of optimization algorithms. Self-written MATLAB algorithms running on an Intel Core i5-10310U2.21 GHz Dual Processor with 8 GB RAM are used in all simulations.

Fig. 4
figure 4

ChBSAOACS implementation results over test10 image using EC-Otsu and EC-Kapur methods.

Fig. 5
figure 5

ChBSAOACS implementation results over test11 image using EC-Otsu and EC-Kapur methods.

Fig. 6
figure 6

ChBSAOACS implementation results over test2 image using EC-Otsu and EC-Kapur methods.

Fig. 7
figure 7

ChBSAOACS implementation results over test3 image using EC-Otsu and EC-Kapur methods.

Fig. 8
figure 8

ChBSAOACS implementation results over test30 image using EC-Otsu and EC-Kapur methods.

Fig. 9
figure 9

ChBSAOACS implementation results over test31 image using EC-Otsu and EC-Kapur methods.

Fig. 10
figure 10

ChBSAOACS implementation results over test4 image using EC-Otsu and EC-Kapur methods.

Fig. 11
figure 11

ChBSAOACS implementation results over test5 image using EC-Otsu and EC -Kapur methods.

Test 10, test 11, test 2, test 3, test 30, test 31, test 4, and test 5 breast thermogram images45 from EC-Otsu-ChBSAOACS and EC-Kapur-ChBSAOACS are displayed in Figs. 4, 5, 6, 7, 8, 9, 10 and 11. Input image, segmented output images at ideal threshold levels, and corresponding energy curve information for Otsu-ChBSAOACS and Kapur-ChBSAOACS are all provided in each Figure. Images segmented using either method may be easily distinguished from one another. The reason for this is that Otsu’s between-class variance and Kapur’s entropy serve as the basis for two distinct methods. Therefore, it is necessary to rely on picture quality measures to evaluate the results. The optimal threshold values achieved using EC-Otsu-ChBSAOACS are compared in Table 3 with those obtained using other methods, including EC-Otsu-based SAO, DA, KHA, GA, and PSO algorithms41.

Table 3 Optimal threshold levels obtained by various optimization algorithms using EC based Otsu method.
Table 4 Optimal threshold levels obtained by various optimization algorithms using EC based Kapur method.

Table 4 presents the comparison of optimal threshold values obtained byEC-Kapur-ChBSAOACS with various existing approaches such as EC-Kapur-based DA, KHA, GA and PSO algorithms24,25,41. To our knowledge, no algorithms have produced the same set of solutions for the matching image as shown in Tables 3 and 4 for a given value of ’th’. Therefore, it is clear that different optimization algorithms use different evolutionary processes to arrive at the optimal solution. The PSNR values for all images at each threshold setting are compared across several Otsu and Kapur-based methods in Table 5. When compared to other current methods, the PSNR values produced using the EC-Otsu-ChBSAOACS and EC-Kapur-ChBSAOACS procedures are consistently superior (see Table 6). The quality of the segmented image increases as PSNR rises. The EC-Otsu-PSO method yields the lowest average PSNR value, at 15.91, while the EC-Kapur-ChBSAOACS method yields the highest, at 20.03.

Table 5 PSNR values obtained by various optimization algorithms using EC based Otsu and Kapur methods.

The suggested Otsu-ChBSAOACS and Kapur-ChBSAOACS methods, as well as other methods, are compared to FSIM values produced by other methods in Table 6. According to common knowledge, an FSIM value near to ‘1’ implies that the input images was segmented effectively with little loss of features in the output image. Average FSIM values range from 0.7188 when using EC-Otsu-PSO to 0.8135 when using EC-Otsu-ChBSAOACS, a significant increase. Table 7 compares the SSIM values found using several methods, including the proposed Otsu-ChBSAOACS and Kapur-ChBSAOACS methods. The SSIM result close to ’1’ implies that the input image was segmented efficiently and structurally comparable to the input image with little loss of information content. Average FSIM values range from 0.6908 for EC-Otsu-PSO to 0.7536 for EC-Otsu-ChBSAOACS, which indicates a significant increase. Based on the quantitative analysis of results for the investigated approaches it is clear that the overall performance of the EC-based thresholding approach using Kapur’s entropy is better than all other approaches for the tested images. However, it appears that image thresholding ability was better with the EC-Otsu approach than the EC-Kapur approach based on qualitative results.

Table 6 FSIM values obtained by various optimization algorithms using EC based Otsu and Kapur methods.
Table 7 SSIM values obtained by various optimization algorithms using EC based Otsu and Kapur methods.

Methods of statistical analysis are compared and contrasted using Wilcoxon’s rank-sum test, which takes into account both alternative and null hypotheses. The lack of a substantial difference is represented by the null hypothesis, whereas its presence is shown by the alternative. The alternative hypothesis is accepted if and only if the null hypothesis is rejected at a significant level (P>0.05). Each method has been executed for 30 times to perform the test. Wilcoxon’s rank-sum test results are shown in Table 8 for each technique. In addition, the computing complexity of each approach is evaluated to maintain a level playing field. ChBSAOACS and SAO has an O(t(N D)) computational complexity, where t is the number of iterations, D is the number of variables, and N is the population size of the solutions. There is no difference in computational complexity between GA, PSO, and DA. However, KHA has a worst-case computational complexity of O(t(Nlog(N))) and an average complexity of O(t(N2)). This means that the computational complexity of KHA is worse than those of GA, PSO, DA, and variants of SAO due to the need to sort the solutions in each iteration.

Table 8 Wilcoxin’s test.

With each repetition, the ChBSAOACS algorithm’s adaptive mechanism enhances the convergence speed. It’s also crucial for maintaining a healthy balance between exploration and exploitation so that selection issues don’t generate too many local solutions and so that an accurate estimate of the best solution can be found. Sniffing mode (Gases Brownian Motion), trailing mode and random mode is used to calculate the new position of a molecule in ChBSAOACS. The fitness values are taken into account when the ChBSAOACS algorithm is updated (the fitness value is related to the molecule mass and temperature). Because of the high velocities of molecules, the ChBSAOACS algorithm can scan the search space more quickly than other algorithms. The local search and global search capabilities of the ChBSAOACS algorithm may be better exploited and explored using this approach. As a result, ChBSAOACS outperformed the other algorithms in this research by a wide margin. The average execution time for investigated ChBSAOACS, SAO, DA, KH, GA and PSO optimization algorithms at 5 threshold levels is 31s, 27s, 28s, 26s, 47s and 29s respectively.

However, the ChBSAOACS algorithm presents a novel and potentially powerful approach to optimization, it has certain limitations related to complexity, computational cost, parameter sensitivity, scalability, and dependence on initial conditions. Addressing these limitations through future research directions such as optimizing computational efficiency, developing robust parameter tuning methods, exploring hybrid approaches, and enhancing scalability and robustness can further improve the algorithm’s performance and broaden its applicability.

Conclusions

Breast thermogram image segmentation based on multilevel thresholding was proposed in this article. More image information was retained while using EC-based multi-level thresholding with Otsu and Kapur objective functions. Experiments have shown that the suggested approach is simple and superior to existing segmentation methods in terms of segmenting breast thermogram images. There are a variety of quantitative indicators used to estimate the suggested method’s performance. The likelihood between the original image and the segmented image is reflected in these quality characteristics. Quantitative evidence demonstrates that EC-based thresholding using Kapur’s entropy yields accurate picture segments. Compared to other approaches, the proposed algorithm produces output images with well-delimited parts that may be evaluated qualitatively. The preliminary findings are encouraging and warrant further investigation on image classification, remote sensing, image denoising and enhancement, and other computational difficulties are only some of the many image processing problems that benefit from the EC-Kapur approach. Due to its speed and robustness, the suggested technique can also be used for other types of applications like control system design, wireless sensor networks, feature selection, data mining, clustering, etc. The hybridization of ChBSAOACS with other metaheuristic algorithms can also improve the algorithm’s performance. The method may be used with various local and global search strategies, such as Pareto distribution, Lévy flight random walk, mutation operators, and more.