Introduction

In recent decades, different modified and improved metaheuristic optimization algorithms have been developed to provide higher-quality solutions for optimization problems, including global optimization1,2. The main idea is to enhance the searchability of a metaheuristics algorithm by modifying its operators using an efficient technique that can boost solution quality and improve the acceleration of the convergence speed. There are various applications for the metaheuristics optimization algorithms, for example, enhancing time-series forecasting3, solving global optimization problems1, manufacturer machine scheduling4, feature selection5,6, data clustering7 and many other applications1,2.

For example, the differential evolution was applied in8 to solve unconstrained global optimization problems. The differential evolution was improved using four crossover operators; called binomial, simple arithmetic, uniform arithmetic, and single arithmetic crossover. The four operators were employed to improve its performance and to avoid the problem of the original crossover operators of the differential evolution. It was evaluated with four standard benchmarking functions and achieved significant results. Houssein et al.9 developed an enhanced version of the marine predators’ algorithm. The opposition-based learning technique was adopted to boost the traditional MPA’s searchability and accelerate convergence speed. It was compared to several old and new metaheuristics methods modified by OBL, and the MPA-OBL obtained the best results in different datasets.

The artificial bee colony algorithm was also used to address the global optimization issue by Chu et al.10. They used adaptive heterogeneous competition to augment the ABC and boost its solution quality. It was compared to several ABC variants and other metaheuristics optimization methods and performed better than them in different benchmark function datasets. Two variants of the grey wolf optimizer were developed in11. The first one is called expanded Ex-GWO. The Ex-GWO has the same main three wolves of the traditional GWO: delta, alpha, and beta. Depending on the first three wolves, the next wolves update their positions in each iteration. The second one, incremental I-GWO, depends on the incremental model. The main idea of the I-GWO is that each wolf updates its position according to the wolves selected before it. Ex-GWO and I-GWO were evaluated on 33 benchmark functions and showed good performance. Cuong-Le et al.12 developed a novel version of GWO, namely NB-GWO, based on a balance function. NB-GWO was utilized to optimize hyperparameters of deep neural networks.

Zhang et al.13 developed an enhanced salp swarm algorithm (SSA) for solving global optimization problems using different strategies, such as generalized oppositional learning, quadratic interpolation, and orthogonal learning. The developed ESSA was evaluated in well-known benchmark functions with extensive comparisons to the traditional SSA and other optimizers and achieved significant results. Additionally, there are various modified metaheuristics algorithms, such as an enhanced sine cosine algorithm14, the modified whale optimization algorithm15, Quantum-inspired differential evolution16, a memory-based optimization algorithm17, a hybrid of butterfly and flower pollination optimization algorithm18 and an enhanced manta ray foraging optimizer19.

Metaheuristic algorithms are widely used these days for solving intricate optimization problems for the reason that they can be derived from a variety of nature and artificial phenomena. Many researchers have tried to understand the base and efficacy of these algorithms and studied their structures, performance, and versatility with respect to different optimization problems. Important developments and understanding have also been achieved in methods known as Grey Wolf, Moth Flame, Whale, Firefly, Bat, and Antlion Optimizations. Moreover, research on algorithms such as Whale Optimization Algorithm (WOA), and Chimp Optimization Algorithm (ChOA) have looked into their operational working and benchmarks for performance metrics and open improvement areas20,21,22. These studies are aimed at revealing the foundations of metaheuristics further to sharpen a more vigorous approach towards the optimization problem.

Like global optimization, metaheuristics methods have also been adopted to solve data clustering problems depending on the hybridization concept. For example, Han et al.23 developed a new gravitational search algorithm variant to address the data clustering issue. The modified variant is called bird flock GSA, where a new mechanism inspired by birds’ collective response behavior was developed to add diversity to the traditional GSA. The developed BFGSA was compared to several basic optimization methods, including the conventional GSA, and showed superior performance. A hybrid Harris Hawks optimization algorithm with differential evolution was developed by24 for data clustering. The operators of the differential evolution were utilized to boost the exploitation process (local search) of the traditional HHO. This hybrid model showed significant performance compared to the conventional differential evolution and HHO. In25, a modified genetic algorithm, called a multi-objective GA, was proposed to be employed with fuzzy c-means for solving data clustering. Kaur et al.26 proposed an efficient clustering approach using a chaos and flower pollination algorithm hybrid over k-means. The chaotic FPA was compared to several well-known optimizers, including the original FPA, which performed better. There are also various boosted metaheuristics algorithms developed for data clustering, such as an improved Black-Hole algorithm27, moth-flame optimization algorithm28, an improved ABC using WOA29 and multi-verse optimizer30.

Metaheuristic algorithms are also referred to as optimization techniques as they are formed from inspiration drawn from natural processes, biological systems, or physical phenomena. Some prominent examples of metaheuristic algorithms include genetic algorithms, particle swarm optimization, ant colony optimization, and differential evolution. There are myriad applications of these algorithms in almost every domain, such as engineering design, machine learning, etc. This is due to their inherent capacity to do two things optimally: global search across the entire solution space and local search in scope in order to refine the solutions for optima. However, no algorithm is free of shortcomings; for instance, some studies suggest that GA suffers from slow convergence. PSO and ACO are prone to becoming stuck in local optima in high-dimensional spaces. These challenges provide the impetus for the development of new algorithms of metaheuristic properties that can target one or more optimization objectives. In The last few years, many cutting-edge algorithms have emerged specifically to tackle these shortcomings found in the traditional metaheuristic approaches. So far, metaheuristic algorithms like Grey wolf optimizer, whale optimization algorithm, and slime mold algorithm have proven to be effective in processing high dimensional and complex optimization problems. These algorithms also adapt new mechanisms like novel strategies for searching the solution space, better balancing between exploration and exploitation, and others, which result in better and faster algorithms. Nevertheless, there remain many challenges that are yet to be solved; issues like scalability, multiple other scenarios in a variety of problem landscapes, and efficiency emerge as ones requiring attention. In this scenario, Aquila Optimizer has been invented as an add-on, but this tool still has a performance gap, specifically in high dimensional optimization; thus, LOBLAO is proposed.

The clustering problem is a ubiquitous challenge across various domains, where the goal is to group similar data points while maintaining separation between different groups. This problem arises in machine learning, data analysis, and pattern recognition. However, due to real-world data’s complexity and feature spaces’ high dimensionality, traditional methods often struggle to find accurate and efficient solutions. Addressing the clustering problem demands sophisticated optimization techniques that can handle large datasets, high-dimensional spaces, and non-convex relationships31,32. These optimization methods play a pivotal role in unraveling underlying patterns within data and are crucial for achieving meaningful insights and decision-making in complex scenarios.

In this paper, we introduce a new clustering method based on metaheuristics, utilizing an enhanced version of the Aquila Optimizer (AO) algorithm, which we call the Locality Opposition-based Learning Aquila Optimizer (LOBLAO). The AO algorithm is a recently developed optimization technique that mimics the natural behaviors of Aquila33. Its original implementation demonstrated strong search capabilities, performing effectively across various optimization tasks and surpassing several earlier metaheuristic algorithms33. However, like many metaheuristic methods, AO has certain limitations in its search capabilities. In order to tackle these problems, we incorporate the Opposition-Based Learning (OBL) technique into traditional AO to extend the population diversity and collaboration of the search strategies. Furthermore, to deal with the problem of local searches focusing on the same areas, we strengthen the exploration of new search areas by using a Mutation Search Strategy (MSS). The LOBLAO method is verified in the context of global optimization and data clustering problems. The effectiveness of the LOBLAO algorithm was compared with both standard AO and various other metaheuristic optimization algorithms on a set of well-known benchmark functions.

In short, this study proposes the following contribution:

  • A new Aquila Optimizer algorithm variant, called LOBLAO, is proposed to solve global optimization and data clustering problems.

  • Enhancing solution quality of the traditional AO algorithm using the Opposition-based Learning (OBL) technique. The OBL can keep the diversity of the solutions and maintain the equilibrium between the search mechanisms.

  • Enhancing the search process of the traditional AO using the Mutation Search Strategy (MSS), which can be used to find new search regions.

  • Implement extensive comparisons to verify the performance of the developed LOBLAO in both global optimization and data clustering problems.

To sum up, the structure of the current study is as follows. Section “Background and algorithms” presents the basic methods used in the main steps of the proposed method. Section “The developed approach (LOBLAO)” highlights the design of the developed LOBLAO and shows the main operators used and their modifications. Section “Experiments and results” illustrates the global optimization and data clustering evaluation experiments and their settings. Section “Conclusion and future works” concludes the current study and proposes a set of future work directions.

Background and algorithms

Data clustering problem

Based on the common property of having a d-dimensional range, the data clustering problem involves dividing N data objects into multiple clusters, denoted as K (groups). A set of N data elements in the d-dimensional space can be represented as \(C = \left\{ c_1, c_2, c_3,\ldots , c_n \right\}\). Conversely, the k groups can be expressed as \(X = \left\{ x_1, x_2, x_3,\ldots , x_n \right\}\). Typically, each group should contain at least one member or data item.

$$\begin{aligned} X_{i} \ne \phi , \forall \in \left\{ 1,2,\ldots ,k \right\} \end{aligned}$$
(1)

At the same time, no more than two clusters should share a data object or a member.

$$\begin{aligned} X_i\cap Xj = \phi , \forall i \ne j \, and \, i,j \in {1,2,\ldots ,k} \end{aligned}$$
(2)

A group should be assigned to each data object.

$$\begin{aligned} C = \bigcup _{i=1} ^{k} X_i \end{aligned}$$
(3)

For information consolidation purposes, the distance test is important. Often, the shortest distance between two data sets is the distance between two points in that space defined by characteristic features of the two data sets. For this purpose, the most known measurement is the Euclidean distance. At the same time, the general Euclidean separation is used in the calculations of clustering to estimate the quality of the clusters24.

The Euclidean distance (E.d) between the data points p, and q can be calculated.

$$\begin{aligned} E.d (D_p, D_q) = \sqrt{\sum _{s=1}^{l} (D_{p,s} - D_{q,t})^2 } \end{aligned}$$
(4)

where q represents the \(\hbox {q}_{th}\) data point, and l indicates the length of the dimension. Additionally, the total intra-cluster distance is the most frequently utilized metric34. The whole intra-cluster remove degree determines the approximate separation between data points of any cluster and the barycenter of that cluster.

$$\begin{aligned} Intra _{sum}= \sum _{q=i}^k \left\| X_{cq} - C_{q} \right\| \end{aligned}$$
(5)

where q is the \(\hbox {q}_th\) data point, and \(\hbox {C}_q\) is given to the \(\hbox {q}_{th}\) cluster’s barycenter.

Aquila optimizer (AO)

This section provides a basic overview of the Aquila Optimizer (AO). The AO algorithm, introduced by33, mimics the social behavior of Aquila when hunting for prey in the wild. The AO is classified as a population-based optimization method, starting with the generation of an initial population X that has N dimensions. This process is described by Eq. (6).

$$\begin{aligned} X_{ij}=r_1 \times (UB_j-LB_j)+LB_j,\, \,\, j=1,2,3,\ldots ,Dim, \, \, i=1,2,3,\ldots ,N \end{aligned}$$
(6)

The search domain’s limitations are \(UB_j\) and \(LB_j\). Dim is the population’s dimension. \(r_1 \in [0,1]\) is the random value.

The next steps are to explore or exploit the search space, and they are repeated till founding the best solution. According to33, the AO uses two methods for applying the exploration and exploitation.

The agent’s best (\(X_b\)) and the average (\(X_M\)) are used in the first strategy to perform the exploration as in Eqs. (78).

$$\begin{aligned} & X_i(t+1)=X_b(t)\times \left( \frac{1-t}{T} \right) + ( X_{M}(t) -X_b(t) *rand), \end{aligned}$$
(7)
$$\begin{aligned} & X_{M}(t)=\frac{1}{N}\sum _{i=1}^{N}X(t) \end{aligned}$$
(8)

Here, \(\left( \frac{1-t}{T} \right)\) keeps the search active during the exploration phase. T represents the maximum number of iterations.

The second technique, formulated as follows, employs the Levy flight (Levy(D) distribution and \(X_b\) to update the solutions’ exploration capabilities.

$$\begin{aligned} & X_i(t+1)=X_b(t)\times Levy(D)+X_R(t)+(y-x)*rand, \end{aligned}$$
(9)
$$\begin{aligned} & Levy(D)=s \times \frac{u \times \sigma }{|\upsilon |^{\frac{1}{\beta }}}, \, \sigma =\left( \frac{\Gamma (1+\beta ) \times sine(\frac{\pi \beta }{2} )}{\Gamma (\frac{1+\beta }{2}) \times \beta \times 2^{(\frac{\beta -1}{2})}} \right) \end{aligned}$$
(10)

where \(\upsilon\) and u denote random numbers. \(beta=1.5\) and \(s=0.01\) are constants. \(X_R\) is an agent selected randomly in Eq. (9). In addition, x and y are utilized to simulate the spiral shape as in the following equations:

$$\begin{aligned} & x= sin(\theta ) \times r, \, \, \, \ y=cos(\theta ) \times r \end{aligned}$$
(11)
$$\begin{aligned} & r=r_1+U \times D_1, \, \theta = -\omega \times D_1 +\theta _1, \, \theta 1=\frac{\pi \times 3 }{2} \end{aligned}$$
(12)

here \(U=0.00565\) and \(\omega =0.005\). \(r_1\) is selected randomly \(\in [0,20]\).

The first way is used in33 to update agents in the exploitation stage using \(X_M\) and \(X_b\) as follows:

$$\begin{aligned} X_i(t+1)= ((UB-LB) \times rand+LB)\times \delta + ( X_b(t)-X_{M}(t) )\times \alpha - rand \end{aligned}$$
(13)

The parameters for exploitation adjustment are denoted by \(\delta\) and \(\alpha\), with a random value (rand) ranging from [0, 1].

The agent’s update is influenced by the quality of the function (QF), Levy, and \(X_b\) in the subsequent exploitation strategy. This process is defined as follows:

$$\begin{aligned} & X_i(t+1)= X_b(t) \times QF -(G_1\times X(t)\times rand)-G_2\times Levy(D)+ G_1 \times rand \end{aligned}$$
(14)
$$\begin{aligned} & QF(t)=t^{(\frac{rand \times 2 -1}{(1-T)^2})} \end{aligned}$$
(15)

Furthermore, \(G_1\) denotes several motions applied to find the optimal individual solution, as follows:

$$\begin{aligned} G_1=2 \times rand-1, \, \, \, G_2=(1-\frac{t}{T}) \times 2 \end{aligned}$$
(16)

where \(G_2\) is used to decrease the values from 2 to 0. The exact steps of the AO method are shown in Fig. 1.

Figure 1
figure 1

Aquila Optimizer (AO).

Opposition-based learning

Opposition-based learning (OBL) is a machine intelligence approach35 that has been utilized to enhance the performance of various optimization techniques36,37. The OBL strategy focuses on generating an opposition solution, which aims to identify a better candidate solution that yields a superior fitness value and moves closer to the optimal solution.

The opposite value of \({\overline{X}}\) for a given value \(X\in [UB,LB]\) is defined as:

$$\begin{aligned} {\overline{X}}=UB+LB-X \end{aligned}$$
(17)

Assuming \({\overline{X}}\) = (\(X_1\), \(X_2\), ..., \(X_n\)) represents a point in a multi-dimensional space, where \(X_1\), \(X_2\), ..., \(X_D\) \(\in\) R and \(X_j\) is within [\(UB_j\), \(LB_j\)], for j \(\in\) 1, 2, ..., D. The definition of \({\overline{X}}\) in n dimensions can be expressed as:

$$\begin{aligned} \overline{X_j}=UB_j+LB_j-X_j, \quad \quad {\text {where}} \quad j=1\ldots D. \end{aligned}$$
(18)

Furthermore, the supplied two solutions (\({\overline{X}}\)and X) are compared according to their fitness functions in the optimization phase, with the best solution being stored and the other being discarded. If f(X) leq f(\({\overline{X}}\)), X is saved in the minimization case; otherwise, \({\overline{X}}\)is saved.

Mutation search strategy (MSS)

In genetic algorithms, the mutation operator is very significant. By creating a uniformly distributed random value between [0, 1],38, the mutation probability Mu is used as a control parameter for tuning the mutation operator. The mutation operator is defined as follows:

$$\begin{aligned} X_{ij}= {\left\{ \begin{array}{ll} Xb_j+\mu (X_{pj}-X_{qj}) & if \, rand()>\mu _r\\ X_{ij}& Otherwise \end{array}\right. } \end{aligned}$$
(19)

where \(\mu _r = 0.5\), \(p, q \in \{1, 2,\ldots ,i - 1, i + 1,\ldots , N\}\) and \(\mu \in [0,1]\).

The developed approach (LOBLAO)

This section describes the so-called Locality Opposition-based Learning Aquila Optimizer (LOBLAO), which is a new approach developed to increase the robustness of the basic Aquila Optimizer (AO) in relation to high dimensional and complex optimization problems. The technique contains three principal strategies-Aquila Optimizer Operators, Opposition Based Learning (OBL), and Mutation Search Strategy (MSS)-that interact with each other such that the identified limitations of the original AO are obviated. The incorporation of these strategies leads to LOBLAO achieving an improved balance between exploration and exploitation, enhancement of solution diversity as well and alleviation of early convergence to a solution.

Strategies for improvement in LOBLAO

The proposed LOBLAO in Fig. 2 uses three additional methods to mitigate the weaknesses of the original AO:

Figure 2
figure 2

The proposed LOBLAO.

Aquila optimizer operator

The core components of Aquila Optimiser are still retained, which are the building blocks of LOBLAO, and this is necessary since these components form a suitable and effective framework of modeling, exploration, and exploitation based firmly on the hunting and flight movement of the Aquila bird. These operators also assist in scaling the searching of strategies and in maintaining that LOBLAO retains notable features as AO, which includes being adaptable and simple.

Opposition-based learning (OBL) strategy

The strategy helps to increase the diversity of the solutions by using both the current and the current opposite solution. Instead of following the conventional OBL methods, the OBLAO model uses a *least OBL* approach, which creates oppositional solutions within a limited boundary of the search space. This centripetal technique aids in eliminating the unfocused and unwarranted movement in other areas while still retaining diversity. With the addition of localized OBLs, LOBLAO is able to effectively advance and make sure that the algorithm is always active and self-adjusting to a myriad of problem topologies.

Mutation search strategy (MSS)

In the way that LOBLAO is able to locate high dimensional local maxima, MSS has been able to due to it being an inclusion strategy. The Modification Search Strategy takes controlled variations of possible solutions, leading to the exploration of new regions that have not been explored before around the solution space. Combined with robust search patterns and other measures like MSS, the modification search strategy achieves tremendous effectiveness in overcoming the challenge of early convergence. MSS’s inclusion, along with robust search patterns, ensures that LOBLAO will not be trapped in local optimal solutions and will constantly progress its searching direction over complex multi-dimensional spaces.

Workflow of LOBLAO

Figure 2 outlines the specifications of the newly LOBLAO procedure. In every iteration of the model, three strategies are applied: Aquila Optimizer operators, OBL, or MSS. In every iteration, given probabilities are randomly assigned to these three strategies, which are biased towards one of them, leading to a certain strategy being selected and applied. In this manner, the algorithms ensure that the contribution of all the strategies in a team-based optimization task is equal.

1. Initialization: The algorithm generates a random initial candidate population throughout the search space.

2. Fitness Evaluation: The second step involves measuring the fitness value for every candidate solution through the objective function.

3. Search Strategy Selection: One of the three strategies (AO, OBL, or MSS) is implemented by the algorithm in every iteration to perform updates to the candidate solutions

4. Updating Population: The population is updated according to the selected strategy, and the best solutions are kept for the next iteration.

5. Termination: This final stage starts when the maximum number of evaluations of the fitness or iterations has been reached.

Advantages of LOBLAO

New enhancements are done on LOBLAO so that it can work on the concurrent limitations of the original AO as well as other modern algorithms:

  • Enhanced Exploration and Exploitation: The combination of the AO operators and OBL and MSS is good for LOBLAO as it helps achieve a good balance between global and local search ranges.

  • Improved Solution Diversity: Applying a localized OBL allows the search process to target a wide range of potential solutions and, therefore, improves the diversity of the solution set.

  • Resilience to Local Optima: With the help of MSS, which involves randomness and variability, the algorithm is able to break away from local optima and look for other areas.

  • Efficiency in High-Dimensional Problems: The specific and directed approach of LOBLAO’s strategies ensures that it can be employed efficiently in optimization tasks that are complex and high-dimensional, and the scaling is also reasonable.

The modifications incorporated in LOBLAO make it an efficient optimization tool that can address a wide range of complex problems. This section illustrates the workings and efficiency of the proposed method using detailed experiments with benchmark test functions and data clustering problems.

Computational complexity of the proposed LOBLAO

The recommended computational complexity of LOBLAO is influenced by how candidate solutions are initialized, how the objective function of the current solutions is evaluated, and how candidate solutions are updated iteratively.

Let’s assume that N represents the total number of employed solutions, and O(N) indicates the time complexity for initializing these solutions. The time complexity for updating the solutions can be expressed as O(T \(\times\) N) + O(T \(\times\) N \(\times\) Dim), where T is the total number of iterations and Dim refers to the spatial dimension of the problem. Therefore, the time complexity for the LOBLAO can be described as follows.

$$\begin{aligned} O(LOBLAO)=(N)\times O(AO)+ O(OBL) + O(MSS) \end{aligned}$$
(20)

The time complexity of the method under consideration is contingent on the interplay of three primary search operators: AO, OBL, and MSS. The computation of the time complexity for these methods is delineated below.

$$\begin{aligned} & O(OBL)=O(N\times (T\times Dim+1)) \end{aligned}$$
(21)
$$\begin{aligned} & O(AO)=O(N\times (T\times Dim+1)) \end{aligned}$$
(22)
$$\begin{aligned} & O(MSS)=O(N\times Dim) \end{aligned}$$
(23)

Henceforth, the comprehensive time complexity of LOBLAO can be expressed as follows.

$$\begin{aligned} & O(LOBLAO)=O(T \times N\times (Dim+1) +( N\times Dim) +( N\times Dim)) \end{aligned}$$
(24)
$$\begin{aligned} & O(LOBLAO)=O\left( T \times N \times \left( Dim+N\right) \right) \end{aligned}$$
(25)

Experiments and results

This section outlines the experiments conducted and the results obtained from both the proposed methods and other comparative approaches. Two experiments were carried out: the first focused on global optimization, while the second addressed clustering problems. All experiments utilized MATLAB R2015a, running on an Intel(R) Core(TM) i7 processor with 16GB of RAM. The global parameters were configured with a population size of 30 and a total of 500 iterations. To ensure statistical validity, each experiment was executed with 30 independent runs.

Experiments 1: Benchmark functions problems

This experiment assesses the performance of the LOBLAO by utilizing 23 established benchmark functions39 and 30 CEC 2017 benchmark functions40. The LOBLAO was executed for 500 iterations with 30 candidate solutions to tackle these test functions. To evaluate the reliability of the LOBLAO, the algorithm was run independently 30 times; we documented the best function value, average outcomes, worst function value, standard deviation (STD), p-value, and rank. The subsequent sections will compare the proposed algorithm with 11 prominent state-of-the-art algorithms. To ensure a fair comparison, all algorithms were set to the same population size and iteration counts of 30 and 500, respectively.

Description of benchmark functions

To evaluate the exploratory and exploitative behaviors of the LOBLAO, we utilized 23 benchmark functions that encompass various problem types: multimodal, fixed-dimension multimodal, and unimodal functions39. The LOBLAO will be tested on unimodal functions (F1–F7) (refer to Fig. 3) to assess its exploitation tendencies. Additionally, the multimodal benchmark functions (F8–F13) will be employed to evaluate the exploration capabilities of the LOBLAO. Both 10 and 100 dimensions are used in these two groups of functions. To further investigate the exploration ability of the LOBLAO in lower dimensions, we will use the fixed-dimension multimodal functions (F14–F23) (see Table 3). A variety of well-known algorithms are compared with the proposed algorithm to demonstrate the superior performance of the LOBLAO. The parameter values for the comparative algorithms are detailed in Table 1. The code was developed using the MATLAB R2015a platform and executed on a PC equipped with 16 GB RAM and an Intel(R) Core(TM) i7 CPU.

Figure 3
figure 3

Details of the tested benchmark functions.

Table 1 Parameters values of the tested algorithms.

Analysis of LOBLAO convergence

To illustrate the behavior of the LOBLAO algorithm, we plot the trajectory and convergence curves in Fig. 4. This figure presents qualitative measures, including the 2D function topology (first column), the best values of the first dimension (second column), the average fitness value of the LOBLAO algorithm (third column), and the convergence curves for both the AO and LOBLAO algorithms (fourth column). In the second column, the curve for the best values of the first dimension indicates that the solution begins with a high frequency and magnitude. However, in the last iterations, these values become obscured in functions F4, F9, and F10. This suggests that LOBLAO demonstrates strong exploration capabilities initially, followed by effective exploitation towards the end. This pattern indicates that the proposed algorithm has a significant chance of reaching the optimum. The third column shows the average fitness value of all candidate solutions over the iterations. It is observed that the average fitness value is initially high. Still, it decreases before the 40th iteration, suggesting that the LOBLAO algorithm requires only a few iterations to find the optimum. In the fourth column, the convergence curves for the AO and LOBLAO algorithms indicate that LOBLAO converges more quickly than AO. In some cases, such as with function F8, LOBLAO achieves a significantly better solution than AO. Additionally, the convergence curves are smooth, and LOBLAO improves with just a few iterations.

Figure 4
figure 4

Qualitative results for the studied problems.

Parameter analysis of the LOBLAO algorithm

In this part, Table 2 shows the effect of changing the solution number (N) of the LOBLAO on its behavior. To make a comprehensive analysis, several solution numbers are used (i.e., 5, 10, 15, 20, 25, 30, 35, 40, 45, and 50) to see the impact of this parameter throughout 500 iterations. Table 2 shows that the proposed algorithm keeps its power and robustness. From the rank in Table 2, one can see that the higher the solutions number (N) is, the higher the performance is. The LOBLAO with (\(N=50\)) has the first or second ranks in 11 functions, except for F7 and F13. Table 2 shows that the change in the solutions number in the functions F9, F10, and F11 does not affect the performance of LOBLAO. One can conclude that the best solution number is 45 because the LOBLAO with (\(N=45\)) has the first final rank. The Wilcoxon signed-rank test (\(\alpha = 0.05\)) is performed between the LOBLAO and the other cases (\(N=50\)). The p-value and h are provided. For a certain N, its h equals zero means that the LOBLAO with (\(N=50\)) is statistically different from this N value. Otherwise, the LOBLAO with (\(N=50\)) is statistically different from this N value.

Table 2 The effect of the number of solutions (N) on the performance of the proposed method.

Exploitative ability of the LOBLAO algorithm

To test the exploitation ability of the proposed algorithm, one can use unimodal functions (F1–F7) for this purpose. The results include the worst function value, the best function value, the average, and the standard deviation over the independent runs. The pairwise Wilcoxon signed-rank test is performed between the proposed algorithm and the other counterparts (see p-value and h in the last three rows of each function). The rank provided in Table 3, where the dimension is 10, shows that the proposed algorithm secures the first or second ranks in all unimodal functions except F6. The PSO got the first rank for F6. Generally, the LOBLAO outperforms the other competitors for the six unimodal functions and achieves high consistency (minimum standard deviation) with excellent exploitive behavior. This is because the proposed algorithm inherited from its parent, the AO algorithm33, has two exploitation strategies (i.e., narrowed and expanded exploitation), which encourages the LOBLAO algorithm to narrowly and widely intensive local search.

Explorative ability of the LOBLAO algorithm

To examine the exploration capability of the LOBLAO algorithm, two groups of multimodal functions are utilized for this purpose: multidimensional functions (F8–F13) and fixed-dimensional functions (F14–F23). Because these functions have many local optima, the results include the worst function value, the best function value, the average, and the standard deviation over the independent runs; Table 3 contains the results of the multidimensional functions (F8–F13) and Table 5 contains the results of the fixed-dimensional functions (F14–F23). Moreover, the h and p-value of the Wilcoxon signed-rank test are reported in both tables. From the rank values of Table 3, where the dimensionality equals 10, one can find that the LOBLAO algorithm secures the first rank in all multimodal, multidimensional functions except F12 and F13. The PSO achieves the first rank for F12, and the AO algorithm is ranked first for F13. Accordingly, the proposed algorithm is superior for four multimodal, multidimensional functions and has good exploration capability. This is because the proposed algorithm inherited from its parent, the AO algorithm33, two exploration strategies (i.e., narrowed and expanded exploration). Consequently, the proposed algorithm can discover the search space effectively compared to the other algorithms. The Penultimate row of Table 3 reports the Friedman mean rank test results. In conclusion, for all unimodal and multimodal multidimensional functions (F1–F13), the LOBLAO is the best performer, where its final rank occupies the first one (see the last row of Table 3). The rank value for the fixed-dimensional functions (F14–F23) shows that the LOBLAO is superior for half functions (F14, F17, F20, F22, and F23). For functions F15 and F18, the MPA algorithm has the best rank. For function F21, the AO algorithm is the best one. The Penultimate row of Table 5 reports the Friedman mean rank test results. In the last row of Table 5, the LOBLAO is the best performer for all fixed-dimensional functions (F14–F23), where its final rank came first. This finding confirms that the LOBLAO algorithm can deal with low-dimensional, fixed-dimensional, and multimodal functions. It achieves the highest position because it can explore the problem’s landscape and then exploit the best solution until it reaches the global optimum.

Table 3 Results of the comparative algorithms using 13 problems, where the dimension is 10.

Stability analysis of the LOBLAO

To assess the performance stability of the LOBLAO in tackling high-dimensional problems, thirteen benchmark functions are utilized (Table 4). These functions operate in a 100-dimensional space. Each algorithm is run independently 20 times for 500 iterations, using a population size of 30. Table 4 presents the worst, best, average function values and the standard deviation from these independent runs. Additionally, the h and p-values from the Wilcoxon signed-rank test are calculated. The ranking results demonstrate that the LOBLAO algorithm is dependable, robust, and stable in higher dimensions, achieving the top rank for nine functions and the second rank for four functions (F5, F7, F12, and F13) (refer to Table 4). These results indicate that the LOBLAO algorithm effectively balances exploitation and exploration when addressing both unimodal and multimodal functions in high dimensions. Among the other algorithms, AO secured the second overall rank following LOBLAO. Conversely, the SCA, SSA, DA, and ALO struggled with high-dimensional tasks, landing in the last four ranks. The penultimate row of Table 4 displays the results from the Friedman mean rank test. Overall, for all unimodal and multimodal functions (F1–F13), LOBLAO consistently outperforms the other algorithms, achieving the highest overall rank (see the last row of Table 4).

Table 4 The results of the comparative algorithms using 13 problems, where the dimension is 100.
Table 5 The results of the comparative algorithms using 10 problems.

Analysis of convergence behavior of the LOBLAO

The convergence of the LOBLAO towards the global optimum across iterations would be a good idea. In Fig. 5, we can see the best solutions achieved so far plotted against the number of iterations. The convergence behavior curves of the proposed LOBLAO demonstrate the quickest convergence for most unimodal functions, such as F1–F4 and F7, as well as for many multimodal functions, including F8, F9, F10, F11, and F13. Meanwhile, other algorithms tend to get stuck in local optima. This suggests that the LOBLAO effectively balances exploration and exploitation, allowing it to approach the near-optimum quickly. Subsequently, the LOBLAO performs an efficient local search around the global optimum, avoiding stagnation in any local optima. The convergence speed of the LOBLAO is comparable to that of other algorithms in cases like F14, F16, F17, F19, F21, and F22. In function F23, the proposed algorithm exhibits the fastest convergence. However, for function, F15, EO, WOA, and MPA demonstrate superior convergence. The proposed algorithm also shows a smooth transition from exploration to exploitation stages for the multimodal functions.

Figure 5
figure 5

Convergence behavior of the comparative algorithms on the test functions (F1–F23).

The results of the comparative algorithms using 10 problems

Table 6 appears to display a comparison of the basic Aquila Optimiser with the LOBLAO while using the CEC 2017 benchmark functions on a 30-D sphere. Furthermore, it details the performance of both algorithms by conducting a set of analyses which include computation of minimums, medians, means, maximums, and standard deviation.

Through the data collected, it appears LOBLAO outperforms AO by a significant amount in nearly every displayed parameter. The use of metrics such as minimum, median, mean, and maximum will help support the statement, as LOBLAO is expected to have lower numbers in every category while also having standard deviations smaller than AO. In other words, it highlights how robust the combination of OBL and MSS in LOBLAO has been. An example of this can be seen in the F1 score, where LOBLAO scores an impressive 1.00000e+03 while AO manages to get only an improvement of 1.23000e+05 with LOBLAO.

It is apparent from the reduced standard deviations that LOBLAO’s performance is consistent across the varying test issues. These more stable and reliable convergence tendencies underscore LOBLAO’s ability to reduce variability and sustain trustworthiness which for enhanced multi-dimensional optimization tasks is critical. Furthermore, the algorithm’s greatly improved management of complex benchmark functions complements the already mentioned validation of OBL and MSS mechanisms in addressing AO’s limitations, in particular, avoiding local optima and maintaining the exploration-exploitation balance.

Friedman’s ranking incorporated in the evaluation has been contrary in favor of LOBLAO. From all the problems tackled, LOBLAO has been first all the time, with AO remaining second, which emphasizes LOBLAO’s supremacy over others. This ranking substantiates LOBALAO’s strength and versatility when dealing with difficult optimization problems. Test problem F2 is an exception where results are absent (hence called “NAN”) for both tests. Even though this does not affect the general observations, it has been pointed out as a limitation in the presented data.

In conclusion, the results confirm that LOBLAO is a very useful and dependable optimization algorithm. Its performance in relation to AO in some difficult test problems commends it as an indispensable asset in complex optimization activities in a variety of fields. The combination of other components of LOBLAO, such as OBL and MSS, not only facilitates effective coverage of the search space but also improves the performance of simple AO.

Table 6 Comparison of results between basic AO and proposed LOBLAO on 30-dimensional CEC 2017 benchmark problems.

Analysis of the developed behaviors

Figure 6 depicts the search paths of the Aquila Optimizer (AO) and the enhanced Locality Opposition-Based Learning Aquila Optimizer (LOBLAO) on the Rastrigin function, a well-known problematic 2D target for optimization algorithms. This comparison sheds light on how both approaches are fundamentally different in their exploration and exploitation strategies and, in turn, how the improvements made in LOBLAO solve the issue of AO.

Figure 6
figure 6

Search Trajectories in a 2D Rastrigin Function.

The blue path traced by AO resides within the high area of the trajectory plane, indicating the performance of AO, which performs a cluster search within specific boundaries when searching for targets that probability displays a relatively easy search originating around the two-dimensional plane. In other words, LOs are less than L or equal to 1start para I when considering single-point parameters. L is a narrow scope, which makes the ability of AO more effective. Such a pattern of behavior portrays AO as prone to early convergence to locally optimal solutions, which is a severe weakness given that we are working with high dimensional tasks, as finding global optimal ones usually requires extensive exploration of the entire search space.

On the other hand, the red path indicates the benefits of the OBL-based OBL strategy incorporated into SMA LOBLAO. The OBL mechanism creates opposing solutions to replace the normal evaluation of the solution, so its diversity is increased. This allows for more relative diversity in the process and explains in part the greater extent of the AO and SMA coverage than previous iterations. Notably, the use of OBL in LOBLAO is spatially targeted, ensuring that any exploration that does not have expected results is sidestepped without any prejudice, which promotes greater efficiency where appropriate without loss of diversity caused by stopping in one place.

The green trajectory depicts LOBLAO’s MSS strategy. A mutation candidate solution perturbs that was used by LOBLAO. This strategy allowed novel regions to be accessed but closed solutions to be escaped from. With this blend of flexibility and versatility, LOBLAO can handle things highly efficiently while keeping the need to exploit and seek reasonably laissez-faire. The net is an increase of coverage as solutions reach closer to the universal maximum.

The disparities in search spaces shown in the visualizations give enough room for the persuasive claims that support the idea that LOBLAO performs better than AO. The use of OBL and MSS in LOBLAO also works hand in hand with other optimizers, as these can complement Other optimizers by alleviating convergence issues associated with the rest dimensional structure convergence headaches, stiffness, and flow, to mention a few. In managing the exploring and exploiting parts, LOBLAO has shown itself to be much more adjustable and also reliable.

In conclusion, this figure clearly illustrates how far metrics related to the LOBLAO search have come from the previous AO performance metrics. With MSS and OBL, complex solving heuristics are coupled with resilient exploitation behind LOBLAO, allowing the system to handle deep, highly multi-dimensional matrices effortlessly. Details are very indicative of comparative analysis and graphically better performance than traditional methods.

Experiments 2: Data clustering problems

In this section, we evaluate the proposed method for addressing data clustering challenges. The experiment utilizes eight well-known datasets: Glass, Cancer, Iris, CMC, Seeds, Vowels, Heart, and Water41. Table 7 presents the characteristics of these datasets.

Table 7 UCI benchmark datasets.

Results and discussion

The results of the LOBLAO were compared with AOA, PSO, GWO, the African vultures optimization algorithm (AVOA)42, WOA, and the artificial gorilla troops optimizer (AGTO)43, as detailed in Table 8. Additionally, the Wilcoxon rank-sum test was conducted for each dataset to determine if there were significant differences between the LOBLAO and the other algorithms.

According to Table 8, the results for the Cancer dataset indicate that the proposed LOBLAO method achieved the best average measure, securing the top rank. The AVOA algorithm followed in second place, with AOA, AO, AGTO, WOA, PSO, and GWO ranking subsequently.

In terms of the Best measure, LOBLAO demonstrated performance comparable to both AGTO and AO, as they all yielded the same result (i.e., 3025.9) and were ranked second after AVOA. Conversely, GWO and PSO recorded the lowest results.

For the worst measure, LOBLAO, AOA, and WOA exhibited similar performance to some degree, while both the original AO and AGTO were ranked fifth with an identical result (i.e., 3511). Furthermore, the LOBLAO, AOA, and WOA algorithms displayed similar stability and were ranked third, following WOA and AOA. In contrast, PSO exhibited unstable behavior relative to the other algorithms.

Moreover, from the CMC dataset results, the proposed LOBLAO and AOA algorithms obtained nearly similar results in the average measure, followed by WOA. The AVOA, AGTO, and AO performed similarly and were ranked fourth. The PSO algorithm obtained the last rank. Similar algorithm performance was also shown in the Best measure; the AVOA, AGTO, AO, and LOBLAO showed very close results, followed by AOA and WOA, respectively. The AOA was the most stable algorithm with (0.51), followed by WOA and LOBLAO with (0.93) and (1.13), respectively. In this dataset, the PSO and WOA algorithms were ranked last.

The proposed LOBLAO performed superior in the Average and worst measures regarding the glass dataset and was ranked first. In the average measure, the PSO came in the second rank with (30.34), followed by WOA and AOA with (33.96) and (34.12), respectively, whereas the GWO, AGTO, AO, and AVOA algorithms obtained the same results (34.5). The PSO was also ranked second in the worst measures, while the rest of the algorithms obtained close results (34). Regarding the Best measure, the LOBLAO and PSO showed close results to (25) and (24), respectively, followed by WOA and AOA. Although LOBLAO showed the Best performance, it showed the worst stability in the glass dataset, followed by PSO. This can be due to the characteristics of the datasets and their different results during the independent runs.

The proposed LOBLAO’s good results were also shown in the Iris dataset; it was ranked first in the average measure, followed by AOA, GWO, and WOA, whereas the PSO recorded the worst results compared to the other algorithms. In the Best measure, both GWO and LOBLAO showed close results and ranked first and second, respectively, while the rest algorithms showed the same performances to some extent. In addition, all algorithms showed the same performances in the worst measure except for PSO and GWO; they were ranked last. Regarding the Std measure, the LOBLAO showed acceptable stability behavior equaled (0.59), whereas the PSO and GWO obtained (1.68) and (1.53), respectively.

In the Seeds dataset, the PSO and GWO were ranked first and second, respectively, whereas, in all measures except the Std, they were ranked last. The proposed LOBLAO, AGTO, and AO showed the same performance and ranked third in the average, Best, and Std measures, followed by WOA, AOA, and AVAO. The most stable algorithm in this dataset was the AVOA.

Furthermore, the proposed LOBLAO showed good performance in the Heart dataset and was ranked first in both Average and Best measures with (925.16) and (775.58), respectively. The PSO came in the second rank, followed by AOA, WOA, and AVAO; the other algorithms showed the same results in all measures. The most stable algorithms were AOA and AVAO.

In both Vowels and Wine datasets, the LOBLAO was ranked second, whereas the PSO was ranked first; the stability of the LOBLAO was better than that of the PSO. The AOA came in third rank, followed by WOA and AVOA. The GWO, AGTO, and AO performed similarly in both datasets.

The Wilcoxon rank-sum test was also considered. From this test, we can conclude that there are significant differences among the LOBLAO, PSO, and GWO in all datasets and among AOA, AVOA, and WOA in Iris, Seeds, and Vowels, respectively.

Table 8 The results of the comparative algorithms using eight data clustering problems.

Figure 7 illustrates the convergence curves of all algorithms in all datasets. From this figure, we can see that the LOBLAO effectively reaches the minimum fitness value over iterations.

Figure 7
figure 7

Convergence behavior of the comparative algorithms using the tested data clustering problems.

Sub-figures in Fig. 7 depict how different optimization algorithms converge across various datasets (Cancer, Cmc, Glass, Iris, Seed, Statlog, Vowel, and Wine), focusing on the best fitness value relative to the number of iterations. Each plot contrasts the performance of the proposed Locality Opposition-based Learning Aquila Optimizer (LOBLAO) with leading optimization algorithms, including AOA, PSO, GWO, WOA, AVA, and AGTO. The following discussion delves into the results across these datasets.

In the Cancer dataset, the convergence curve indicates that LOBLAO outperforms all the other methods by being able to reach a lower fitness value. LOBLAO is able to achieve quick convergence during the first few iterations and is able to continue improving. On the other hand, AOA, PSO, and other algorithms were reported to converge at higher final fitness values, indicating that the dataset was not optimally solved.

When using the Cmc dataset, LOBLAO outperforms every other method, as it has the best fitness values compared to them all. Its convergence is steady throughout, which shows its ability to explore and later effectively exploit the search space fully. The grey wolf optimization and the whale optimization algorithms have moderate performance but do not come close to the levels that LOBLAO reaches. What is more noticeable is the steepness of LOBLAO’s curve, which shows how efficient this method is in solving optimal problems.

The glass dataset has LOBLAO outperforming all other algorithms once again since it does better at achieving optimization results as compared to the other algorithms. Some algorithms, like the AGTO and PSO, do a little bit better but not enough to make a significant impact since they have already plateaued, which means there was not enough exploration done in the search space. On the other hand, LOBLAO keeps improving, as one would expect when working with more difficult datasets.

LOBLAO comes in first again in the Iris dataset, and the algorithm works by obtaining lower fitness values than others and by showing faster convergence. One thing to note is that there is a reduced difference between LOBLAO, AOA, and GWO, indicating that this dataset is less complex for most methods. Still, LOBLAO does beat all of the other algorithms in the relevant experiments, showing that it works well in relatively simple datasets.

The convergence curve in the Seed dataset also confirms the efficiency of LOBLAO since it takes the least number of iterations to converge to the best fitness values. The other typified algorithms, such as WOA and PSO, end up taking quite some time to converge and achieve higher final fitness values. This goes on to demonstrate the effectiveness of LOBLAO in solving datasets that are of a moderate level of complexity.

In the Statlog dataset, LOBLAO differentiates as well performing outstandingly since it posted results that are considerably higher in fitness values than all other methods. As evidenced in the plots, LOBLAO achieves fast convergence during the early iterations and never stops improving, achieving the goal of optimally solving the problem. First, it is noticeable that all other algorithms are rather slow, and progress is very erratic. This further emphasizes the strength of LOBLAO in terms of stability.

The Vowel dataset appears to be a more complex optimization problem as we can see that there is a wider range of fitness values from different algorithms. This is AOA and GWO that are many times unfit in this area but LOBLAO performs excellently as well by getting the best fitness value while massively decreasing. However, many algorithms, such as AOA and GWO, are not very effective here as they do not show a continuous decrease, which is LOBLAO’s strength for high dimensional or complex data sets.

In the context of the competition on the Wine dataset, LOBLAO outperforms every other method and has the lowest fitness value. Its convergence has a steady and rapid increase, enabling better searching and rewarding ensemble methods. Although AGTO and PSO show signs of improvement in the initial stages, they reach a plateau in the later stages, where they are unable to keep up with the optimization techniques employed by LOBLAO.

These results are self-explanatory and appreciation should be accorded to the newcomers who introduced LOBLAO as it beats others on most datasets. The following considerations seem to be most essential for the aim pursued by the authors:

  • Fast Convergence: LOBLAO demonstrates quick convergence in the early iterations, highlighting its efficiency in exploring the search space.

  • Better Final Fitness Values: In all datasets, LOBLAO achieves the lowest fitness values, showcasing its strong optimization abilities.

  • Adaptability: LOBLAO is effective in managing a wide range of datasets, from simpler ones like Iris to more complex ones such as Vowel and Statlog.

  • Consistent Performance: Unlike other algorithms that vary in success across datasets, LOBLAO consistently ranks at the top, underscoring its robustness and dependability.

These findings confirm LOBLAO’s effectiveness in tackling complex optimization problems, emphasizing its potential as a versatile and powerful optimization tool.

Figure 8
figure 8

Example of the clustering plot for five datasets.

In addition, Fig. 8 shows the clustering plot images conducted by the proposed LOBLAO, where each dataset is tested using a different number of clusters (i.e., K 2, 4, and 8). In this figure, the algorithm detected the groups of each dataset correctly.

The figures presented illustrate the clustering outcomes for various datasets (Cancer, Cmc, Glass, Iris, and Seed) using different numbers of clusters (K = 2, K = 4, and K = 8). Each dataset reveals its own unique patterns and challenges for clustering, offering valuable insights into how the clustering algorithm performs across a range of data distributions.

Using the K = 2 setting with the Cancer dataset, the algorithm splits the data into two clusters. It is able to produce the major arrangement of the data. A significant degree of overlap is, however, observed, implying that there are aspects of the model that could perhaps use some enhancement in the centers of the clusters.

In comparison to K = 2, K = 4 produced four more numerous groups, and the division into groups was more advanced; the groups produced in K = 2 were more broad and grouped the data into larger quadrants than K = 4. This shows the algorithmic potential to capture the smaller details in the dataset.

Finally, when level K was set to 8, the data produced clearer images of clumping, but some clumps were blurred, meaning they should not have been grouped, given they were distinctly different. This could represent cluttering or mere sensitivity to interference of the data used.

With K = 2, the first division leads to two primary clusters, illustrating the overall distribution of the data set. Yet the clear interference abounds, indicating that the data set is more complex. When K increases to K = 4, the separation improves since the method identifies more subgroups within the data. The clusters seem to be well-formed, suggesting that the revision to the algorithm is indeed working in dealing with more sub-groups in the data. But with K = 8 nearer to these clusters, it becomes more distinct with smaller scattered picked clustering. Although beneficial for detailed information, this may lead to local overfitting.

For the Glass Dataset, the K = 2 level indicates that points clusters contain some of the members, and ambiguity of some points settles on the borders of the clusters. But when K is made 4 the boundaries mark disappears and subsumes the overlying structure depicting those datasets. This further demonstrates the algorithm’s ability to strategically manage clustering by being general when necessary but more focused at other times. However, K = 8 shows that most of the clustering objectives have been met with the sub-clusters, unlike that of K = 4, which makes sampling difficult. This emphasizes that larger values of K may not be ideal for this data set.

In the Iris Dataset, at K = 2, the clusters correspond with the well-known separability of the Iris dataset, effectively splitting the data into two primary groups. For K = 4, the algorithm captures more subtle details within the dataset, forming distinct sub-clusters. This aligns with the known structure of the Iris dataset, where some species exhibit overlapping characteristics. With K = 8, the clustering becomes more intricate but may add unnecessary complexity, as the dataset inherently contains fewer distinct groups.

With K = 2, the algorithm effectively captures the overall distribution of the Seed dataset, resulting in two large, distinct clusters. When K is increased to 4, the granularity improves, and the clusters begin to reflect more specific patterns within the data. The outcomes appear well-balanced, showcasing the algorithm’s adaptability. At K = 8, however, the clusters become more fragmented, indicating that such a high level of granularity may not be suitable for this dataset, as it could lead to overfitting to minor variations.

Throughout all datasets, the clustering algorithm shows its ability to adapt to different data distributions and varying K-values. The findings emphasize the balance between granularity and simplicity:

  • Lower K-Values: Yield broad, general clusters that are appropriate for datasets with fewer inherent groups.

  • Higher K-Values: Provide more detailed segmentation but may lead to over-segmentation or increased sensitivity to noise.

This analysis highlights the significance of choosing an appropriate K-value based on the dataset’s characteristics and the specific application. Adjusting this parameter can help strike a balance between trade-offs and enhance clustering performance.

Limitations

Despite the impressive improvement that the proposed Locality-based Opposition Learning Aquila optimizer system has on the benchmark problems and the data clustering, some points limit the efficiency of the proposed system. One parameter that negatively affects the performance of the system is the charge associated with the use of opposition-based learning in combination with the mutation search strategy. These revisions help enhance the solution diversity and improve the search efficiency. Still, at the same time, they are associated with high computational costs, especially in high-dimensional cases or cases that require many iterations to obtain a solution. Such added complexity is likely to undermine the compactness of LOBLAO, resulting in it benefitting neither real-time nor large-scale optimization problems.

Another limitation relates to the competencies of the method since it is problem-dependent. Even though LOBLAO was successful at the benchmarks and in the clustering problems that were tested, the performance level achieved may not even be useful when considering other formulations, such as dynamic optimization problems and multi-objective optimization problems that were beyond the scope of this paper. This underscores a gap that this paper’s results cannot fill, hence a recommendation for more studies to examine its performance over more complex optimization problems.

In addition, the performance of the algorithm is sensitive to parameters such as the rate of mutation and the size of the population. Therefore, in order to obtain optimal results, these meta-heuristic algorithms will have to be finely tuned, a task that is easy for an experienced user and difficult for a new user. This sensitivity may, in turn, affect the wider use of the algorithm in real-life applications.

Finally, the evaluation method employed in this study centers on mathematical benchmark functions and typical clustering data sets. Even though these benchmarks solve general optimization problems, They often over-simplify the issues encountered in practice, such as those problems that are corrupted by noise or some uncertainty in data.

Understanding these drawbacks provides a basis for further work, making it possible for LOBLAO to be more num resistant and versatile, enabling it to solve even more difficult problems.

Conclusion and future works

The clustering of data can be considered one of the most complex problems in the optimization domain as it has to be accompanied by strong and efficient search techniques for it to source competent solutions. Apart from clustering, other forms of optimization problems are found, for instance, in mathematics, engineering, data mining, and the Internet of Things. The Aquila Optimizer developed based on the hunting and searching features of the Aquila in nature, has posed to be one of the solutions to such problems. Nonetheless, it did not fare well in solving complicated and high-dimensional optimization problems; thus, a better variant was developed.

In this paper, we develop the topic of the improvement of two algorithms by integrating: first, the Locality Optimised Based Learning Opposition Aquila Optimiser (LOBLAO) into the Aquila Optimiser. Moreover, they combine the newly developed algorithms, called OBL, aimed at improving the diversity of the solutions, thus providing a better exploration-exploitation tradeoff. Furthermore, the MSS also prevents early convergence by expanding the search space and finding new regions of search. They performed a large number of experiments, and LOBLAO was found to exhibit better search capabilities on twenty-three benchmark functions and eight common data clustering problems. The efficiency of LOBLAO was demonstrated in numbers, and its results were most of the time between the top ones when compared against the leading methods such as the Arithmetic Optimization Algorithm, Salp Swarm Algorithm, Whale Optimization Algorithm, and many others.

This paper highlights the potential of the LOBLAO as a good optimization approach, but there is scope to explore its efficacy further using more algorithms. Further work can aim at enhancing the performance through the use of more sophisticated search techniques for example, hybrid metaheuristic approaches or procedures for dynamic parameter tuning. Furthermore, the research could also focus on dynamic approaches to learning that will improve the performance of LOBLAO for dynamic or real-time optimization problems.

Also, LOBLAO’s application can be extended to tackle more complicated, real-world problems. Such examples comprise classification and feature selection tasks, parameter estimation for renewable energy systems like solar cells, task scheduling in distributed systems, optimization problems in big data, and complex engineering designs. The application of LOBLAO under large-scale and multi-objective optimization eventually extends its coverage abstraction and performance in reality.

By pursuing these options, the proposed technique may evolve into an invaluable device with high applicability in various optimization tasks. Such developments would serve to bolster LOBLAO’s status as a benchmark in the optimization field and pave the way to new approaches for solving intricate real-life problems.