Abstract
The Dung Beetle Optimization (DBO) algorithm is a relatively recent metaheuristic known for its simplicity, versatility, and low parameter dependence, making it a valuable tool for solving complex optimization problems. Despite its potential, DBO suffers from limitations such as slow convergence and premature stagnation in local optima. To address these critical issues, this paper introduces a novel enhanced variant named Elite Bernoulli-based Mutated Dung Beetle Optimizer with Local Escaping Operator (EBMLO-DBO), specifically designed to improve the convergence speed, search capability, and robustness of the original DBO algorithm. The motivation for this enhancement stems from DBO’s limited performance in high-dimensional and non-convex problems, where it often fails to maintain an effective balance between exploration and exploitation. The novelty of the proposed EBMLO-DBO lies in the integration of four key strategies tailored to overcome these weaknesses: (i) Bernoulli map-based initialization to enhance population diversity and ensure a better global search foundation; (ii) Morlet Wavelet mutation to introduce adaptive local refinements and help the algorithm escape local optima; (iii) elite guidance to accelerate convergence by directing the population toward high-quality regions; and (iv) a local escaping operator (LEO) to dynamically refine the search process and strengthen exploitation without sacrificing exploration. The performance of EBMLO-DBO is rigorously validated using the CEC2017 and CEC2022 benchmark suites, where it achieves Friedman ranks of 1.83 and 2.7 respectively, consistently surpassing eleven state-of-the-art algorithms including PSO, HHO, WOA, and advanced methods like CMAES and IMODE. In benchmark function optimization, EBMLO-DBO demonstrates superior performance by achieving first rank in 50% of CEC2022 functions and obtaining the lowest average fitness values in 18 out of 29 CEC2017 functions. For photovoltaic parameter estimation applications, EBMLO-DBO exhibits exceptional accuracy with RMSE values of 9.8602E-4 for single diode models, 9.81307E-4 for double diode models, and 2.32066E-3 for PV module models, achieving top performance ranks of 1.45, 1.42, and 1.74, respectively. Statistical analysis using Wilcoxon signed-rank test at significance level \(\alpha =0.05\) confirms the significant superiority of EBMLO-DBO over all compared algorithms, thereby validating the effectiveness and reliability of the proposed enhancements. Overall, the results state that EBMLO-DBO offers a significantly improved search performance and solution quality compared to the original DBO and related methods, thereby justifying the necessity and effectiveness of the proposed enhancements.
Similar content being viewed by others
Introduction
Global optimization has become a necessary component in various scientific disciplines in recent years. The real-world problems’ complexity has increased in recent decades, raising significant concerns in the optimization field. It is possible to view almost all issues in the actual world as optimization problems. The purpose of high-performance algorithms is to manage potentially difficult problems. Consequently, scientists are trying to create effective optimization strategies for these types of issues. Many optimization techniques have been introduced and modified to provide the best solutions for a range of problems in a variety of fields, including economics, engineering, machine learning, energy, medical, networks, and more1,2. According to the characteristics of their behaviors, these algorithms can be categorized as either metaheuristic or deterministic3. Within deterministic methodologies, the solution obtained from the preceding iteration is employed to deduce the revised solutions for the present iteration. The initial solution may affect the use of deterministic methods, which may affect the result. Furthermore, the search space gradient might be required in some circumstances. Besides, the solutions may converge to local optima4. Hard optimization problems pose serious obstacles in front of the deterministic methods which provides comparatively low success and efficiency rates. Therefore, new methodologies must be developed for optimization capable of determining near-optimal solutions when only limited computational effort is available. Over the last twenty years there has been an ever-increasing interest by the researchers in applying metaheuristic algorithms (MAs) to solve real-world problems5,6. Ease and simplicity, ease of usage, flexibility, and effective efficiency in computation in their application have attracted the interest of researchers towards the application of MAs for solving practical issues7. Also, since MAs do not rely on any operations involving gradients, it can work with less memory. It has the ability to quickly and simply address a variety of problems8.
It follows that the categories of MAs are human-activities based algorithms, natural law-based algorithms, evolution-based algorithms, and group-intelligence based algorithms9,10. Evolutionary algorithms are Genetic Algorithm (GA)11, for instance, quantum evolutionary algorithm (QEA)12, and the evolutionary programming (EP)13. Group-intelligence based algorithms are simulated on the routines of various natural swarming species. For instance, the Particle Swarm Algorithm (PSO)14, the Secretary Bird Optimizer (SBOA)15, the Nutcracker Optimizer (NOA)16, the Sled Dog Optimizer (SDO)17, the Coati Optimization Algorithm (COA)18, the Snake Optimizer (SO)19, the Crayfish Optimization Algorithm20, the Parrot Optimizer21, the Elk Herd Optimizer22, the Dung Beetle Optimizer (DBO)23, and the Sea Horse Optimization Algorithm (SHO)24, among others.
Furthermore, the third category of algorithms can be categorized into physics, and mathematical algorithms. The mathematical inspired algorithms include Sinh-Cosh algorithm25, runge kutta optimization algorithm (RUN)26, the Triangulation Topology Aggregation Optimizer (TTAO)27, and Newton Raphson Optimizer (NRO)28. In addition, the physics-based algorithms include Polar Lights Optimizer (PLO)29, Kepler law optimizer (KOA)30, geyser inspired algorithm31, lightning search algorithm (LSA)32, Fick’s Law optimizer (FLA)33. A group of intelligent algorithms has also been proposed that are based on the activities of human beings. These include the political optimizer (PO)34, the Poor and rich optimization algorithm (PRO)35, Human memory search algorithm (HMO)36, Human evolutionary algorithm (HEO)37, the Mother optimization algorithm (MOA)38 and the Skill optimizer (SOA)39 and etc.
Since MAs have efficient and flexible computational performance, they are frequently applied in practical optimization scenarios across a wide range of fields40,41,42,43. In recent years, the expansion of numerous disciplines and the advancement of technology and modern science have led to the emergence of complex, non-convex and large-scale problems44. While many intelligent algorithms have proven to be more effective than one another in cracking actual optimization issues, they still encounter challenges such as inadequate convergence, limited exploration capability, and performance decline when confronted with increasingly difficult and intricate problems across different fields45,46. Furthermore, the "No Free Lunch" (NFL) theory asserts that there is no algorithm now in use that can consistently outperform all other algorithms in any optimization situation that may arise47. Based on this investigation, it is critical to investigate quick and effective solutions for real-world application instances. Thus, it is necessary to continuously improve and enhance the algorithms’ performance in order to deliver more precise and ideal solutions to actual optimization scenarios.
On the other hand, the Dung Beetle Optimizer (DBO) is a new MA that was presented in 2023 and relies on the population notion to solve multiple problems23. The foraging habits of dung beetles served as the model for the DBO algorithm23. The DBO has fewer parameters and is much more resilient than other MAs48. The primary motivation for developing a novel EBMLO-DBO stems from critical limitations observed in current MAs, particularly the recently proposed DBO, which significantly impact their practical applicability in complex engineering problems. Despite DBO’s demonstrated efficiency in basic optimization scenarios, four fundamental weaknesses limit its effectiveness in high-dimensional and non-convex optimization problems. First, premature convergence occurs when dealing with complex multimodal landscapes, leading to suboptimal solutions and inadequate exploration of the search space, particularly evident in problems with numerous local optima. Second, existing algorithms including DBO fail to maintain an effective balance between exploration and exploitation phases, often either getting trapped in local regions or spending excessive computational resources on unnecessary exploration. Third, conventional initialization strategies rely heavily on random generation, which may not provide sufficient population diversity for comprehensive search space exploration, limiting the algorithm’s ability to locate promising regions efficiently. Fourth, the absence of effective local optima escape mechanisms restricts the algorithm’s capability to find global optima in challenging optimization scenarios, particularly in real-world applications like photovoltaic parameter estimation where high precision and reliability are crucial for practical implementation. Moreover, the "No Free Lunch" (NFL) theorem states that it is impossible for any MA to handle all optimization problems47. This implies that while the MA may provide good solutions for certain optimization problems, it may perform poorly for other problems. Therefore, appropriate, and efficient solutions are required to enhance the algorithm’s efficiency. Also, researchers are continuously devising novel and efficient algorithms as well as refining already-existing optimization algorithms to better tackle real-world engineering optimization problems. Consequently, pertinent literature is required to examine and enhance DBO. These justifications drive the improvement of a new Dung beetle optimization algorithm in this study. These limitations directly impact the performance and applicability of optimization algorithms in engineering applications, motivating the systematic development of EBMLO-DBO with targeted enhancement strategies designed to address each identified weakness comprehensively.
According to these motivations, this paper presents the enhanced Dung beetle optimization algorithm EBMLO-DBO with stronger performance, based on the aforementioned motivation. It does this by integrating four primary strategies including the Bernoulli search map, Elite guidance strategy, mixed mutation strategy and Local escape operator into the original DBO algorithm. The Bernoulli map is used to provide diverse population-based agents at the initialization phase which provide more away regions into the race of search process. Like many other MA, DBO agents are susceptible to be trapped in local solutions which affect the search ability of DBO, so Morlet mutation is presented to mutate the population and provide mutated enhanced agents to escape from the local optima. The majority of the MA’s processes are exploration and exploitation; the elite guidance strategy can enhance the quality of solutions presented and enrich the population’s exploration routes before iteration begins. As iteration progresses, the exploitation operator gradually gains the upper hand, improving the efficiency of global exploration. Furthermore, we present the local escape operator is used to enhance the exploitation ability of DBO and provide more accurate and high-quality solutions for the next generations which boosts the convergence speed at early iterations and remove the detrimental effect of invalid solutions. This study evaluates the enhanced EBMLO-DBO algorithm in comparison to a variety of advanced and traditional algorithms on different test sets CEC’17 and CEC’22. The statistical analysis is conducted with the Friedman rank and Wilcoxon rank sum tests. The results indicate a substantial enhancement in the performance of the DBO algorithm, and EBMLO-DBO shows notable competitiveness in comparison to other methods. Furthermore, the enhanced EBMLO-DBO method is utilized to address a practical issue of estimating the parameters of three photovoltaic (PV) models, which consist of single and double diode as well as PV module. Moreover, the EBMLO-DBO algorithm is more stable in comparison to most algorithms; thus, the solutions present higher accuracy and are likely to provide a suitable development methodology.
Objectives of the paper
The objectives of the paper are:
-
(1)
To develop EBMLO-DBO algorithm by integrating four novel enhancement strategies: Bernoulli map-based initialization, Elite Leadership Strategy, Morlet Wavelet mutation, and Local Escaping Operator to overcome DBO’s limitations in high-dimensional optimization problems;
-
(2)
To integrate Bernoulli map that replaces conventional random initialization with chaotic sequences providing superior search space coverage, while the Elite Leadership Strategy dynamically adjusts selection probability between elite random leadership and best leadership to maintain optimal exploration–exploitation balance throughout optimization;
-
(3)
The proposed mutation strategy dynamically adjusts mutation space using iteration-dependent scaling parameters, providing strong exploration in early stages and fine-tuning capabilities in later optimization phases. The utilization of the Local Escaping Operator (LEO) to enhance solution quality and accelerate convergence through strategic position updates based on elite guidance;
-
(4)
And, to demonstrate practical engineering applicability of EBMLO-DBO through successful implementation in photovoltaic parameter estimation problems for single diode, double diode, and PV module models, establishing the algorithm’s real-world value in renewable energy system optimization and parameter identification applications.
Organization of paper
The rest of the paper is organized as: Related works to the proposed algorithm are presented in Section "Literature review". The mathematical model and the main description of the original DBO algorithm is presented and discussed in Section "Related terminologies". Section "Proposed methodology: EBMLO-DBO algorithm for global complex problem and parameter estimation of solar photovoltaic model" presents the main proposed improvements integrated with DBO to enhance and boost its convergence and solution quality. Section "Experimental results" displays the conducted experiments for EBMLO-DBO using global optimization sets of functions while Section "PV parameter estimation using EBMLO-DBO" test and evaluate the EBMLO-DBO for solving a real problem for estimating the parameters of three PV models. The main discussions related to the overall presented experiments are presented in Section "Discussion". And, finally, section "Conclusion and future scope" concludes the paper with future scope.
Literature review
In recent years, the field of PV system optimization has seen significant advancements through the development and application of various MAs. These approaches have been pivotal in addressing the challenges of parameter estimation and optimization in PV models, leading to improvements in accuracy, convergence, and robustness. This section summarizes key contributions from the literature related to the optimization problems and PV models, providing a foundation for identifying the research gaps addressed by our proposed work.
Zhu et al.49 proposed a multi-strategy DBO algorithm that integrated both quantum computing and multi-strategy fusion and was effective with a number of complex engineering problems. Mai et al.50 adopted the advanced variant of DBO to optimize wind speed forecasting, outperforming many other techniques such as inverse learning and Fuch chaotic functions.
Li et al.51 added a new version of DBO to optimize the parameters of LSTM models for identification of concrete dynamic principles, including integration with some novel methods like the greedy lens imaging back-learned one. The DBO algorithm with OBL improved by Zilong and Peng52 significantly outperformed the results on CEC’17 benchmarks and other engineering problems.
Zhong et al.53 proposed an enhanced variant of the Sine Cosine Algorithm by introducing a hierarchical multi-leadership search strategy to address limitations in population diversity and exploration–exploitation balance, enabling parallel guidance across multiple search trajectories. The algorithm demonstrated superior optimization performance across 18 classical benchmarks and 30 CEC’17 functions, consistently surpassing other state-of-the-art MAs. HMLSCA was also successfully applied to optimize SVM parameters and feature weights for medicine data classification across eight datasets, achieving the top Friedman rank of 1.00, and attained a 98% accuracy rate in COVID-19 diagnosis, validating its effectiveness in real-world applications.
Parizi et al.54 presented a hybrid MA approach by combining the exploration strength of the Woodpecker Mating Algorithm with the exploitation efficiency of the Sine Cosine Algorithm, enhanced further through Levy flight-based modifications and a shared local memory mechanism to dynamically balance search behaviors. The algorithm was validated on 28 complex benchmark functions, where it consistently outperformed recent metaheuristic algorithms in solving nonconvex and inseparable optimization tasks. Additionally, HSCWMA was applied as a trainer for Multi-Layer Perceptron networks in software development effort estimation problems, achieving superior results across three real-world datasets, demonstrating its adaptability and high-performance potential.
Karimzadeh Parizi et al.55 introduced an enhanced version of the Woodpecker Mating Algorithm by incorporating three key mechanisms: Distance Opposition-Based Learning (DOBL) for improved exploration and diversity, a local memory strategy to enhance exploitation, and a self-adaptive adjustment of the \(H\alpha\) parameter to optimize the Running Away function’s effectiveness. The algorithm demonstrated strong performance across 23 benchmark functions, surpassing recent metaheuristic techniques in terms of convergence reliability and search efficiency. Additionally, OWMA was successfully applied as a trainer for Multi-Layer Perceptron neural networks on multiple biomedical and regression tasks, and further validated on five real-world constrained optimization problems, confirming its robustness and versatility in complex scenarios.
Xiao et al.56 proposed an advanced version of the Snow Ablation Optimizer by incorporating Good Point Set initialization, greedy selection, Differential Evolution-based search, and Dynamic Lens Opposition-Based Learning to overcome premature convergence, enhance diversity, and improve accuracy in complex high-dimensional problems. Extensive comparisons on the IEEE CEC2017 and CEC2022 benchmark functions showed that MSAO outperformed classical SAO and several leading algorithms, achieving the best overall performance with Friedman mean rankings of 1.66 and 1.25, respectively. Moreover, the proposed MSAO demonstrated strong practical effectiveness in solving six constrained engineering design problems and a photovoltaic model parameter estimation case, confirming its broad applicability and high optimization capability57.
A guaranteed convergence arithmetic optimization algorithm combined with a modified Newton–Raphson technique, GCAOAEmNR, for parameter optimization of PV models, was proposed by Ridha et al.58, indicating very high accuracy and stability under various conditions. Similarly, El-Dabah et al.59 used the Northern Goshawk Optimization algorithm for parameter estimation in triple diode PV models. The results demonstrated increased convergence speed and improved solution accuracy when it was applied to a few commercial PV modules.
Abbassi et al.60, presented an enhanced Arithmetic Optimization Algorithm (AOA) elaborated for parameter extraction in photovoltaic models, offering better applicability over varied conditions. The approach was applied to single and double diode PV models and gave results superior to other algorithms that already exist.
Long et al.61 presented the hybrid Seagull Optimization Algorithm for estimating PV parameters, much superior to the available algorithms, since it exhibits better precision and the power to avoid local optima. Janamala and Radha Rani62 presented an Archimedes Optimization Algorithm for optimal placement of solar photovoltaic with minimum grid dependence and improved voltage profiles.
In this respect, Lu et al., in63, proposed the HMSCPSO algorithm to enhance global search and avoid a possible local optima for an improved accuracy of PV model parameter estimation. Another option was NOA algorithm proposed by Duan et al.64 for extracting the PV parameters that present improved results against PSO, FWA, and WOA in terms of accuracy and efficiency of the parameter extraction process.
Ramadan et al. in65 presented the Enhanced Harris Hawk Optimization (EHHO) algorithm used to optimize PV model parameters by enriching the search phase through the incorporation of sine and cosine functions, leading to improved accuracy for multiple PV models. Additionally, Yousri et al.66 developed an Adaptive Fractional-order Archimedes Optimization Algorithm (A-FAOA), which outperformed state-of-the-art techniques for finding the best PV parameters under different scenarios.
Mohamed et al.67 proposed a Hybrid Kepler Optimization Algorithm (HKOA) for enhancing parameter estimation within PV models, while being able to balance exploration and exploitation effectively, outperforming existing advanced methodologies in the process. Kullampalayam Murugaiyan et al.68 introduced the Opposition-Based Exponential Distribution Optimizer (OBEDO) applied for photovoltaic parameters extraction which enhanced computational efficiency and accuracy.
A better algorithm in PV model parameter extraction was presented by Izci et al.69, which combines random learning with a logarithmic spiral search algorithm, and it was called En-PDO. A better accuracy in the estimation of the PV parameter was shown by Saadaoui et al.70 by combining BSA with DE.
Ekinci et al.71 proposed a hybrid Gazelle-Nelder-Mead algorithm for PV parameter extraction, showing its convergence and accuracy for the superiority of many benchmark tests. Qaraad et al.,72, proposed the QPSOL technique, based on PSO, increasing accuracy and efficiency of solution in the estimation of PV parameters.
Javed et al.73 suggested MPPT-based Flying Squirrel Search Optimization for hybrid PV-TEG systems. The algorithm provided very good efficiency and fast response in tracking. On the other hand, Kahraman et al.74 introduced a dynamic-fitness-distance-balance technique into the Stochastic Fractal Search algorithm, which led to increased accuracy and robustness in the estimation of photovoltaic parameters. Izci et al.75 developed a pattern search algorithm that would overcome the Mountain Gazelle Optimizer problems regarding low parameter estimation accuracy and reliability.
While important contributions have been made in putting metaheuristic algorithms to the estimation of PV parameters, no solid trade-off has been obtained so far between global exploration and local exploitation, especially in what concerns DBO variants. The proposed contributions always provide less-than-satisfactory convergence speed and accuracy, with inconsistent performance across different PV models. The proposed work aims to address these gaps by introducing a novel DBO variant (EBMLO-DBO) designed to enhance global optimization and parameter estimation accuracy in single, double, and PV models, thereby improving the overall efficiency and reliability of PV system modeling.
Related terminologies
This section gives definitions to the most important terms connected with the proposed work as well as provides a broad overview of the original DBO algorithm and the reason why and how PV model parameters are determined in this procedure.
Dung beetle algorithm (DBO)
A new algorithm called the DBO was presented in 202323. Its goal is to find the best solutions through a variety of activities like ball rolling, dancing, foraging, breeding, and thieving. Therefore, there are four stages in this algorithm: foraging, breeding, rolling balls, and stealing.
Population initialization
Like any swarm-based optimization algorithm, DBO’s first phase is the population initialization which comprises the set of agents/ solutions used to crack the optimization problem. Random initialization is used in the original DBO to initialize the values of starting solutions in DBO. For a set of \(N\) solutions and each solutions comprises a set of \(D\) dimensions that represent the dimensionality of the cracked problem, the initialization phase can be represented as follows:
where the solution within the population is denoted by \({X}_{i}\) and the limits for the search pace of the cracked problem are defined by \(Lb\) and \(Ub\). In addition, a random value is generated between 0 and 1 using the built-in function \(rand()\)
Roller-ball dung beetle
In its native habitat, the dung beetle relies on the position of the sun to maintain a straight course when rolling a ball of excrement. During the rolling of the dung ball, the location of the dung beetle was updated using Eq. (2):
The current iteration and the current position of the dung beetle are represented by t and \({X}_{i}(t)\), respectively, in accordance with Eq. (2). The dung beetle’s divergence from the average direction is defined by the parameter \(\alpha\). It accepts two random numbers, 1 and -1, where 1 denotes no deviation and -1 denotes a deviation. The defect factor, with a default value of 0.1, is defined by the variable \(k\), which ranges from 0 to 0.2. It is assumed that the constant \(b\) has a value of 0.3 and is restricted to the interval [0, 1]. Since the simulation of solar illumination is described by \(\Delta x\) and the worst global value is \({X}^{w}\), a higher \(\Delta x\) value would indicate a greater distance from the light source. To simulate the potential of hitting an obstruction while rolling a ball, an obstacle-hitting event is used. To mimic the dancing behavior of dung beetles, the tangent function is utilized to determine a new path for the ball to roll in the event of an impediment. Consequently, Eq. (3) describes the rolling beetle’s position:
The range of 0 to \(\pi\) encompasses the parameter θ and the position is not updated when it takes the values 0, \(\pi /2\), and \(\pi\) because there is no position change.
Spawning dung beetles
Dung beetles typically reproduce in safe places based on their natural surroundings. This approach, which is described as in Eq. (4), takes into account the selection of safe borders in the surroundings to symbolize secure zones:
The variables \(L{b}^{*}\) and \(U{b}^{*}\), which define the lower and upper bounds, respectively, define the spawning region limits. The current local optimum is indicated by the variable \({X}^{*}\). The convergence parameter \(R\) is defined as \(1-t/T\), where \(T\) is the maximum number of iterations. The dung beetle will therefore choose where to lay its egg within this predetermined range when it discovers the ideal location for breeding. In order to prevent the search from becoming isolated while trying to locate the current best solution, the spawning area is dynamically set so that the best position remains within the range of interest. The spawning beetle’s position update is explained as follows:
The two randomly generated vectors, \({b}_{1}\) and \({b}_{2}\), each have a size of \(1\times D\), where \(D\) is the size of the optimization issue.
Foraging dung beetles
In their natural environment, the dung beetles forage in close proximity to the designated region when they choose a safe spot, which is close to their egg-laying process. The original study suggests the following formula for a new definition of this area:
From this angle, \({X}^{b}\) would represent the ideal global position. The boundaries of the foraging area are denoted by \(L{b}^{b}\) and \(U{b}^{b}\), respectively. Furthermore, the range of the problem-solving domain is represented by \(Lb\) and \(Ub\). Equation (7) describes how the foraging dung beetle updates its position each time it engages in foraging behavior as follows:
where \({C}_{1}\) has a normal distribution and \({C}_{2}\) is a vector of size \(1\times D\), and \(D\) is the dimensionality of the problem and whose elements are between 0 and 1.
Stealing dung beetles
In their natural habitat, dung beetles on locations are not directly taken from other dung beetles; instead, they move one another in a competition for ownership. As a result, the authors estimated this behavior in the initial search using the best global location, or \({X}^{b}\). The thieving dung beetle repositions itself in such a way that Eq. (8) specifies the location update:
Where \(g\) be the random variable’s magnitude and let \(S\) be a constant with a fixed value of 0.5 in Eq. (8). Six dung beetles are rolling dung, seven are foraging, eleven are thieving, and six are reproducing, according to the original paper.
PV parameter estimation problem
Determining parameters in PV models is typically approached as an optimization problem aimed at minimizing the disparity between simulated results from estimated parameters and experimental data76. The objective of this optimization is to quantify the discrepancy between the current reported by the manufacturer and predicted current by the model. Various algorithms are utilized to find parameter sets that reduce this difference. Solar cell modeling involves three main steps: selecting appropriate equivalent models, developing mathematical models, and accurately determining cell parameter values. Firstly, identifying the electrical characteristics of a PV system is to carefully select the appropriate models. The double-diode and single-diode models are used to define the electrical properties of PV modules and cells77.
-
The single diode is shown in Fig. 1. This model consists of five fundamental parameters: the \({I}_{rsD}\) which represents the diode reverse saturation current, \({I}_{PC}\) that represents the photovoltaic current, \({R}_{se}\), \({R}_{SH}\) which are the series and shunt resistances, respectively and \(N\) denotes the ideality factor. Figure 1 depicts the lighted solar cell as a composite of two resistors, a diode and the current source. The series resistor symbolizes the resistance that is faced in the route of the current, whereas the leakage current is demonstrated by the shunt resistor. The series resistance comprises the contact resistance, electrode resistance, and resistance from the material itself. The output current \({I}_{out}\) can be calculated using Kirchhoff’s current law and is specified as follows:
$${I}_{\text{out }}={I}_{PC}-{I}_{rsD}\left({e}^{\frac{Q\left({V}_{l}+{I}_{\text{out }}{R}_{\text{se }}\right)}{(NKT)}}-1\right)-\frac{\left({V}_{l}+{I}_{\text{out }}{R}_{\text{se }}\right)}{\left({R}_{SH}\right)}$$(9)
According to Eq. (9), \({V}_{l}\) represents the resultant voltage, \(K\) symbolizes the Boltzmann constant, \(Q\) symbolizes the electric charge of an electron, and \(T\) represents the temperature measured in Kelvin. Hence, it is necessary to ascertain the values of the five unidentified parameters, namely \({I}_{rsD}\), \({I}_{PC}\), \({R}_{se}\), \({R}_{SH}\), and \(N\), for the single diode model. Optimization approaches can be employed to precisely estimate these parameters in order to appropriately depict the solar cell’s efficiency. Accurate and strong parameter estimation is crucial for obtaining dependable outcomes.
-
The single-diode model has been broadly utilized to analyze the solar cells’ static characteristics because of its accuracy and simplicity. Nevertheless, this model is limited due to its assumption of a constant diode ideality factor throughout the whole range of output voltage. The diode ideality factor changes in response to the voltage applied across the device. In addition, the single-diode model fails to account for the effect of current recombination losses in the depletion zone. To tackle these concerns and improve the authenticity of the model, a double-diode model can be employed. Figure 2 demonstrates the inclusion of an extra recombination diode that runs parallel to the original diffusion diode in this model. The double-diode model’s output current is determined by the following equation:
$$\begin{array}{l}{I}_{\text{out }}={I}_{PC}-{I}_{rsD1}\left({e}^{\frac{Q\left({V}_{l}+{I}_{\text{out }}{R}_{se}\right)}{(N1KT)}}-1\right)-{I}_{rsD2}\left({e}^{\frac{Q\left({V}_{l}+{I}_{\text{out }}{R}_{se}\right)}{(N2KTT)}}-1\right)\\ -\frac{\left({V}_{t}+{I}_{\text{out }}{R}_{\text{se }}\right)}{\left({R}_{SH}\right)}\end{array}$$(10)
In this case, the diffusion current is denoted by \({I}_{rsD1}\) while the saturation current is represented by \({I}_{rsD2}\). Moreover, \({N}_{1}\) and \({N}_{2}\) represent the ideality factors for the diffusion and saturation currents, respectively. According to Eq. (10), the seven parameters that need to be estimated for the double-diode model are, \({N}_{1}\), \({I}_{rsD1}\), \({I}_{PC}\), \({R}_{SH}\), \({I}_{rsD2}\), \({R}_{se}\) and \({N}_{2}\). These parameters are crucial for accurately modeling the double-diode behavior.
-
Finally, the Fig. 3 displays the layout of the PV module model. In this case, many solar cells are configured in parallel or series organization. Equations (11) and (12) can be used to describe the output current of PV modules based on both the double-diode and single-diode models, respectively.
$$\begin{array}{l}{I}_{\text{out }}\\ ={C}_{p}{I}_{PC}-{C}_{p}{I}_{rsD1}\left({e}^{\frac{Q\left({V}_{l}+{C}_{s}{I}_{\text{out }}{R}_{se}/{C}_{p}\right)}{\left(N1KT{C}_{s}\right)}}-1\right)\\ -{C}_{p}{I}_{rsD2}\left({e}^{\frac{Q\left({V}_{l}+{C}_{s}{I}_{\text{out }}{R}_{se}/{C}_{p}\right)}{\left(N2KT{C}_{s}\right)}}-1\right)-\frac{\left({V}_{l}+{I}_{\text{out }}{R}_{se}{C}_{s}/{C}_{p}\right)}{\left({R}_{sH}{C}_{s}/{C}_{p}\right)}\end{array}$$(11)$${I}_{\text{out }}={C}_{p}{I}_{PC}-{C}_{p}{I}_{rsD}\left({e}^{\frac{Q\left({V}_{l}+{C}_{s}{I}_{\text{out }}{R}_{se}/{C}_{p}\right)}{\left(NKT{C}_{s}\right)}}-1\right)-\frac{\left({V}_{l}+{I}_{\text{out }}{R}_{se}{C}_{s}/{C}_{p}\right)}{\left({R}_{sH}{C}_{s}/{C}_{p}\right)}$$(12)
where the number of parallelly configured solar cells are denoted by \({C}_{p}\), while the number of serially configured solar cells are denoted by \({C}_{s}\). Therefore, the five parameters that must be estimated for the single-diode model-based PV module are \(N\), \({I}_{rsD}\), \({I}_{PC}\), \({R}_{se}\) and \({R}_{SH}\) while the seven parameters required to be estimated for the double-diode model-based PV module are \({I}_{rsD1}\), \({I}_{PC}\), \({R}_{SH}\), \({I}_{rsD2}\), \({N}_{1}\), \({R}_{se}\), and \({N}_{2}\).
Proposed methodology: EBMLO-DBO algorithm for global complex problem and parameter estimation of solar photovoltaic model
Among global search algorithms, the DBO algorithm stands out due to its major features: simplicity, minimum number of parameters and fast calculation power. In certain cases, its capacity to converge rapidly after a limited number of iterations allows it to find the global optimum. Diversity declines as iterations go on because agents in the population try to gravitate toward the best location. It might thus be vulnerable to local optima, which would prevent it from reaching optimal solutions. Reduced computational precision and sluggish convergence are the results of this contributing factor. Early on, good performance was guaranteed by the founding population’s equitable distribution and high quality. As a result, the Bernoulli map (BM) is utilized to provide diverse solutions and enhance the exploration phase of DBO. Furthermore, to improve the EBMLO-DBO’s exploitation potential and expand the search space, the Morlet Wavelet mutation (MWM) strategy is utilized. As a result, it improves convergence rate and computational accuracy. In addition, at the end of each EBMLO-DBO iteration, the Local Escape Operator (LEO) strategy is applied to improve the quality of solutions and establish a more optimal trade-off between investigating novel possibilities and utilizing already-existing solutions in the domain-based optimization process. Finally, the elite leadership strategy (ELS) is proposed to enhance the population depending on more elite with lower fitness values agents. The details of the proposed strategies are discussed in the following subsections.
System model
This section explains the four main strategies applied to DBO algorithm in detail. Then, the working principle of the EBMLO-DBO algorithm is outlined to give a general overview for the proposed algorithm in its inherent working principle. Finally, the computational complexity of the proposed work is provided in terms of the time and space complexity.
Bernoulli map-based population initialization (BM)
MAs rely on iteratively refining candidate solutions, so their initial values are crucial for their performance in terms of convergence and exploration78. Commonly, the initialization phase of these algorithms involves generating random values within a specified range, often using a Gaussian distribution. This approach, however, depends significantly on randomness, which can hinder deeper insight and progress. Alternatively, chaotic maps can produce chaotic sequences—random-like sequences generated by simple deterministic systems. These maps exhibit properties such as non-linearity, sensitivity to initial conditions, randomness, ergodicity, fractional stability, chaotic attractors, local instability, long-term unpredictability, and global stability. In optimization, chaotic maps are often preferred over pseudo-random numbers for generating values between 0 and 1. Studies have demonstrated that employing chaotic sequences for tasks such as population initialization, selection, crossover, and mutation can enhance algorithm performance and frequently results in better convergence than using random numbers79. An example of a chaotic system is the Bernoulli map, a segmented chaotic system defined as follows:
where α, the mapping parameter, is between 0 and 1 which is usually set to 0.2980. One unique instance where these maps exist is:
Therefore, the novel population initialization approach can be modeled as:
Therefore, new population initialization strategy using Bernoulli mechanism to enhance the diversity of the initial population. This approach leverages the chaotic properties of Bernoulli maps to ensure a more varied set of candidate solutions from the start, which significantly improves the exploration capabilities of the DBO algorithm. By increasing the diversity at the initialization phase, this strategy helps prevent premature convergence and enhances the algorithm’s overall performance.
Elite leadership strategy (ELS)
Upon each iteration’s completion, the elite leadership strategy81 divides the current population \(X(t)\) into two subsets: the elite group \(Y1(t)\) and the non-elite group \(Y2(t)\). This division ensures that \(Y1(t)\) and \(Y2(t)\) are distinct and together comprise \(X(t)\) where \(Y1\left(t\right)\cup Y2\left(t\right)=X(t)\) and \(Y1\left(t\right)\cap Y2\left(t\right)=\varnothing\). The size of the elite group is denoted as \(\Gamma\). Three random members \({X}_{r1}\), \({X}_{r2}\), and \({X}_{r4}\) are picked from the elite group \(Y1(t)\), while \({X}_{r3}\) and \({X}_{r5}\) are selected from the non-elite one \(Y2(t)\). Therefore, the elite leadership strategy is formulated as:
According to Eq. (16), \({X}_{best}\) is the top individual of generation \(t\), \(F\) represents the scaling factor, and rand is a random value between 0 and 1. The selection probability \(\tau\) is calculated as:
The selection in Eq. (16a) utilizes a process where individuals from \(Y1\) are picked at random to provide guidance to the population. This selection process is referred to as elite random leadership. On the other hand, Eq. (16b), which utilizes the most optimal individual to generate new individuals, is referred to as best leadership. If the mutation strategy gives higher importance to global exploration, the algorithm may move towards the global best solution, but it will need a greater number of iterations. On the other hand, focusing on exploiting local resources can decrease the variety of search results, which in turn hampers the ability to explore new possibilities. As stated in Eq. (17), at the start of the iteration, the selection probability for elite random probability is almost 1, whereas the selection probability for optimal leadership is nearly 0. As the iterations progress, the algorithm gradually approaches the optimal solution by raising the probability of selecting the best leadership and decreasing the probability of selecting random leadership from the elite group. During this process of mutation, \(Y1\) offers a favourable path for speeding up convergence, while \(Y2\) aids in adjusting the search orientation, so improving population variety to avoid premature convergence. Finally, if the new position \(Xnew1\) improves the objective function value of the current population solutions, as detailed in Eq. (18), it is accepted for replacement of the current solution as follows:
The ELS strategy plays a critical role in enhancing the performance of the DBO algorithm. The population is split into elite and non-elite groups, and scope for mutation is induced by choosing certain individuals which guarantees that both exploration and exploitation strategies are maintained. The early usage of elite random leadership allows exploring diversity and global search during the initial phases while over time starting to introduce the better leadership allows to better progress the search towards better solutions. This dynamic adjustment helps reduce or minimizes the possibility of early convergence as well as enhances the performance of the solutions when solving local optima. As a result, the ELS strategy considerably raises the efficiency and effectiveness of the EBMLO-DBO algorithm at solving the problem of high-quality solutions.
Morlet wavelet mutation strategy (MWM)
Swarm intelligence optimization algorithms rely on jumping out of the local optimum when solving problems with dense distributions of extreme points. The primary driver of biological evolution is mutation10. In order to increase the stability of the solution, wavelet mutation (MWM) can dynamically modify the mutation space using the wavelet function’s translation and scaling capabilities82. This study integrates the DBO algorithm with wavelet mutation. In order to achieve the fine-tuning impact of the mutation operation, improve the ability to jump out of the local optimum, and increase the algorithm’s calculation accuracy, the number of iterations in the wavelet function’s scaling parameters is used to confine the mutation space. Mutation is only carried out during the exploratory phase of the algorithm since its goal is to break out of the local optimum.
Let \(p\) be the mutation probability on the interval [0,1], and \(s\) be a random number. Mutation probability indicates that when \(s\ge p\), dung beetles experience Morlet wavelet mutation, specifically:
when \(i=1\),
when \(i=2,\dots ,N\),
The point, \({X}_{rand}^{d}\), is randomly generated within the search space, \(U{b}^{d}\) and \(L{b}^{d}\). represents the upper and lower boundaries of the \({d}^{th}\) dimension, respectively. The wavelet mutation coefficient, \(\sigma\), is defined as \(\sigma =1/\sqrt{a}\psi (\varphi /a)\), \(\psi (x)\) represents the Morlet wavelet function, which may be expressed as \(\psi (x)={e}^{-{x}^{2}/2}\cdot cos(5x)\). The scaling parameter \(a\) is incremented as the number of iterations increases in order to prevent the algorithm from overlooking the global best solution as a result of excessive mutation in the later stages of iteration which is defined as follows:
where \(\lambda\) is a given constant selected randomly from the range \([800, 1200]\).
Incorporation of the MWM strategy in EBMLO-DBO can be expected to further improve the performance, as it tends to provide better convergence speed and avoidance of local optima. This strategy will change the mutation space dynamically by translation and scaling properties in the wavelet function and solve solutions in much more stable and precise ways. Wavelet mutation can provide the ability for the algorithm to fine-tune its search capabilities, especially during the exploratory phase, hence increasing the possibility of the algorithm to get out of the local optima. The progressive adjustment of the scaling parameter \(a\) helps maintain an effective balance between exploration and exploitation and ensures the algorithm will not miss the global best solution due to excessive mutation. This, therefore, assures such a procedure that it accelerates the convergence of the EBMLO-DBO algorithm, its robustness, and its accuracy toward the determination of an optimal solution.
Local escaping operator (LEO)
The Local Escaping Operator (LEO) is an additional local search approach83. Its main contribution is that it improves the search capability of the DBO algorithm, helping to locate unexplored areas of the search space-especially in those difficult real-world applications. This will result in an improvement of the general quality of a solution. LEO improves the convergence behavior of the optimization algorithm by updating solution positions based on specific criteria, hence preventing it from becoming trapped in local optima. LEO utilizes range of important tactics to offer alternative solutions that demonstrate superior performance. These tactics involve using the most favorable position, labeled as \({X}_{best}\), together with two randomly generated solutions known as \({X}_{r1}\) and \({X}_{r2}\). In addition, LEO randomly chooses two solutions, \(X{1}_{i}\) and \(X{2}_{i}\). Finally, LEO develops a novel solution, \({X}_{z}\), which is generated randomly. The suggestion outlines a mathematical calculation that can be used to calculate the value of \({X}_{LEO}\) as follows:

The given equations have several variables. The variable \(f1\) follows a uniform distribution with a range from -1 to 1, including both endpoints. The variable \(f2\) follows a normal distribution with a mean of 0 and a standard deviation of 1. In addition, there are three additional random variables: \(u1\), \(u2\), and \(u3\). Also, \(\rho 1\) denotes the likelihood.
where the variable \(rand\) indicates a randomly generated number that is within the range of 0 and 1. In contrast, the variable \(\mu\) represents a number that is also constrained inside the range of 0 to 1. The provided equations can be simplified as shown in Eqs. (26–28):
The binary parameter \(Q1\) can only have the values of 0 or 1. The value is determined by a condition: if \(Q1\) is less than 0.5, then \(Q1\) is assigned a value of 1. Alternatively, it is given a numerical value of zero. In addition, to ensure a fair allocation of resources between exploration and exploitation in search processes, the variable \(\rho 1\) is included. The variable is determined using Eqs. (29–31):
The values of \({\beta }_{min}\) and \({\beta }_{max}\) are assigned as 0.2 and 1.2, respectively. The variable \(t\) denotes the current iteration, whereas \(T\) is the maximum number of iterations. In order to maintain a proper equilibrium between exploration and exploitation, the parameter \(\rho 1\) dynamically adapts itself according to the sine function \(y\). The solution for \({X}_{z}\) in the given scheme can be computed by applying the method described in Eqs. (32) and (33):
From Eq. (32), the solution \({X}_{rand}\) denotes a novel solution, whereas \({X}_{p}\) represents a solution that has been chosen randomly from a population. Moreover, \(\mu\) represents a random variable that follows a uniform distribution with values ranging from 0 to 1. Equation (32) can be simplified in the following manner:
where the parameter \({Q}_{2}\) is a binary variable that can only take the values of 0 or 1. The value is determined based on the condition of \(\mu\) being less than 0.5 or not. The stochastic selection of parameter values, \({u}_{1}\), \({u}_{2}\), and \({u}_{3}\), enhances the heterogeneity of the population and mitigates the likelihood of suboptimal solutions.
EBMLO-DBO for local optima avoidance
Local optima represent a fundamental challenge in optimization where algorithms become trapped in suboptimal solutions, mistakenly treating them as global optima due to insufficient exploration or premature convergence84. In the EBMLO-DBO, local optima situations are systematically identified and addressed through multiple complementary strategies designed to maintain population diversity and exploration capability throughout the optimization process. Local optima conditions typically manifest when population diversity decreases below critical thresholds, fitness improvement stagnates over consecutive iterations, and search agents converge prematurely to limited regions of the search space, effectively terminating the search process before discovering superior solutions.
Thus, EBMLO-DBO employs four synergistic mechanisms to avoid local optima entrapment effectively. The Bernoulli map-based initialization generates chaotic sequences with inherent properties of non-linearity, sensitivity to initial conditions, and ergodicity, ensuring diverse starting positions that significantly reduce the probability of early convergence to local optima by providing better coverage of the search space. Morlet Wavelet mutation is strategically applied when stagnation is detected through the condition \(s\ge p\) in Eqs. (19) and (20), where the wavelet function dynamically adjusts mutation space using the scaling parameter '\(a\)' that decreases with iterations according to Eq. (21), providing strong exploration capabilities in early optimization stages and fine-tuning adjustments in later stages. The ELS strategy prevents premature convergence by maintaining both elite random leadership and best leadership mechanisms, where the selection probability \(\tau\) defined in Eq. (17) starts at approximately 1 and gradually decreases, ensuring continued exploration while progressively focusing on promising regions. The LEO strategy, activated with probability \(pr\) through Eqs. (22–34), generates alternative solutions using the best position \(Xbest\) and randomly selected solutions, effectively enabling the algorithm to “jump out” of local optima through strategic solution perturbations. The synergistic effect of these mechanisms ensures that even when individual agents become trapped in local optima, the overall population maintains sufficient diversity and exploration capability to escape and continue searching for global optima, with statistical analysis confirming that EBMLO-DBO successfully avoids local optima in most of test functions compared to the original DBO algorithm.
Architecture and working
To balance the trade-offs between exploration and exploitation in solving complex optimization problems with enhanced efficiency, the proposed EBMLO-DBO is equipped with four advanced strategies. The step-by-step procedure of the EBMLO-DBO algorithm is elaborated in detail in this section. The algorithm progresses through a series of iterative updates, applying different strategies to various subsets of the population, with the ultimate goal of finding the optimal solution.
-
Step 1 Initialization: The algorithm initializes the population according to Bernoulli maps presented in Eq. 15; this generates a highly diversified set of initial solutions since it gives a good starting point to efficiently explore the search space.
-
Step 2 Fitness evaluation: This step involves the calculation and sorting of fitness values in each iteration. Ranking provides important information about the best solutions and about the splitting of the population and its update.
-
Step 3 Population division: This is followed by the division of the population, based on the ranking, into two subgroups, \(Y1(t)\) and \(Y2(t)\), on which different strategies can be applied for an effective optimization of both exploration and exploitation.
-
Step 4 Rolling Dung Beetles Update: For each rolling dung beetle, a random value \(\alpha\) is generated. If \(\alpha \le 0.9\), then the position of the rolling dung beetle is updated by Eq. (2). Otherwise, if the beetle hits the obstacle, its position will be updated by Eq. (3).
-
Step 5 Calculation of Convergence Factor: The convergence factor \(R\) will be computed using the relation \(R=1-t/T\), where \(t\) denotes the iteration at each run and \(T\) signifies the maximum number of iterations. This factor is affecting the dynamics of the search during the runtime; thus, the balance between exploration and exploitation will be changing with time passing.
-
Step 6 Spawning Dung Beetles Update: The position of spawning dung beetles is updated by using Eq. (4) and Eq. (5). Both these updates ensure that the spawning dung beetles contribute towards maintaining the diversity in population and enhance the process of searching.
-
Step 7 Foraging Dung Beetles Update: If a random generated value is less than a predefined probability \(p\), update the foraging dung beetle using Eq. (6) and Eq. (7); otherwise, apply Morlet Wavelet Mutation using Eq. (19) or Eq. (20) to introduce diversity for the avoidance of premature convergence.
-
Step 8 Stealing Dung Beetles Update: The stealing dung beetles are updated using Eq. (8). This strategy focuses on exploiting the best areas of the search space, driven by the behavior of elite individuals.
-
Step 9 Elite Leadership and Greedy Selection: Each solution undergoes an update using the elite leadership strategy as defined in Eq. (16). Afterward, a greedy selection process, guided by Eq. (18), compares new and old solutions to retain only the best candidates for the next iteration.
-
Step 10 LEO Operator Application and Boundary Check: The Local Escape Operator (LEO) is applied to each solution using Eqs. (22–34), helping to escape local optima. Following this, all solutions are checked to ensure they remain within the predefined boundary limits.
-
Step 11 Update Optimal Solution: The current best solution \({X}_{best}\) is updated based on the latest evaluations. This ensures that the algorithm continuously tracks the most promising solution throughout the optimization process.
-
Step 12 Iteration and Completion: The iteration counter \(t\) is incremented, and the process repeats until the maximum number of iterations \(T\) is reached. Upon completion, the algorithm returns the optimal solution \({X}_{best}\) along with its corresponding fitness value.
The procedures and operators of the proposed EBMLO-DBO are depicted in Algorithm 1 and Fig. 4.

Algorithm 1: EBMLO-DBO algorithm
Computational complexity
The computational complexity of an algorithm is an essential factor that affect its performance. The computational of an algorithm can depend on two parameters: the time complexity and the space complexity85. The time complexity assures and evaluate the theoretical performance of an algorithm in terms of its effectual operations while the space complexity is the required memory and storage for the algorithm while execution. The next subsections discuss the time and space complexity of EBMLO-DBO in details and compare it with the original DBO to assure the difference between them.
Time complexity
The variables dimension \(D\), population size \(N\), and number of iterations \(T\) determine the algorithm’s temporal complexity. The original DBO is mainly depend on its internal operators beside the initialization phase. The main loop of the DBO algorithm include the four main phases of ball rolling, foraging, spawning and stealing. The main population is divided among these operators therefore, the main loop complexity includes the complexity of the four phases. The complexity of DBO can be described as follows:
The inclusion of LEO, BM, MWM and ELS strategies are the extended operators found in EBMLO-DBO but not in DBO. In other words, every solution undergoes the BM which is used for the initialization require \(O(N\times D)\) while the LEO is applied for every solution at each iteration, resulting in a complexity of \(O(T\times N\times D)\). Since the MWM strategy is a component of the EBMLO-DBO update methodology, its complexity when paired with the operators of the DBO algorithm is \(O(T\times N\times D)\). Since the ELS strategy is used at the conclusion of every iteration, its complexity is \(O(T\times N\times D)\). In summary, EBMLO-DBO’s temporal complexity is:
Therefore, when comparing DBO with the suggested EBMLO-DBO, the time complexity remains the same, but there are variations in performance.
Space complexity
The quantity of storage space required for an algorithm to execute is referred to as its space complexity. Thus, beginning from the initialization step, the space complexity of \(O(N\times D)\) is the same for both DBO and the proposed EBMLO-DBO.
Experimental results
This section highlights the experimentation results of the proposed EBMLO-DBO algorithm. First, we provide details about the experimental setup and parameter settings, which are crucial to replicate the study or take cognizance of the operating conditions in the algorithm evaluation. Then, the performance metrics used in the effectiveness evaluation of the algorithm are presented, completing the framework of comprehensive evaluation. Lastly, the section highlights, the test results from the experiments, presenting a comprehensive analysis in order to outline the strengths and weaknesses, along with the overall performance of the algorithm in question for the various test scenarios.
Experimentation setup and parameter settings
This work compared the performance of the EBMLO-DBO algorithm against eight state-of-the-art algorithms that experience superior performance, together with eleven conventional and recent algorithms. In this work, a total of 41 benchmark functions were tested, including 12 from CEC’2286 and 29 from CEC’1787. This is supported by five modern algorithms: Crayfish Optimization Algorithm (COA)20, Levy Flight Distribution optimizer (LFD)88, Nutcracker Optimizer (NOA)16, Gradient-based Optimizer (GBO)83, original DBO, and six traditional algorithms: Particle Swarm Optimizer (PSO)14, Harris Hawk Algorithm (HHO)89, Arithmetic Optimization Algorithm (AOA)90, Whale Optimization Algorithm (WOA)91, Slime Mould Optimizer (SMA)92, and Artificial Rabbit Algorithm (ARO)93. Table 1 summarizes the conditions of these comparative schemes. In Table 1, all algorithms are performed under the same condition with the initial population size 30 and the same maximum iteration limit 1000. For each algorithm, results are generated by 30 runs for every test function in order to avoid random effects94,95. The mean best-of-run (AVG) and standard deviation (STD) of solutions are then recorded. All the experiments are conducted on an Intel(R) i7-10750H CPU computer running Microsoft Windows 10 with 32 GB of RAM. Besides, MATLAB R2020a is used as a development environment for coding to ensure computational power and dependability when conducting the experiments. The rank rows of the Tables show the ordering of the mean values. A rank of 1 means that, after 30 runs, the algorithm had the lowest mean of the solution; it therefore means higher search capacity.
Performance indicators
To accurately differentiate between algorithms and ensure a fair and credible comparison, this section introduces three key performance evaluation metrics—Mean and Standard Deviation—along with two statistical testing methods. These metrics and tests are utilized to compare the performance of the involved algorithms. For the ensuing, let \(M\) be the number of independent runs of each algorithm, and for every run \({g}_{opt}\) will denote the best solution obtained. The specific definitions and calculation methods for these metrics are outlined below:
-
Average Value (AVG): The mean value is a statistical measure that reflects the central tendency of the data96. It provides an overview of the typical performance of an algorithm over multiple runs. The calculation of the mean value is given by the following equation:
$$\text{AVG }=\frac{1}{M}\times \left({g}_{opt}^{1}+{g}_{opt}^{2}+{g}_{opt}^{3}+\cdots +{g}_{opt}^{M}\right)$$(35) -
Standard Deviation (STD): Standard deviation describes the dispersion of the individual observations with respect to the mean value. The higher is the STD, the larger is the dispersion, which in turn signifies that the algorithm produces more unstable results. A lower STD therefore means that a particular algorithm will produce more consistent and hence more stable results. The STD is calculated as follows:
$$\text{STD }=\sqrt{\frac{{\left({g}_{opt}^{1}-AVG\right)}^{2}+{\left({g}_{opt}^{2}-AVG\right)}^{2}+\cdots +{\left({g}_{opt}^{M}-AVG\right)}^{2}}{M-1}}$$(36) -
Wilcoxon Rank Sum Test: The Wilcoxon rank sum test is a non-parametric statistical test used to determine the significant median of differences between two independent samples. The test ranks all data from the two groups together to obtain the rank sums for each group and then compares these rank sums in order to show that the difference between the two groups is statistically significant. This approach is useful when considering two algorithms’ performances under different conditions.
-
Friedman Test: The Friedman test is another non-parametric statistical test, but it compares three or more matched samples. In the case that there is more than one algorithm, the Friedman test will rank their performances from best to worst. Accordingly, each of these algorithms will be given a rank according to this test in terms of performance, and if there is no significant difference among the algorithms, their ranks should be evenly distributed. This test is supposed to indicate the performance of several algorithms relative to one another for various datasets or problem instances.
These performance metrics and statistical tests collectively provide a comprehensive framework for evaluating and comparing the effectiveness, stability, and reliability of different algorithms in a robust and statistically sound manner.
Exploration- exploitation balance test
The exploration–exploitation percentage plots presented in Fig. 5 demonstrates the EBMLO-DBO algorithm’s superior ability to maintain an optimal dynamic balance between exploration and exploitation phases throughout the optimization process across diverse function landscapes from the CEC’17 benchmark suite. The blue line representing exploration percentages and the orange line representing exploitation percentages reveal that EBMLO-DBO consistently exhibits a well-coordinated transition from high exploration dominance in early iterations to increased exploitation in later stages, which is particularly evident in unimodal functions C17-G1 and C17-G3 where the algorithm maintains approximately 80–90% exploration during the initial 50 iterations before systematically transitioning to exploitation dominance with orange line values reaching 60–70% by the final iterations. The balanced approach is clearly visible in multimodal functions C17-G6 and C17-G9, where EBMLO-DBO demonstrates intelligent phase transitions with blue exploration percentages gradually decreasing from initial high values while orange exploitation percentages correspondingly increase, ensuring effective navigation through complex search landscapes with multiple local optima. The algorithm shows remarkable adaptability by maintaining sufficient exploration capability when encountering challenging regions while smoothly transitioning to exploitation mode once promising areas are identified.
The exploration–exploitation balance validation is further reinforced through EBMLO-DBO’s performance on hybrid functions C17-G12, C17-G15, C17-G19, and C17-G20, where the algorithm demonstrates exceptional balance control with blue exploration lines starting at high percentages and gradually decreasing while orange exploitation lines correspondingly increase, maintaining moderate exploration levels during mid-stage iterations to ensure comprehensive search coverage without premature convergence to suboptimal solutions. The composite functions C17-G22, C17-G25, C17-G28, and C17-G30 present the most challenging scenarios for balance maintenance, yet EBMLO-DBO consistently achieves stable exploration–exploitation transitions with smooth percentage curves, where the blue exploration line typically exhibits initial dominance of 80–90% that systematically decreases as the orange exploitation line increases to achieve dominance of 60–80% in later iterations. This dynamic balance mechanism is achieved through the coordinated interaction of the four enhancement strategies integrated within EBMLO-DBO, where the Bernoulli map initialization promotes diverse blue exploration percentages in early stages, the elite leadership strategy provides intelligent guidance for balanced transitions between blue and orange phases based on population quality, the MWM maintains necessary exploration capabilities during critical phases to escape local optima, and the LEO enhances orange exploitation effectiveness during convergence phases. The consistent and adaptive exploration–exploitation balance exhibited by EBMLO-DBO across all twelve function categories, as demonstrated by the smooth percentage transitions and appropriate phase dominance patterns in Fig. 5, directly contributes to its enhanced optimization performance and robust solution quality.
Ablation study
This section presents a comprehensive ablation study designed to evaluate the individual contribution of each proposed enhancement strategy to the overall performance of the EBMLO-DBO algorithm. To systematically assess the impact of each component, six algorithmic variants were developed and tested: BDBO incorporates only the Bernoulli map-based initialization strategy, EDBO implements solely the ELS strategy, MDBO utilizes only the MWM approach, LDBO employs exclusively the LEO strategy, DBO represents the original baseline algorithm, and EBMLO-DBO integrates all proposed strategies collectively. The experimental evaluation was conducted using the CEC’22 benchmark suite comprising twelve distinct optimization functions.
The results presented in Table 2 demonstrate the substantial individual impact of each enhancement strategy when compared to the original DBO algorithm, with the EBMLO-DBO achieving the most significant improvements across all function categories. The LDBO variant, incorporating only the LEO strategy, showed remarkable performance gains over the original DBO, particularly in unimodal function C22-G1 where it achieved 2.04E + 04 compared to DBO’s 2.28E + 07, and in hybrid function C22-G6 with 39,527.408 versus DBO’s catastrophic 1.75E + 10, demonstrating LEO’s effectiveness in escaping local optima and enhancing exploitation capabilities. The MDBO variant, utilizing solely the MWM strategy, exhibited consistent improvements across most functions, achieving near-optimal performance in C22-G1 with 333.975 compared to DBO’s 2.28E + 07, and maintaining competitive results in multimodal and composite functions, indicating the wavelet mutation’s capacity to enhance exploration while preventing premature convergence. The EDBO variant with ELS strategy showed moderate but consistent improvements over the original DBO, particularly effective in hybrid functions C22-G5 and C22-G6 where it achieved 2,161.254 and 1.27E + 04 respectively compared to DBO’s significantly higher values of 2.64E + 04 and 1.75E + 10, while the BDBO variant with Bernoulli map initialization demonstrated mixed results, showing substantial improvements in some functions like C22-G1 but performing worse than DBO in certain composite functions, suggesting that chaotic initialization alone requires complementary strategies to achieve consistent optimization performance. The synergistic integration of all strategies in EBMLO-DBO consistently outperformed both the original DBO and all individual variants, achieving the best average rank of 1.83 and final Friedman rank of 2.16, while the original DBO ranked consistently last with an average rank of 5.75, confirming that each enhancement strategy addresses specific algorithmic limitations and their collective implementation provides the most robust optimization framework.
CEC’22 results analysis
The CEC’22 test set comprises 12 distinct optimization test functions. In the experimental configuration, a dimension of 20 is employed. The comparisons have encompassed a range of algorithms, including DBO, WOA, HHO, LFD, PSO, GBO, NOA, AOA, COA, SMA, and ARO. CEC’22 has a greater degree of intricacy and challenge. The obtained results of applying EBMLO-DBO compared to its rivals using CEC’22 is shown in Table 3. EBMLO-DBO exhibits the best rank in the functions C22-G1, C22-G5, C22-G7, C22-G8, C22-G11 and G12, as evidenced by the statistical values presented in Table 3. Furthermore, the mean ranking of the solution outcomes for EBMLO-DBO and its corresponding algorithms, acquired from 30 separate runtimes of each function, is depicted in Fig. 6. Based on the data provided, it can be inferred that EBMLO-DBO consistently identifies superior solutions for these functions and maintains a high level of solution stability. EBMLO-DBO outperforms all other algorithms by achieving the highest rank in 50% of the functions and dominating them. On the other hand, EBMLO-DBO attains the second-highest position in 41.6% of the functions and the third-highest position in the remaining functions, surpassing all competitors. It is also analyzed that, the functions whose EBMLO-DBO is ranked the second, the gap between the second rank and the first is small. For example, for C22-G2, the PSO obtained the best rank in terms of the average fitness value but the stability of C22-G2 is the best indicating the superior performance of DBMLO-DBO at functions with the second rank. For the function C22-G3, the GBO was the best among all competitors this is due to the LEO operator which enhances the search ability of GBO at this function which indicates the powerful ability of the LEO operator as a new improvement proposed to EBMLO-DBO. Regarding the average rank, it is clear that the exceptional performance of EBMLO-DBO outperform all its rival while GBO is the second in terms of the average rank. The aforementioned investigations imply that EBMLO-DBO exhibits exceptional global search and optimization prowess in nearly all CEC’22 functions. It showcases the consistent capacity to consistently discover exceptional solutions and generate solutions that are of superior quality and stability.
Comparison of EBMLO-DBO with advanced algorithms
To further evaluate the performance of EBMLO-DBO against high-performing algorithms, an experiment was conducted using eight advanced optimization techniques. These include five advanced algorithms: DAOA97, SCADE98, RLTLBO99, , and LWOA100 along with three top-performing algorithms in the CEC competitions: CMAES101, LSHADE-cnEpSin102, IMODE103, and AGSK104. The experiment involved 30 independent runs on twenty-nine CEC’17 benchmark functions with a dimensionality of 50, and the results are summarized in Table 4, highlighting the mean and standard deviation of the fitness values.
The investigation reveals that the EBMLO-DBO algorithm achieves the lowest average fitness value in 17 out of 29 functions, outperforming all other algorithms in the study. EBMLO-DBO also has the highest number of superior functions, demonstrating exceptional performance, particularly in unimodal functions CEC17-G1 and CEC17-G3, where it shows superior standard deviation and mean fitness values. For function CEC17-G4, while IMODE achieves the best standard deviation values, EBMLO-DBO secures the best average fitness values.
The original DBO method, in contrast, does not excel in any of the benchmark functions. The remarkable performance of EBMLO-DBO can be attributed to several proposed improvements. These include the MWM, which enhances exploration by introducing diverse solutions; the LEO, which helps the algorithm avoid local optima; the Bernoulli operator for initializing the population, which ensures diversity from the start; and the ELS strategy, which boosts convergence speed by leveraging the best individuals to guide the search.
These enhancements significantly improve the balance between exploration and exploitation, ensuring high-quality solutions and optimal global optima. As evidenced by the results in Table 4, EBMLO-DBO is a highly promising algorithm in the field of optimization, outperforming several well-established algorithms. Its superior ability to solve global optimization tasks makes it an invaluable tool for complex optimization problems.
Convergence behavior analysis
This section examines the convergence behavior and scalability performance of the EBMLO-DBO algorithm compared to eleven state-of-the-art optimization algorithms across varying dimensional complexities, utilizing selected benchmark functions from different categories of the CEC’17 test suite to evaluate algorithmic robustness and effectiveness as problem dimensionality increases from 10 to 100 dimensions105. These results are shown in Figs. 7, 8, 9, 10
The convergence behavior analysis across varying dimensional complexities from 10 to 100 dimensions reveals the robust scalability and superior performance characteristics of the EBMLO-DBO algorithm compared to eleven state-of-the-art optimization algorithms including AOA, ARO, CMAES, COA, DBO, GBO, HHO, IMODE, LFD, LSHADE_nEpSin, PSO, SMA, and WOA. The comprehensive evaluation conducted using selected benchmark functions from different categories of the CEC’17 test suite demonstrates that EBMLO-DBO consistently maintains its competitive advantage across all dimensional settings, with particularly notable performance improvements becoming more pronounced as problem dimensionality increases. In the 10-dimensional setting, EBMLO-DBO exhibits superior convergence characteristics across all function categories, achieving the lowest fitness values in unimodal functions F1 and F3, where it demonstrates rapid initial descent followed by stable convergence, while competitors like DBO, DAOA, and COA show premature stagnation or oscillatory behavior. The multimodal functions F6 and F9 reveal EBMLO-DBO’s enhanced exploration capabilities, as it successfully navigates complex landscapes with multiple local optima, consistently outperforming algorithms such as PSO, GBO, and WOA that frequently become trapped in suboptimal regions.
As dimensionality increases to 30 dimensions, the performance gap between EBMLO-DBO and competing algorithms becomes more pronounced, particularly evident in the unimodal function F1 where EBMLO-DBO achieves remarkable fitness improvements from approximately 26 to below 8, while most competitors struggle to achieve comparable convergence rates. The hybrid functions F12 and F19 in 30 dimensions showcase EBMLO-DBO’s adaptive nature, maintaining smooth convergence trajectories while algorithms like ARO, SMA, and CMAES exhibit irregular convergence patterns with frequent plateaus and limited improvement in later iterations. The composite functions F25 and F29 present the most challenging optimization scenarios in 30 dimensions, yet EBMLO-DBO demonstrates exceptional stability, achieving consistent convergence to near-optimal solutions while maintaining computational efficiency throughout the optimization process. The 50-dimensional results further emphasize EBMLO-DBO’s scalability advantages, where the algorithm continues to exhibit superior performance across all function categories, with particularly impressive results in multimodal function F9 where it achieves convergence to approximately 9.2 compared to competitor algorithms that remain above 10.5, and in hybrid function F12 where EBMLO-DBO reaches fitness values below 17 while most alternatives plateau above 20.
The 100-dimensional analysis represents the most challenging optimization environment, where the curse of dimensionality typically causes significant performance degradation in traditional optimization algorithms, yet EBMLO-DBO maintains its superior convergence characteristics with remarkable consistency. In the 100-dimensional unimodal function F1, EBMLO-DBO demonstrates exceptional scalability by achieving continuous improvement from initial values to final fitness values, while competing algorithms like DBO, DAOA, and ARO show minimal improvement or complete stagnation after early iterations. The multimodal functions F6 and F9 in 100 dimensions reveal EBMLO-DBO’s robust exploration–exploitation balance, where it consistently achieves the lowest fitness values across all evaluation points, with F9 showing particularly dramatic improvements, significantly outperforming all competitors that struggle to achieve fitness values below 10. The hybrid and composite functions F12, F19, F25, and F29 in 100 dimensions demonstrate EBMLO-DBO’s ability to handle the most complex optimization landscapes, maintaining smooth convergence trajectories and achieving superior final solutions, while most competing algorithms exhibit premature convergence, irregular oscillations, or complete stagnation. The consistent performance superiority across all dimensional settings can be attributed to the synergistic integration of the four enhancement strategies within EBMLO-DBO, where the Bernoulli map initialization ensures diverse starting points that become increasingly crucial as dimensionality increases, the ELS strategy provides intelligent guidance that scales effectively with problem complexity, the MWM maintains exploration capabilities essential for high-dimensional escape from local optima, and the LEO enhances exploitation effectiveness that becomes increasingly important for fine-tuning solutions in complex high-dimensional spaces.
Statistical analysis of EBMLO-DBO
To establish the statistical significance of the EBMLO-DBO algorithm’s performance superiority, the Wilcoxon signed-rank test was implemented as a non-parametric statistical analysis method to determine whether the proposed EBMLO-DBO demonstrates statistically meaningful differences compared to competing optimization algorithms106. This statistical validation was conducted through 30 independent experimental runs for each comparative algorithm across the comprehensive benchmark function suite, ensuring robust statistical reliability and eliminating potential bias from random variations. The Wilcoxon signed-rank test was executed with a confidence level of 95% (α = 0.05) to evaluate the statistical significance of performance differences between the solution quality achieved by EBMLO-DBO and that obtained by each competitor algorithm. The comprehensive statistical analysis results are systematically organized and documented in Tables 5 and A1 for detailed examination. Following standard statistical hypothesis testing protocols, when the computed p-value falls below the threshold of 0.05, the null hypothesis of no significant difference is rejected, thereby confirming the existence of statistically significant performance distinctions between the compared algorithms. Conversely, when the p-value exceeds 0.05, the performance differences between the two algorithmic approaches are considered statistically insignificant. The results presented in Table 5 provide conclusive evidence that the EBMLO-DBO algorithm exhibits statistically significant performance improvements when compared to all benchmark algorithms including WOA, NOA, COA, HHO, PSO, DBO, GBO, AOA, SMA, ARO, and LFD, thereby establishing the statistical validity of EBMLO-DBO’s superior optimization capabilities. Furthermore, the table demonstrates the statistical differentiation between EBMLO-DBO and other advanced optimization methodologies, confirming that EBMLO-DBO maintains statistically significant advantages across most benchmark functions, thus validating the robustness and reliability of the proposed algorithmic enhancements.
In addition, the Friedman test was used to rank each optimization algorithm, and the test findings are provided in Fig. 11. Critically looking at the data of Fig. 11 reveals that the improved dung beetle algorithm EBMLO-DBO has a lower average rating than the other algorithms. This means that the effectiveness of the augmentation can be established. In other words, the figure shows the Friedman rank test results for CEC’17-D50, compared with advanced algorithms and CEC’22-D20 compared to a set of classical and recent algorithms which provides an overall summary of the ranks for different occurrences and dimensions of the problem. It is clear that for both comparisons the EBMLO-DBO is the best by obtaining the lowest Friedman rank and is ranked the first. This indicates that EBMLO-DBO is a powerful promising algorithm that can handle different types of functions with different dimensions this is due to the proposed improvements that enhanced the solution quality, accuracy, and convergence speed of DBO toward the best global solutions.
PV parameter estimation using EBMLO-DBO
The process of determining parameters for PV models can be modeled as an optimization problem. As previously mentioned, minimizing the disparity between simulated data and experimental data is the main objective of this optimization problem. Hence, the objective function in this case can be formulated as computing the disparity between the manufacture measured current and the estimated current. Subsequently, optimization approaches are employed to investigate and ascertain the parameter values minimizing this disparity. The absolute difference between the expected and observed current is defined as the individual absolute current error (AEC) which can be formulated in Eq. (37) and (38) for both the single diode and double diode, respectively:
Moreover, the root mean square error (RMSE) is computed by squaring the average of the observed and predicted current values, and then taking the square root of this value. This term represents the cumulative discrepancy between the two sets of data:
where \(N\) denotes the entire quantity of data points that have been measured. The vector \(x\) comprises the parameter values ascertained by the optimization approach. The variables included in the single-diode model, denoted by \(x\), are \({I}_{rsD}\), \({I}_{PC}\), \(N\), \({R}_{se}\) and \({R}_{SH}\). On the other hand, the double-diode model contains the variables \({I}_{PC}\), \({I}_{rsD1}\), \({I}_{rsD2}\), \({R}_{SH}\), \({R}_{se}\), \({N}_{1}\), and \({N}_{2}\).
Experimental settings and comparative algorithms
An ideal scenario to exhibit the effectiveness of the proposed method is to illustrate its ability to precisely represent the properties of PV module models, including the double-diode, PV module, single-diode models. The data from107 were selected as benchmarks because to their widespread usage in evaluating various methods for predicting parameters of photovoltaic models. The investigations involved the serial connection of 36 polycrystalline PV cells with a commercial silicon PV cell from RTC France, which had a diameter of 57 mm. The experiments were carried out at an irradiation level of 1000 W/m2 and a 33 °C temperature. Furthermore, a monocrystalline STM6-40/36 module underwent testing under specific conditions, including an irradiation level of 1000 W/m2 and a temperature of 45 °C. Previous research has utilized parameter ranges that ensure a reliable search space for all problems. The parameter ranges for PV cells and modules are illustrated in Table 6.
A wide range of approaches has been engaged for estimating the parameters of the PV models, including but not limited to the analytical heuristic and deterministic approaches108. Analytical procedures are widely adopted due to their practicality and ease of use109. In addition, the results cannot always be reliable or precise enough since the basic assumptions used may be too general110. In deterministic approaches, the estimation is bonded by the assumptions made within the model and may be caught up in local maxima. To be certain, the convexity and differentiability of the model can guide the efficiency of the models especially for multi-modal nonlinear PV models. On the contrary, the heuristic approaches often outperform the analytical and deterministic methods in terms of performance. These are flexible methods that do not impose constraints on the problem at hand and are therefore suitable for a wide range of practical problems. Heuristics approaches have indeed appeared as promising alternatives when it comes to the estimation of parameters in PV models.
The proposed EBMLO-DBO is tested using the recently developed classical and advanced algorithms to overcome the complexity of parameters extraction for different solar systems. The approaches encompassed are RTLBO111, SSA112, IJAYA113, BES114, CGO-LS115, MLBSA 116, NOA64, GBO83, and DBO. Since swarm intelligence algorithms are non-deterministic, all runs are made with the same setting. In fair comparisons, the maximum iteration number of each approach is set to 500 for all experiments. For each problem, the process repeats 30 independent times in order to ensure statistically valid comparisons. The accuracy and robustness of the solution obtained by each scheme are measured with the statistical measure. After that, a deeper comparison is drawn based on the best RMSE values obtained using 30 runs in each method. The smallest magnitude of the RMSE values is highlighted in bold.
Results analysis
In this section, the proposed EBMLO-DBO is compared with its competitors for all three PV models: double-diode model, single-diode model, and PV module model. I-V and P–V curves are shown for this experiment. In addition, the estimated parameters and the statistical metrics for each model are presented in a comprehensive manner.
Regarding to the single diode model: The EBMLO-DBO algorithm accurately predicted the current–voltage (I-V) and power-voltage (P–V) curves for the Single Diode Model (SDM), as shown in Fig. 12. The results obtained via EBMLO-DBO exhibit a high degree of correlation with the experimental data across the whole range of voltages. Figure 13 illustrates the differences, in absolute error, between the values obtained from experiments and those obtained from simulations. Table 7 displays the recorded current measurements acquired during a simulated test session. The computed values for the Individual Absolute Error (IAE) are included in the analysis. These values are determined using the following Equation:
The results indicate that the IAE values are below 0.0025667. This demonstrates that the single-diode model, when simulated using the suggested algorithm, closely corresponds to the real features of solar cells. In addition, the EBMLO-DBO algorithm has superior performance compared to other approaches, including RTLBO, SSA, IJAYA, BES, CGO-LS, MLBSA, GBO, and DBO. This is clearly seen in Table 8, which provides detailed information on the five parameter values and the corresponding RMSE. In addition, Table 9 clearly shows that both the proposed EBMLO-DBO and algorithm RTLBO consistently produced the lowest RMSE value of 9.8602E-4 in both circumstances. These findings indicate that the suggested technique is very efficient in determining the parameters of the SDM.
-
Regarding the result analysis of double diode model: Fig. 14 presents a juxtaposition of the estimated characteristics generated by EBMLO-DBO and the factual measured data. In addition, Fig. 15 visually illustrates the IAE for the double-diode module. The simulations demonstrate a close resemblance between the simulated I-V and P–V characteristic curves and the experimental data. Table 9 presents the power, current and IAE parameters of the simulation. At every voltage measurement point, the IAE is lower than 0.002597, and the. This indicates that the double-diode module designed by EBMLO-DBO effectively represents the key properties of solar cells. Furthermore, Table 10 displays a juxtaposition of the outcomes achieved by EBMLO-DBO in relation to those of comparative algorithms. The table provided contains a comparison of the RMSE as well as the values of the seven retrieved parameters. Out of the seven ways that were assessed, the proposed EBMLO-DBO algorithm obtains the lowest RMSE value.
-
Regarding the result analysis of PV module mode: Fig. 16 presents the I-V and P–V curves for the Photowatt-PWP 201 module model, demonstrating a strong correlation between the measured and simulated data, indicating a high level of consistency. Figure 17 displays the measured and simulated currents for the IAE of the PV module model. In addition, Table 11 provides comprehensive information on the power and current values obtained from this experiment, as well as the accompanying IAE. The IAE value is explicitly stated to be lower than 0.00585036. These results indicate that the PV module model parameters advised by EBMLO-DBO are dependable. Estimating all five parameters that characterize the PV module model is essential. Table 12 presents the extracted values for these five parameters, along with the optimal RMSE value achieved by seven algorithms across 30 tests. A summary of this data is provided in Table 13, showing that EBMLO-DBO achieves the highest performance in terms of RMSE. Therefore, it can be concluded that the proposed EBMLO-DBO method excels in parameter estimation for the PV module model, outperforming other approaches.
The evaluation of the EBMLO-DBO algorithm’s performance employs non-parametric statistical tests and multiple metrics. Accuracy is determined by analyzing the mean and best values, while a lower worst value shows the algorithm’s capability to prevent premature convergence. Consistency is indicated by the standard deviation. The Wilcoxon signed-rank test is used to compare the proposed algorithm with baseline methods, and the Friedman test ranks the algorithms based on overall performance. A lower Friedman test rank value indicates superior performance.
In the Wilcoxon signed-rank test, different symbols represent levels of significance: a \(+\) symbol means the algorithm performs worse than the proposed one at \(p < 0.05\); a \(-\) symbol indicates significantly better performance at \(p < 0.05\); and a \(=\) symbol denotes similar performance to the proposed algorithm. The results are detailed in Table 13, with the Friedman test rank values listed in the final row. Table 13 shows the statistical results for various algorithms after 30 independent trials, each consisting of 500 iterations. The analysis includes worst, best, standard deviation, average, and rank values.
Based on the data in Table 13, the Wilcoxon signed-rank test results, at a significance level of 0.05, indicate that EBMLO-DBO outperforms the other algorithms. This is shown by all \(+\) signs for double and single diode models, and all \(+\) signs except for EBMLO-DBO in the PV module scenario, where \(=\) indicates comparable performance. Furthermore, the Friedman test supports the dominance of EBMLO-DBO, ranking it as the best algorithm for average performance of double, single, and PV modules, respectively, followed by the RTLBO and other algorithms.
Finally, the obtained results for this work ascertain that the EBMLO-DBO approach is very effective at finding the optimal photovoltaic model parameters. It presents the minimum RMSE values and outperforms complex algorithms such as RTLBO, IJaya, and DBO in terms of accuracy and speed of results. With respect to \(\alpha = 0.05\), the performed Wilcoxon signed-rank test and the Friedman test proved that EBMLO-DBO enjoys a significant statistical advantage over other compared algorithms. The integration of the strategies of BM, LEO, MWM, and ELS enhances the ability of DBO to tackle complex optimization problems with several modes and local valleys. It makes this optimization algorithm versatile for a wide range of applications. This is achieved by avoiding local optima.
Discussion
This section provides a comprehensive analysis of the experimental results obtained from EBMLO-DBO and critically evaluates its performance against existing optimization methods. The proposed EBMLO-DBO demonstrates several significant advantages over conventional optimization algorithms, particularly in terms of convergence performance and solution quality. The algorithm achieved superior Friedman ranks of 1.83 and 2.7 on CEC’22 and CEC’17 benchmark suites respectively, indicating exceptional optimization capability across diverse problem landscapes. The Bernoulli map-based initialization strategy effectively addresses premature convergence issues commonly encountered in traditional MAs by providing more diverse initial solutions, while the integration of MWM and LEO strategy enables effective escape from local optima, as evidenced by superior performance on multimodal functions.
When compared to classical algorithms such as PSO, GBO, HHO, DBO, and WOA, EBMLO-DBO shows remarkable improvements in both solution quality and consistency. Against advanced algorithms including CMAES, IMODE, LSHADE-cnEpSin and AGSK, EBMLO-DBO achieved the lowest average fitness values in 18 out of 29 CEC2017 functions, demonstrating its competitiveness with state-of-the-art methods. In photovoltaic parameter estimation applications, EBMLO-DBO consistently achieved the lowest RMSE values across all three models: single diode (9.8602E-4), double diode (9.81307E-4), and PV module (2.32066E-3), outperforming specialized algorithms like RTLBO and SSA. The ELS strategy dynamically adjusts selection probability between elite random leadership and best leadership, maintaining optimal balance throughout the optimization process and contributing significantly to the algorithm’s superior performance.
Despite these advantages, EBMLO-DBO exhibits certain limitations that warrant consideration. The integration of four enhancement strategies increases computational overhead compared to the original DBO, resulting in higher execution times. The algorithm introduces additional parameters including scaling factors, mutation probabilities, and elite group size, which require careful tuning for optimal performance across different problem domains. Memory requirements are increased due to the ELS strategy’s need to maintain separate elite and non-elite populations, which may impact scalability for large-scale problems. While effective on tested dimensions up to \(D=50\), the algorithm’s performance on extremely high-dimensional problems requires further investigation. Each enhancement strategy contributes uniquely to overall performance: Bernoulli Map initialization provides an improvement in population diversity, ELS strategy effectively guides convergence but requires appropriate elite group size selection, MWM prevents premature convergence in most tested functions, and LEO significantly improves local optima escape with an increased computational overhead.
Conclusion and future scope
This study proposed an enhanced version of the DBO algorithm, named Elite Bernoulli-based EBMLO-DBO algorithm. The main contributions of this work include the design of four strategic mechanisms—Bernoulli map-based initialization for improved population diversity, MWM for enhanced search precision, an ELS strategy to accelerate convergence, and a LEO strategy to escape local stagnation. These innovations collectively address the core weaknesses of the original DBO, particularly its slow convergence and vulnerability to premature convergence in complex landscapes.
EBMLO-DBO was rigorously validated across 41 benchmark functions from the CEC’22 and CEC’17 test suites, where it demonstrated superior performance in terms of convergence speed, solution accuracy, and robustness. The algorithm achieved Friedman rankings of 1.83 and 2.7 on the CEC’22 and CEC’17 benchmarks, respectively, outperforming many recent and classical algorithms, including PSO, HHO, COA, CMAES, and IMODE. It achieved the first rank in 50% of the CEC’22 functions and produced the best fitness values in 18 out of 29 functions on CEC’17. The statistical significance of these results was confirmed by Wilcoxon signed-rank tests, yielding p-values below 0.05 across all comparisons. In real-world scenarios, EBMLO-DBO achieved unprecedented accuracy in photovoltaic parameter estimation, producing RMSE values of 9.8602 × 10⁻4, 9.81307 × 10⁻4, and 2.32066 × 10⁻3 for the single-diode, double-diode, and PV module models, respectively.
Although EBMLO-DBO has proven to be highly effective, certain limitations remain. The current implementation depends on fixed hyperparameters, which may affect performance across different problem domains. The inclusion of additional operators slightly increases computational cost, and its behavior in extremely high-dimensional and dynamic environments requires further investigation.
Future research directions are categorized into immediate, medium-term, and long-term priorities to expand the capabilities and applicability of EBMLO-DBO. Immediate efforts will focus on improving scalability for problems exceeding 1000 dimensions through the development of adaptive control strategies and integrating learning-based mechanisms for dynamic parameter adjustment. The algorithm will also be hybridized with machine learning techniques to facilitate automated tuning and problem-specific adaptation. Medium-term objectives include extending the algorithm to multi-objective optimization scenarios, enabling real-time adaptation to dynamic problems with time-varying objectives, and exploring quantum-enhanced variants that utilize quantum parallelism for faster convergence. Additionally, enhancing the exploratory capacity through opposition-based learning and comprehensive learning strategies will be considered. Long-term research will target applications in engineering design, structural optimization, renewable energy, deep learning hyperparameter tuning, neural architecture search, and large-scale industrial systems such as smart grids and supply chain management. These research directions collectively aim to establish EBMLO-DBO as a powerful and adaptable optimization framework suitable for diverse and increasingly complex real-world challenges.
Data availability
All data generated or analyzed during this study are included directly in the text of this submitted manuscript. There are no additional external files with datasets.
References
Abdel-Salam, M. et al. Harnessing dynamic turbulent dynamics in parrot optimization algorithm for complex high-dimensional engineering problems. Comput. Methods Appl. Mech. Eng. 440, 117908 (2015).
Villoth, J. P. et al. Two-tier deep and machine learning approach optimized by adaptive multi-population firefly algorithm for software defects prediction. Neurocomputing 630, 129695 (2025).
Karimzadeh Parizi, M., Keynia, F. & Khatibi Bardsiri, A. Woodpecker mating algorithm for optimal economic load dispatch in a power system with conventional generators. Int. J. Ind. Electron. Control Optim. 4(2), 221–234 (2021).
Elhoseny, M., Abdel-salam, M. & El-hasnony, I. M. Web service prediction and composition based on Harris Hawks optimization and deep learning. SN Comput. Sci. 6(5), 429 (2025).
Wu, L., Huang, X., Cui, J., Liu, C. & Xiao, W. Modified adaptive ant colony optimization algorithm and its application for solving path planning of mobile robot. Expert Syst. Appl. 215, 119410 (2023).
Abdel-salam, M., Kumar, N. & Mahajan, S. A proposed framework for crop yield prediction using hybrid feature selection approach and optimized machine learning. Neural Comput. Appl. 36(33), 20723–20750 (2024).
Jin, Z., Sun, X., Lei, G., Guo, Y. & Zhu, J. Sliding mode direct torque control of SPMSMs based on a hybrid wolf optimization algorithm. IEEE Trans. Ind. Electron. 69(5), 4534–4544 (2021).
Jia, Y. H., Mei, Y. & Zhang, M. A bilevel ant colony optimization algorithm for capacitated electric vehicle routing problem. IEEE Trans. Cybernet. 52(10), 10855–10868 (2021).
Abdel-salam, M. Quadruple strategy-driven hiking optimization algorithm for low and high-dimensional feature selection and real-world skin cancer classification. Knowl.-Based Syst., 315, 113286 (2025).
Zhang, J., Li, H. & Parizi, M. K. HWMWOA: A Hybrid WMA–WOA algorithm with adaptive cauchy mutation for global optimization and data classification. Int. J. Inf. Technol. Decis. Mak. 22(04), 1195–1252 (2023).
Sivanandam, S. N., Deepa, S. N., Sivanandam, S. N. & Deepa, S. N. Genetic algorithms 15–37 (Springer, 2008).
Han, K. H. & Kim, J. H. Quantum-inspired evolutionary algorithm for a class of combinatorial optimization. IEEE Trans. Evol. Comput. 6(6), 580–593 (2002).
Yao, X., Liu, Y. & Lin, G. Evolutionary programming made faster. IEEE Trans. Evol. Comput. 3(2), 82–102 (1999).
Kennedy, J., & Eberhart, R. (1942). Particle swarm optimization Proceedings of the International Conference on Neural Networks. Australia IEEE, 1948.
Fu, Y., Liu, D., Chen, J. & He, L. Secretary bird optimization algorithm: a new metaheuristic for solving global optimization problems. Artif. Intell. Rev. 57(5), 123 (2024).
Abdel-Basset, M., Mohamed, R., Jameel, M. & Abouhawwash, M. Nutcracker optimizer: A novel nature-inspired metaheuristic algorithm for global optimization and engineering design problems. Knowl.-Based Syst. 262, 110248 (2023).
Hu, G., Cheng, M., Houssein, E. H., Hussien, A. G. & Abualigah, L. SDO: A novel sled dog-inspired optimizer for solving engineering problems. Adv. Eng. Inform. 62, 102783 (2024).
Dehghani, M., Montazeri, Z., Trojovská, E. & Trojovský, P. Coati Optimization Algorithm: A new bio-inspired metaheuristic algorithm for solving optimization problems. Knowl.-Based Syst. 259, 110011 (2023).
Hashim, F. A. & Hussien, A. G. Snake optimizer: A novel meta-heuristic optimization algorithm. Knowl.-Based Syst. 242, 108320 (2022).
Jia, H., Rao, H., Wen, C. & Mirjalili, S. Crayfish optimization algorithm. Artif. Intell. Rev. 56(2), 1919–1979 (2023).
Lian, J. et al. Parrot optimizer: Algorithm and applications to medical problems. Comput. Biol. Med. 172, 108064 (2024).
Al-Betar, M. A., Awadallah, M. A., Braik, M. S., Makhadmeh, S. & Doush, I. A. Elk herd optimizer: A novel nature-inspired metaheuristic algorithm. Artif. Intell. Rev. 57(3), 48 (2024).
Xue, J. & Shen, B. Dung beetle optimizer: A new meta-heuristic algorithm for global optimization. J. Supercomput. 79(7), 7305–7336 (2023).
Zhao, S., Zhang, T., Ma, S. & Wang, M. Sea-horse optimizer: a novel nature-inspired meta-heuristic for global optimization problems. Appl. Intell. 53(10), 11833–11860 (2023).
Bai, J. et al. A sinh cosh optimizer. Knowl.-Based Syst. 282, 111081 (2023).
Ahmadianfar, I., Heidari, A. A., Gandomi, A. H., Chu, X. & Chen, H. RUN beyond the metaphor: An efficient optimization algorithm based on Runge Kutta method. Expert Syst. Appl. 181, 115079 (2021).
Zhao, S., Zhang, T., Cai, L. & Yang, R. Triangulation topology aggregation optimizer: A novel mathematics-based meta-heuristic algorithm for continuous optimization and engineering applications. Expert Syst. Appl. 238, 121744 (2024).
Sowmya, R., Premkumar, M. & Jangir, P. Newton-Raphson-based optimizer: A new population-based metaheuristic algorithm for continuous optimization problems. Eng. Appl. Artif. Intell. 128, 107532 (2024).
Yuan, C. et al. Polar lights optimizer: Algorithm and applications in image segmentation and feature selection. Neurocomputing 607, 128427 (2024).
Abdel-Basset, M., Mohamed, R., Azeem, S. A. A., Jameel, M. & Abouhawwash, M. Kepler optimization algorithm: A new metaheuristic algorithm inspired by Kepler’s laws of planetary motion. Knowl.-Based Syst. 268, 110454 (2023).
Ghasemi, M. et al. Geyser inspired algorithm: a new geological-inspired meta-heuristic for real-parameter and constrained engineering optimization. J. Bionic Eng. 21(1), 374-408 (2024).
Shareef, H., Ibrahim, A. A. & Mutlag, A. H. Lightning search algorithm. Appl. Soft Comput. 36, 315–333 (2015).
Hashim, F. A., Mostafa, R. R., Hussien, A. G., Mirjalili, S. & Sallam, K. M. Fick’s Law Algorithm: A physical law-based algorithm for numerical optimization. Knowl.-Based Syst. 260, 110146 (2023).
Askari, Q., Younas, I. & Saeed, M. Political Optimizer: A novel socio-inspired meta-heuristic for global optimization. Knowl.-Based Syst. 195, 105709 (2020).
Moosavi, S. H. S. & Bardsiri, V. K. Poor and rich optimization algorithm: A new human-based and multi populations algorithm. Eng. Appl. Artif. Intell. 86, 165–181 (2019).
Zhu, D., Wang, S., Zhou, C., Yan, S. & Xue, J. Human memory optimization algorithm: A memory-inspired optimizer for global optimization problems. Expert Syst. Appl. 237, 121597 (2024).
Lian, J. & Hui, G. Human evolutionary optimization algorithm. Expert Syst. Appl. 241, 122638 (2024).
Matoušová, I., Trojovský, P., Dehghani, M., Trojovská, E. & Kostra, J. Mother optimization algorithm: A new human-based metaheuristic approach for solving engineering optimization. Sci. Rep. 13(1), 10312 (2023).
Givi, H., & Hubalovska, M. (2023). Skill Optimization Algorithm: A New Human-Based Metaheuristic Technique. Comput. Mater. Contin. 74(1).
Antonijevic, M. et al. Intrusion detection in metaverse environment internet of things systems by metaheuristics tuned two level framework. Sci. Rep. 15(1), 3555 (2025).
Abdel-Salam, M., Chhabra, A., Braik, M., Gharehchopogh, F. S. & Bacanin, N. A Halton enhanced solution-based human evolutionary algorithm for complex optimization and advanced feature selection problems. Knowl. Based Syst. 311, 113062 (2025).
Askr, H., Abdel-Salam, M., Snášel, V. & Hassanien, A. E. A green hydrogen production model from solar powered water electrolyze based on deep chaotic Lévy gazelle optimization. Eng. Sci. Technol. Int. J. 60, 101874 (2024).
Abdel-Salam, M. et al. An adaptive enhanced human memory algorithm for multi-level image segmentation for pathological lung cancer images. Comput. Biol. Med. 183, 109272 (2024).
Nadimi-Shahraki, M. H., Taghian, S. & Mirjalili, S. An improved grey wolf optimizer for solving engineering problems. Expert Syst. Appl. 166, 113917 (2021).
Gong, J. & Karimzadeh Parizi, M. GWMA: The parallel implementation of woodpecker mating algorithm on the GPU. J. Chin. Inst. Eng. 45(6), 556–568 (2022).
Abdel-Salam, M., Alzahrani, A. I., Alblehai, F., Zitar, R. A. & Abualigah, L. An improved Genghis Khan optimizer based on enhanced solution quality strategy for global optimization and feature selection problems. Knowl.-Based Syst. 302, 112347 (2024).
Wolpert, D. H. & Macready, W. G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1(1), 67–82 (1997).
Li, Y., Sun, K., Yao, Q. & Wang, L. A dual-optimization wind speed forecasting model based on deep learning and improved dung beetle optimization algorithm. Energy 286, 129604 (2024).
Zhu, F. et al. Dung beetle optimization algorithm based on quantum computing and multi-strategy fusion for solving engineering problems. Expert Syst. Appl. 236, 121219 (2024).
Mai, C. et al. A novel MPPT technology based on dung beetle optimization algorithm for PV systems under complex partial shade conditions. Sci. Rep. 14(1), 6471 (2024).
Li, P., Zhao, H., Gu, J. & Duan, S. Dynamic constitutive identification of concrete based on improved dung beetle algorithm to optimize long short-term memory model. Sci. Rep. 14(1), 6334 (2024).
Zilong, W. & Peng, S. A multi-strategy dung beetle optimization algorithm for optimizing constrained engineering problems. IEEE Access 11, 98805–98817 (2023).
Zhong, M. et al. A hierarchical multi-leadership sine cosine algorithm to dissolving global optimization and data classification: The COVID-19 case study. Comput. Biol. Med. 164, 107212 (2023).
Parizi, M. K., Keynia, F. & Bardsiri, A. K. HSCWMA: A new hybrid SCA-WMA algorithm for solving optimization problems. Int. J. Inf. Technol. Decis. Mak. 20(02), 775–808 (2021).
Karimzadeh Parizi, M., Keynia, F. & Khatibi Bardsiri, A. OWMA: An improved self-regulatory woodpecker mating algorithm using opposition-based learning and allocation of local memory for solving optimization problems. J. Intell. Fuzzy Syst. 40(1), 919–946 (2021).
Xiao, Y., Cui, H., Hussien, A. G. & Hashim, F. A. MSAO: A multi-strategy boosted snow ablation optimizer for global optimization and real-world engineering applications. Adv. Eng. Inform. 61, 102464 (2024).
Xiao, Y., Cui, H., Khurma, R. A. & Castillo, P. A. Artificial lemming algorithm: a novel bionic meta-heuristic technique for solving real-world engineering optimization problems. Artif. Intell. Rev. 58(3), 84 (2025).
Ridha, H. M. et al. Parameter extraction of single, double, and three diodes photovoltaic model based on guaranteed convergence arithmetic optimization algorithm and modified third order Newton Raphson methods. Renew. Sustain. Energy Rev. 162, 112436 (2022).
El-Dabah, M. A., El-Sehiemy, R. A., Hasanien, H. M. & Saad, B. Photovoltaic model parameters identification using Northern Goshawk Optimization algorithm. Energy 262, 125522 (2023).
Abbassi, A., Mehrez, R. B., Touaiti, B., Abualigah, L. & Touti, E. Parameterization of photovoltaic solar cell double-diode model based on improved arithmetic optimization algorithm. Optik 253, 168600 (2022).
Long, W. et al. Parameters estimation of photovoltaic models using a novel hybrid seagull optimization algorithm. Energy 249, 123760 (2022).
Janamala, V. & Radha Rani, K. Optimal allocation of solar photovoltaic distributed generation in electrical distribution networks using Archimedes optimization algorithm. Clean Energy 6(2), 271–287 (2022).
Lu, Y., Liang, S., Ouyang, H., Li, S. & Wang, G. G. Hybrid multi-group stochastic cooperative particle swarm optimization algorithm and its application to the photovoltaic parameter identification problem. Energy Rep. 9, 4654–4681 (2023).
Duan, Z., Yu, H., Zhang, Q. & Tian, L. Parameter extraction of solar photovoltaic model based on nutcracker optimization algorithm. Appl. Sci. 13(11), 6710 (2023).
Ramadan, A., Kamel, S., Korashy, A., Almalaq, A. & Domínguez-García, J. L. An enhanced Harris Hawk optimization algorithm for parameter estimation of single, double and triple diode photovoltaic models. Soft. Comput. 26(15), 7233–7257 (2022).
Yousri, D., Shaker, Y., Mirjalili, S. & Allam, D. An efficient photovoltaic modeling using an Adaptive Fractional-order Archimedes Optimization Algorithm: Validation with partial shading conditions. Sol. Energy 236, 26–50 (2022).
Mohamed, R. et al. Novel hybrid kepler optimization algorithm for parameter estimation of photovoltaic modules. Sci. Rep. 14(1), 3453 (2024).
Kullampalayam Murugaiyan, N., Chandrasekaran, K., Manoharan, P. & Derebew, B. Leveraging opposition-based learning for solar photovoltaic model parameter estimation with exponential distribution optimization algorithm. Sci. Rep. 14(1), 528 (2024).
Izci, D., Ekinci, S. & Hussien, A. G. Efficient parameter extraction of photovoltaic models with a novel enhanced prairie dog optimization algorithm. Sci. Rep. 14(1), 7945 (2024).
Saadaoui, D. et al. A hybrid optimization algorithm to identify unknown parameters of photovoltaic models under varying operating conditions. Eng. Appl. Artif. Intell. 133, 108544 (2024).
Ekinci, S., Izci, D. & Hussien, A. G. Comparative analysis of the hybrid gazelle-Nelder–Mead algorithm for parameter extraction and optimization of solar photovoltaic systems. IET Renew. Power Gener. 18(6), 959–978 (2024).
Qaraad, M. et al. Quadratic interpolation and a new local search approach to improve particle swarm optimization: Solar photovoltaic parameter estimation. Expert Syst. Appl. 236, 121417 (2024).
Javed, M. Y. et al. Improving the efficiency of photovoltaic-thermoelectric generator system using novel flying squirrel search optimization algorithm: Hybrid renewable and thermal energy system (RTES) for electricity generation. Process Saf. Environ. Prot. 187, 104–116 (2024).
Kahraman, H. T. et al. Dynamic-fitness-distance-balance stochastic fractal search (dFDB-SFS algorithm): an effective metaheuristic for global optimization and accurate photovoltaic modeling. Soft. Comput. 28(9), 6447–6474 (2024).
Izci, D. et al. A new modified version of mountain gazelle optimization for parameter extraction of photovoltaic models. Electr. Eng. 106(5), 6565-6585 (2024).
Çelik, E. et al. Reconfigured single-and double-diode models for improved modelling of solar cells/modules. Sci. Rep. 15(1), 2101 (2025).
Jiao, S. et al. Orthogonally adapted Harris hawks optimization for parameter estimation of photovoltaic models. Energy 203, 117804 (2020).
Abdel-Salam, M., Abualigah, L., Alzahrani, A. I., Alblehai, F. & Jia, H. Boosting crayfish algorithm based on halton adaptive quadratic interpolation and piecewise neighborhood for complex optimization problems. Comput. Methods Appl. Mech. Eng. 432, 117429 (2024).
Abdel-Salam, M., Hu, G., Çelik, E., Gharehchopogh, F. S. & El-Hasnony, I. M. Chaotic RIME optimization algorithm with adaptive mutualism for feature selection problems. Comput. Biol. Med. 179, 108803 (2024).
Yu, Y. et al. CBSO: a memetic brain storm optimization with chaotic local search. Memetic Comput. 10, 353–367 (2018).
Li, Y., Wang, S. & Yang, B. An improved differential evolution algorithm with dual mutation strategies collaboration. Expert Syst. Appl. 153, 113451 (2020).
Ling, S. H. et al. Hybrid particle swarm optimization with wavelet mutation and its industrial applications. IEEE Trans. Syst. Man Cybernet. Part B (Cybernetics) 38(3), 743–763 (2008).
Ahmadianfar, I., Bozorg-Haddad, O. & Chu, X. Gradient-based optimizer: A new metaheuristic optimization algorithm. Inf. Sci. 540, 131–159 (2020).
Hassanien, A. E., Darwish, A. & Elghamrawy, S. M. (eds) Artificial intelligence for environmental sustainability and green initiatives. (Springer, 2024).
Abdel-Salam, M., Askr, H. & Hassanien, A. E. Adaptive chaotic dynamic learning-based gazelle optimization algorithm for feature selection problems. Expert Syst. Appl. 256, 124882 (2024).
Bujok, P., & Kolenovsky, P. Eigen crossover in cooperative model of evolutionary algorithms applied to CEC 2022 single objective numerical optimisation. In 2022 IEEE congress on evolutionary computation (CEC) (pp. 1–8). IEEE (2022).
Wu, G., Mallipeddi, R., & Suganthan, P. N. (2017). Problem definitions and evaluation criteria for the CEC 2017 competition on constrained real-parameter optimization. National University of Defense Technology, Changsha, Hunan, PR China and Kyungpook National University, Daegu, South Korea and Nanyang Technological University, Singapore, Technical Report.
Houssein, E. H., Saad, M. R., Hashim, F. A., Shaban, H. & Hassaballah, M. Lévy flight distribution: A new metaheuristic algorithm for solving engineering optimization problems. Eng. Appl. Artif. Intell. 94, 103731 (2020).
Heidari, A. A. et al. Harris hawks optimization: Algorithm and applications. Futur. Gener. Comput. Syst. 97, 849-872 (2019).
Abualigah, L., Diabat, A., Mirjalili, S., Abd Elaziz, M. & Gandomi, A. H. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 376, 113609 (2021).
Mirjalili, S. & Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 95, 51–67 (2016).
Li, S., Chen, H., Wang, M., Heidari, A. A. & Mirjalili, S. Slime mould algorithm: A new method for stochastic optimization. Futur. Gener. Comput. Syst. 111, 300–323 (2020).
Wang, L., Cao, Q., Zhang, Z., Mirjalili, S. & Zhao, W. Artificial rabbits optimization: A new bio-inspired meta-heuristic algorithm for solving engineering optimization problems. Eng. Appl. Artif. Intell. 114, 105082 (2022).
Elhoseny, M., Abdel-salam, M. & El-Hasnony, I. M. An improved multi-strategy Golden Jackal algorithm for real world engineering problems. Knowl.-Based Syst. 295, 111725 (2024).
Çelik, E. et al. Novel distance-fitness learning scheme for ameliorating metaheuristic optimization. Eng. Sci. Technol. Int. J. 65, 102053 (2025).
Hu, G., Song, K. & Abdel-salam, M. Sub-population evolutionary particle swarm optimization with dynamic fitness-distance balance and elite reverse learning for engineering design problems. Adv. Eng. Softw. 202, 103866 (2025).
Khodadadi, N., Snasel, V. & Mirjalili, S. Dynamic arithmetic optimization algorithm for truss optimization under natural frequency constraints. IEEe Access 10, 16188–16208 (2022).
Nenavath, H. & Jatoth, R. K. Hybridizing sine cosine algorithm with differential evolution for global optimization and object tracking. Appl. Soft Comput. 62, 1019–1043 (2018).
Yu, K., Wang, X. & Wang, Z. An improved teaching-learning-based optimization algorithm for numerical and engineering optimization problems. J. Intell. Manuf. 27, 831–843 (2016).
Ling, Y., Zhou, Y. & Luo, Q. Lévy flight trajectory-based whale optimization algorithm for global optimization. IEEE access 5, 6168–6186 (2017).
Lozano, J. A. (ed.) Towards a New Evolutionary Computation: Advances on Estimation of Distribution Algorithms, Vol. 192 (Springer, 2006).
Awad, N. H., Ali, M. Z., & Suganthan, P. N. Ensemble sinusoidal differential covariance matrix adaptation with Euclidean neighborhood for solving CEC2017 benchmark problems. In 2017 IEEE congress on evolutionary computation (CEC) (pp. 372–379). IEEE (2017).
Fan, J., Xiong, S., Wang, J., & Gong, C. (2008). IMODE: Improving multi-objective differential evolution algorithm. In 2008 fourth international conference on natural computation (Vol. 1, pp. 212–216). IEEE.
Mohamed, A. W., Hadi, A. A., Mohamed, A. K., & Awad, N. H. (2020). Evaluating the performance of adaptive gainingsharing knowledge based algorithm on CEC 2020 benchmark problems. In 2020 IEEE congress on evolutionary computation (CEC) (pp. 1–8). IEEE.
Elhosseny, M., Abdel-Salam, M. & El-Hasnony, I. M. Adaptive dynamic crayfish algorithm with multi-enhanced strategy for global high-dimensional optimization and real-engineering problems. Sci. Rep. 15(1), 10656 (2025).
Derrac, J., García, S., Molina, D. & Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 1(1), 3–18 (2011).
Gao, X. et al. Parameter extraction of solar cell models using improved shuffled complex evolution algorithm. Energy Convers. Manage. 157, 460–479 (2018).
Tong, N. T. & Pora, W. A parameter extraction technique exploiting intrinsic properties of solar cells. Appl. Energy 176, 104–115 (2016).
Yu, K. et al. A performance-guided JAYA algorithm for parameters identification of photovoltaic cell and module. Appl. Energy 237, 241–257 (2019).
Saleem, H. & Karmalkar, S. An analytical method to extract the physical parameters of a solar cell from four points on the illuminated $ J - V $ curve. IEEE Electron Device Lett. 30(4), 349–352 (2009).
Yu, X., Hu, Z., Wang, X. & Luo, W. Ranking teaching–learning-based optimization algorithm to estimate the parameters of solar models. Eng. Appl. Artif. Intell. 123, 106225 (2023).
Maden, D., Çelik, E., Houssein, E. H. & Sharma, G. Squirrel search algorithm applied to effective estimation of solar PV model parameters: a real-world practice. Neural Comput. Appl. 35(18), 13529–13546 (2023).
Yu, K., Liang, J. J., Qu, B. Y., Chen, X. & Wang, H. Parameters identification of photovoltaic models using an improved JAYA optimization algorithm. Energy Convers. Manage. 150, 742–753 (2017).
Nicaire, N. F., Steve, P. N., Salome, N. E. & Grégroire, A. O. Parameter estimation of the photovoltaic system using bald eagle search (BES) algorithm. Int. J. Photoenergy 2021(1), 4343203 (2021).
Bogar, E. Chaos game optimization-least squares algorithm for photovoltaic parameter estimation. Arab. J. Sci. Eng. 48(5), 6321–6340 (2023).
Yu, K., Liang, J. J., Qu, B. Y., Cheng, Z. & Wang, H. Multiple learning backtracking search algorithm for estimating parameters of photovoltaic models. Appl. Energy 226, 408–422 (2018).
Author information
Authors and Affiliations
Contributions
M.H: Methodology, Software, Conceptualization, Data curation, Validation, Writing—review & editing. M.A.S: Supervision, Software, Methodology, Conceptualization, Formal analysis, Investigation, Visualization, Writing—review & editing. A.N: Visualization, Formal analysis, Writing – review & editing, Conceptualization. E.C: Conceptualization, Formal analysis, Resources, Data curation, Writing—original draft, Writing—review & editing. S.M: Writing—review & editing, Formal analysis, Resources, Methodology, Data curation. N.B: Writing—review & editing, Conceptualization, Visualization, Methodology, Data curation. I.H: Writing—review & editing, Visualization, Formal analysis, Methodology, Data curation.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Elhosseny, M., Abdel-Salam, M., Nayyar, A. et al. Elite Bernoulli-based mutated dung beetle algorithm for global complex problems and parameter estimation of solar photovoltaic models. Sci Rep 15, 36304 (2025). https://doi.org/10.1038/s41598-025-06108-4
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-06108-4



















