Introduction

Metaheuristic optimization algorithms have emerged as a prominent class of optimization techniques capable of addressing complex, nonlinear, and high-dimensional problems1,2. Unlike classical optimization methods, metaheuristics do not require derivative information and are more flexible in handling a wide range of objective functions3. By drawing inspiration from natural and physical processes, these algorithms effectively explore and exploit solution spaces. Prominent examples of metaheuristic algorithms include genetic algorithms (GA)4, particle swarm optimization (PSO)5, ant colony optimization (ACO)6, and gravitational search algorithm (GSA)7. There are also the latest algorithms, such as draco lizard optimizer8, eurasian lynx optimizer9, and artificial meerkat algorithm10.

One critical feature of many advanced metaheuristic methods is their hierarchical structure, which organizes the population into different layers or subgroups11. The hierarchical structure facilitates efficient information exchange and diversity preservation by promoting interactions between various segments of the population12. This layered approach is crucial for preventing premature convergence and ensuring comprehensive exploration of the solution space. By structuring the population hierarchically, metaheuristic algorithms can effectively balance the dual demands of exploration (searching new regions) and exploitation (intensifying search around promising solutions).

Manta ray foraging optimization (MRFO)13 has emerged as a powerful technique inspired by the unique foraging behaviors of manta rays, including chain foraging, cyclone foraging, and somersault foraging strategies. These strategies enable manta rays to adapt dynamically to changing environmental conditions, allowing MRFO to maintain a balance between exploration and exploitation. The original MRFO has shown significant effectiveness in addressing a variety of practical optimization challenges. Nevertheless, as is common with many metaheuristic approaches, MRFO faces challenges such as premature convergence and reduced population diversity, which limit its capacity to reliably discover the global optimal solution. Addressing these challenges requires further enhancements to ensure effective exploration of the solution space and to prevent convergence to suboptimal solutions.

In recent years, numerous advancements have been made in the MRFO algorithm, aimed at overcoming its limitations and enhancing its performance. Researchers have focused on strategies to mitigate premature convergence and enhance population diversity. For instance, adaptive control mechanisms for parameters have been proposed, allowing the algorithm to dynamically adjust based on the state of the search process14,15. Hybrid approaches that integrate MRFO with other metaheuristic algorithms, such as GA and PSO, have also been introduced to enhance global exploration capabilities and improve solution quality16,17.

Another line of research has focused on developing novel selection strategies to diversify the population and maintain a robust search process. These strategies often involve elite preservation, and opposition-based learning techniques to help balance exploration and exploitation12,18. Elite preservation ensures that the best solutions are retained across generations, opposition-based learning has been demonstrated to be particularly effective in enhancing the convergence rate of MRFO by periodically reinitializing certain members of the population with more promising candidates19,20. Additionally, the incorporation of chaotic maps and Levy flights has been demonstrated to further enhance MRFO’s search efficiency, minimizing the probability of the algorithm getting trapped in local optima21,22. Chaotic maps help in better initial population distribution, which is crucial for global exploration, whereas Levy flights provide a mechanism for making larger, more diverse jumps in the search space, especially useful for avoiding local optima.

Moreover, multi-objective extensions of MRFO have been introduced to address problems involving multiple conflicting objectives. These multi-objective versions utilize Pareto-based selection mechanisms and crowding distance measures to maintain a diverse set of solutions while achieving an effective trade-off among objectives23,24. Recently, more advanced multi-objective methods such as inertia weight decreasing approaches and adaptive trade-off models have been incorporated to improve the flexibility and robustness of multi-objective MRFO25,26. These improvements allow MRFO to handle practical, real-world scenarios where multiple criteria must be optimized simultaneously, providing decision-makers with a set of Pareto-optimal solutions.

Another noteworthy advancement is the incorporation of memory-based strategies within MRFO, where successful solutions from previous iterations are stored and used to guide future search efforts15,27. Such strategies allow the algorithm to learn from historical data, which significantly improves the efficiency of the search, especially in dynamic environments. Furthermore, adaptive step-size mechanisms have been developed to refine the local search capabilities of MRFO. By adaptively adjusting the search step size based on the improvement rate in the fitness value, these methods help strike a better balance between exploration and exploitation21,28.

MRFO also has been effectively integrated with advanced machine learning methodologies, leading to significant advancements in feature selection, hyperparameter optimization, and classification tasks, particularly in medical and biomedical applications. MRFO has been employed to optimize support vector machines (SVM) and convolutional neural networks (CNN), resulting in enhanced classification accuracy and reduced data dimensionality29. MRFO has been hybridized with CNNs for brain tumor detection, demonstrating superior performance compared to traditional optimization methods30. Collectively, MRFO substantially enhances the accuracy, robustness, and generalization capabilities of machine learning models in the medical and biomedical domains, providing a powerful tool for addressing complex classification and feature selection challenges.

Overall, these improvements have made MRFO a more versatile and powerful optimization tool capable of solving a broader range of complex problems. The advancements not only extend the applicability of MRFO to diverse fields such as engineering design, energy systems, and biomedicine but also enhance its effectiveness in handling practical challenges characterized by multiple objectives, dynamic environments, and computational constraints.

In recent years, numerous parameter estimation methods for solar photovoltaic models have emerged. In31, experimental comparisons between hybrid methods and metaheuristic methods for photovoltaic parameter estimation revealed that hybrid methods provide more accurate parameter estimation. In32, the trust-region dogleg (TRD) and trust-region reflective (TRR) algorithms were employed to estimate three unknown parameters in the solar photovoltaic model, while the remaining two parameters were determined using an analytical approach. In33, a hybrid approach combining grey wolf optimization (GWO) and analytical methods achieved promising results in photovoltaic parameter estimation. In34, a comparative analysis of existing photovoltaic parameter estimation techniques revealed that hybrid-based parameter estimation methods achieve more accurate photovoltaic characteristics. In35, a multi-hybrid evolutionary algorithm based on the flower pollination algorithm (FPA), grey wolf optimizer (GWO), and differential evolution (DE) algorithm was used for precise and effective parameter estimation of photovoltaic systems. In36, a new algorithm, the pelican optimization algorithm (POA), was proposed for the optimal parameter estimation of photovoltaic systems, achieving promising results. In37, reinforcement learning and nonlinear hunting strategies were integrated into the golden jackal optimizer, achieving promising results in photovoltaic parameter estimation, making it a reliable tool for this task. In38, a multi-strategy (self-adaptive weighting mechanism, chaotic elite learning approach, and experience-based learning method) tree seed algorithm was used for photovoltaic model parameter estimation, achieving successful results. In39, the study found that the hybrid approach combining the linear slope method and GWO yields photovoltaic parameters that are closer to the real characteristics. In40, a nonlinear least squares (NLS) objective function was constructed using manufacturer datasheet information, and particle swarm optimization (PSO) was employed to estimate the unknown parameters of the photovoltaic model, achieving higher accuracy.

Within these study, we introduce the hierarchical guided manta ray foraging optimization (HGMRFO), which incorporates an adaptive somersault factor to replace the original fixed parameter \(S\), and a hierarchical guidance search mechanism to replace the sole reliance on the current best individual during somersault foraging. These modifications enable effective directionality in individual searches and facilitate an improved equilibrium between exploration and exploitation. During the experimental phase, the optimal parameter configuration for HGMRFO was systematically determined, and the effectiveness of the complete hierarchical structure was demonstrated. Comparative analyses were conducted using the IEEE CEC2017 benchmark and IEEE CEC2011 real-world optimization problem sets, thereby validating the superior performance and practicality of HGMRFO. Furthermore, the proposed algorithm was applied to parameter estimation in solar photovoltaic models, where it exhibited promising real-world applicability and effectiveness.

The main contributions of this research are as follows:

1. Introduction of the hierarchical guidance mechanism: This paper innovatively alters the population structure, guiding individuals toward the global optimal solution through layered interactions. This approach overcomes the limitations of many existing algorithms that focus solely on searching around the current best solution. On the CEC2017 benchmark functions, the algorithm achieved an average win rate of 73.15%, and it obtained the most optimal solutions on the CEC2011 real-world optimization problems. These results effectively demonstrate that the mechanism significantly enhances the algorithm’s global search capability in complex optimization problems.

2. Adaptive parameter adjustment strategy: An adaptive parameter tuning mechanism was implemented to achieve a dynamic balance between the exploration and exploitation stages of the search process. This strategy alleviates premature convergence issues caused by fixed parameter settings. The convergence curves obtained across various dimensional functions clearly show that this strategy significantly improves the algorithm’s robustness and effective convergence.

3. Application in photovoltaic model parameter estimation: The improved algorithm was applied to parameter estimation for six photovoltaic (PV) models, achieving a 97.62% success rate. This not only verifies the algorithm’s comprehensive performance in PV modeling but also demonstrates its potential practical utility and broad applicability in the renewable energy field.

The remainder of this paper is structured as follows: Section “Related work” provides an overview of related work, including a detailed introduction to the fundamental MRFO and classifications of population structures. Section “Hierarchical guided manta ray foraging optimization (HGMRFO)” identifies the limitations of MRFO and subsequently proposes the HGMRFO as an enhancement. Section “Experiments and analysis” presents the experimental and analytical aspects of the study, where the parameter sensitivity and hierarchical interactions of HGMRFO are comprehensively analyzed, and comparative experiments with state-of-the-art algorithms are conducted on benchmark functions and real-world optimization problems. Moreover, the practical applicability of HGMRFO is validated through its application to parameter estimation in photovoltaic models. Section “Conclusions” concludes the paper by summarizing key findings and outlining potential directions for future research.

Related work

Basic MRFO

MRFO algorithm is inspired by the unique foraging behavior of manta rays and employs three distinct mechanisms-chain foraging, cyclone foraging, and somersault foraging-to effectively drive the optimization process. These mechanisms collectively simulate the exploratory and exploitative capabilities of the swarm, ultimately enhancing optimization performance.

In the MRFO framework, a population of \(N\) manta ray individuals is represented as \(X_{i}=\left( x_{i}^{1}, x_{i}^{2}, \ldots , x_{i}^{d} \right)\), where \(i \in \{1,2, \ldots , N\}\) and \(x_{i}^d\) denotes the position of the \(i\)th individual in the \(d\)th dimension. Each individual employs different foraging behaviors to navigate the search space, described in detail below.

Fig. 1
figure 1

Flowchart of MRFO.

The chain foraging mechanism replicates the behavior of manta rays moving in a coordinated chain formation, in which each individual follows the one ahead. The position of each manta ray is influenced by both the current best solution and the individual directly in front of it, which enhances both exploration and convergence toward the optimal solution. The update equation for chain foraging is formulated as:

$$\begin{aligned} \begin{aligned} x_{i}^{d}(t+1)=\left\{ \begin{array}{cl} x_{i}^{d}(t)+r \cdot \left( x_{\text{ best } }^{d}(t)-x_{i}^{d}(t)\right) & \\ +\alpha \cdot \left( x_{\text{ best } }^{d}(t)-x_{i}^{d}(t)\right) , & i=1 \\ x_{i}^{d}(t)+r \cdot \left( x_{i-1}^{d}(t)-x_{i}^{d}(t)\right) & \\ +\alpha \cdot \left( x_{\text{ best } }^{d}(t)-x_{i}^{d}(t)\right) , & i=2, \ldots , N \end{array}\right. \end{aligned} \end{aligned}$$
(1)
$$\begin{aligned} \begin{aligned} \alpha =2 \cdot r \cdot \sqrt{|\log (r)|} \end{aligned} \end{aligned}$$
(2)

where, \(X_{\text {best}}=\left( x_{\text {best}}^{1}, x_{\text {best}}^{2}, \ldots , x_{\text {best}}^{d}\right)\) represents the current best individual, \(x_{\text {best}}^{d}(t)\) denotes the position of the best individual in the \(d\)th dimension at time \(t\). \(r\) is a random number within the range \([0, 1]\). The coefficient \(\alpha\) is introduced to adjust the weight of the search direction, enhancing the effectiveness of the foraging behavior.

In the cyclone foraging phase, manta rays exhibit a spiraling behavior around the current best solution, which enhances their ability to perform local search. Unlike chain foraging, this behavior aims to increase the exploration capability of the population by generating a swirling motion around the optimal solution. The position update in cyclone foraging is modeled as:

$$\begin{aligned} \begin{aligned} x_{i}^{d}(t+1)=\left\{ \begin{array}{cl} x_{\text{ best } }^{d}+r \cdot \left( x_{\text{ best } }^{d}(t)-x_{i}^{d}(t)\right) & \\ +\beta \cdot \left( x_{\text{ best } }^{d}(t)-x_{i}^{d}(t)\right) , & i=1 \\ x_{\text{ best } }^{d}+r \cdot \left( x_{i-1}^{d}(t)-x_{i}^{d}(t)\right) \\ +\beta \cdot \left( x_{\text{ best } }^{d}(t)-x_{i}^{d}(t)\right) , & i=2, \ldots , N \end{array}\right. \end{aligned} \end{aligned}$$
(3)
$$\begin{aligned} \begin{aligned} \beta =2 e^{r_{1} \frac{T-t+1}{T}} \cdot \sin \left( 2 \pi r_{1}\right) \end{aligned} \end{aligned}$$
(4)

where \(\beta\) is a dynamic coefficient dependent on the random number \(r_1 \in [0,1]\), and \(T\) represents the maximum number of iterations. The cyclone foraging mechanism includes an additional feature that selects a random individual as a reference point to preserve variability and mitigate the likelihood of getting trapped in local optima, formulated as:

$$\begin{aligned} \begin{aligned} x_{\text{ rand } }^{d}=L b^{d}+r \cdot \left( U b^{d}-L b^{d}\right) \end{aligned} \end{aligned}$$
(5)
$$\begin{aligned} \begin{aligned} x_{i}^{d}(t+1)=\left\{ \begin{array}{cl} x_{\text{ rand } }^{d}+r \cdot \left( x_{\text{ rand } }^{d}(t)-x_{i}^{d}(t)\right) & \\ +\beta \cdot \left( x_{\text{ rand } }^{d}(t)-x_{i}^{d}(t)\right) , & i=1 \\ x_{\text{ rand } }^{d}+r \cdot \left( x_{i-1}^{d}(t)-x_{i}^{d}(t)\right) & \\ +\beta \cdot \left( x_{\text{ rand } }^{d}(t)-x_{i}^{d}(t)\right) , & i=2, \ldots , N \end{array}\right. \end{aligned} \end{aligned}$$
(6)

where \(x_{\text {rand}}^{d}\) denotes the position of a randomly selected individual, and \(L b^{d}\), \(U b^{d}\) represent the lower and upper bounds for dimension \(d\). This mechanism ensures improved solution diversity, effectively preventing premature convergence.

The final strategy, somersault foraging, simulates the somersault movements of manta rays around the optimal solution. This behavior aims to balance exploration and exploitation by allowing individuals to adjust their positions with respect to the best-known solution, thereby improving convergence accuracy. The somersault foraging update is expressed as:

$$\begin{aligned} \begin{aligned} x_{i}^{d}(t+1)=x_{i}^{d}(t)+S \cdot \left( r_{2} \cdot x_{\text{ best } }^{d}-r_{3} \cdot x_{i}^{d}(t)\right) , \quad i=1, \ldots , N \end{aligned} \end{aligned}$$
(7)

where \(S\) is a constant parameter controlling the somersault magnitude, and \(r_2, r_3 \in [0, 1]\) are random numbers. This mechanism progressively reduces the movement range of each individual as iterations proceed, thereby enhancing the precision of the solution search.

The overall workflow of MRFO is depicted in Fig. 1. Initially, either chain or cyclone foraging is selected based on a random criterion. During the cyclone foraging phase, the decision of whether to use the current best solution or a random reference is made based on the ratio \(t/T\), thereby balancing global and local exploration. Ultimately, each individual performs somersault foraging to refine its position, ensuring steady convergence to the optimal solution. This coordinated use of chain, cyclone, and somersault foraging effectively balances the exploration and exploitation capabilities of the algorithm, thereby enhancing both its robustness and convergence efficiency.

Population structure

Population structure plays a crucial role in optimization algorithms, involving the allocation, organization, and management of individuals to identify superior solutions. The design of population structure significantly impacts the exchange of information among individuals and the maintenance of population diversity, both of which are critical for effective exploration and exploitation within the search space. Common population structures are generally classified into three categories:

  1. (1)

    Panmictic structure: The panmictic structure is the most fundamental and widely adopted population configuration41. In this structure, individuals interact with each other by randomly selecting partners, which expedites the dissemination of information across the population. However, frequent and random interactions often result in insufficient population diversity, leading to premature convergence and hindering comprehensive exploration of the solution space.

  2. (2)

    Neighborhood structures: To address the premature convergence problem inherent in panmictic structures, neighborhood-based configurations have been developed, leveraging both fitness values and spatial topology to enhance information exchange42,43. By promoting diversified neighborhood interactions, these structures facilitate evolutionary processes and enhance algorithmic performance. Neighborhood structures are predominantly categorized into cellular structures and distributed structures44,45.

In cellular structures, individuals are organized into a grid, interacting exclusively with their immediate neighbors, which may partially overlap. This localized interaction helps maintain population diversity and mitigate premature convergence42. For instance, Dorronsoro and Alba introduced a cellular genetic algorithm that demonstrated the effective diffusion of solutions throughout the population45. Cellular structures have also been employed to enhance differential evolution variants46 and address dynamic optimization problems by integrating mechanisms such as clonal selection47. In the context of differential evolution, cellular direction information has been utilized to improve exploration and convergence rates through a combined approach48.

In distributed structures, the population is partitioned into multiple subpopulations that evolve independently while sharing information via migration strategies. Each subpopulation searches the solution space in its specific region, and migration strategies facilitate effective information exchange among different subpopulations49. Through this migration mechanism, distributed structures effectively expand the exploration range and further enhance population diversity. For example, a PSO divided the population into two types of swarms-one composed of elite individuals and the other used for exploring alternative regions of the solution space50. In another study, a multicultural migration strategy was proposed to further increase population diversity in distributed differential evolution51.

The hierarchical structure is a specific type of distributed configuration that allocates individuals across multiple layers, each serving distinct purposes, while enabling dynamic information exchange among the layers52. This stratification facilitates the evolution of subsequent generations by promoting inter-layer communication. For example, in dynamic hierarchical particle swarm optimization, particles can move vertically within the hierarchy, allowing the most promising individuals to ascend and optimize the population at multiple levels53. Hierarchical designs, such as the multi-layered GSA, significantly improve the balance between exploration and exploitation through hierarchical interactions11. In the context of MRFO, Tang et al. introduced an update mechanism that incorporates weighted fitness-distance balance selection, further enhancing diversity12. Similarly, Wang et al. proposed a deep-layered differential evolution approach comprising four distinct layers, effectively balancing diversity, exploration, and exploitation54.

3) Structures with small-world and scale-free attributes: In addition to panmictic and neighborhood structures, some population configurations exhibit small-world and scale-free properties. The Watts-Strogatz and Barabási-Albert models are commonly employed to construct small-world networks and scale-free topologies, respectively55. Small-world networks uniquely blend features of random graphs and regular lattices, significantly improving the balance between exploration and exploitation during the search process41. For example, Zhang et al. used scale-free networks to adapt the PSO population structure, endowing it with scale-free properties that significantly improved population diversity and algorithm performance56. In another study, Yu et al. leveraged neighborhood information from scale-free networks to refine the mutation operator in differential evolution, thus improving the algorithm’s capacity for both exploration and exploitation of the search space57.

In conclusion, population structure is fundamental to the success of optimization algorithms. Through careful design of topological structures and strategic interactions among individuals, population structures can achieve an effective balance between exploration and exploitation, significantly improving both the efficiency and quality of solution searches58,59,60. Different population structures offer unique advantages in terms of information exchange and diversity maintenance, making them suitable for a wide range of optimization problems.

Hierarchical guided manta ray foraging optimization (HGMRFO)

Issues of MRFO

The somersault foraging mechanism in the MRFO algorithm, which aims to replicate the intricate movement behavior of manta rays during their foraging activities, presents notable limitations in its current form. These limitations are evident in the following update equation:

$$x_{i}^{d}(t+1) = x_{i}^{d}(t) + S \cdot \left( r_{2} \cdot x_{\text {best}}^{d} - r_{3} \cdot x_{i}^{d}(t) \right) , \quad i=1, \ldots , N$$

The main limitations are:

  1. (a)

    Fixed parameter \(S\): The parameter \(S\), which controls the magnitude of the somersault movement, is kept constant throughout the optimization process. This static setting imposes significant constraints on the adaptability of the algorithm’s search behavior. During the initial optimization stages, a larger value of \(S\) would be beneficial for promoting global exploration and preventing premature convergence to local optima. Conversely, in the latter stages, a smaller value would enhance the ability to fine-tune the search around promising regions. By keeping \(S\) fixed, MRFO struggles to dynamically balance exploration and exploitation, leading to a less efficient convergence rate and ultimately compromising the quality of the obtained solution.

  2. (b)

    Limited exploration mode centered on the current best individual: The somersault foraging mechanism directs all individuals to perform movements around the current best solution, \(x_{\text {best}}\). While this approach aims to accelerate convergence by intensifying focus on promising areas of the solution space, it inherently suffers from a significant drawback: it considerably reduces population diversity over time. As individuals concentrate around \(x_{\text {best}}\), the search space becomes increasingly constrained, leading to rapid convergence to a localized region. This reduction in diversity weakens the algorithm’s capacity for thorough exploration of the solution space, increasing the risk of MRFO getting stuck in local optima. This problem becomes especially evident in complex, multimodal landscapes, where the presence of numerous local optima makes identifying the global optimum more challenging.

In conclusion, the somersault foraging mechanism’s dependence on a fixed parameter \(S\) and its singular focus on the current best individual substantially limit MRFO’s capacity to balance exploration and exploitation effectively. These limitations heighten the risk of premature convergence and undermine the robustness of the search process. To address these issues, enhancements could include dynamically adapting the parameter \(S\) based on the current state of the search, as well as introducing alternative movement strategies that foster greater exploration diversity. Such improvements are anticipated to bolster MRFO’s adaptability, enhance convergence efficiency, and improve overall solution quality.

Proposed HGMRFO

Adaptive somersault factor (\(S\))

In the MRFO, the somersault factor \(S\) is generally a constant value that determines the range for somersault searches centered around the current best solution. However, a fixed value for \(S\) presents significant limitations at different optimization stages: a larger \(S\) value is beneficial in the early stages to enhance exploration and discover more unknown regions, while a smaller \(S\) is more suitable for fine-tuning solutions during the local exploitation phase. Therefore, a fixed \(S\) value lacks the flexibility needed to adapt to the dynamic requirements of various optimization phases, which ultimately restricts the effectiveness of global exploration and local exploitation.

To address this limitation, an adaptive somersault factor can be introduced to replace the fixed constant \(S\), allowing it to be dynamically adjusted based on the evolving conditions of the search process. The adaptive \(S\) is calculated as follows:

$$\begin{aligned} \begin{aligned} S = 2 \times \text {arctan}(|f_{\text {best}} - f_{\text {global}} |) \end{aligned} \end{aligned}$$
(8)

where \(f_{\text {best}}\) represents the best fitness value among the current population, and \(f_{\text {global}}\) denotes the global best fitness value found so far. By applying the arctangent function to the absolute difference between \(f_{\text {best}}\) and \(f_{\text {global}}\), a dynamically adjusted value of \(S\) is obtained.

The arctangent function exhibits a characteristic where its output changes rapidly for small inputs and becomes more gradual for larger inputs. This property implies that when the difference between \(f_{\text {best}}\) and \(f_{\text {global}}\) is large, the computed value of \(S\) will also be relatively large, thereby enhancing exploration ability and helping the population escape local optima. Conversely, when the difference is small, \(S\) will decrease, allowing for more focused local search and improving exploitation during the later stages of optimization.

Through this adaptive adjustment mechanism, the somersault factor \(S\) varies dynamically according to the current search state, thereby autonomously achieving an appropriate equilibrium between global exploration and local refinement. This strategy effectively addresses the challenges associated with a fixed parameter \(S\), substantially enhancing the rate of convergence as well as the algorithm’s capability to perform a thorough global search, thereby enabling it to adapt more effectively to the changing requirements of the search process across different phases.

Hierarchical guidance mechanism in MRFO

Fig. 2
figure 2

Illustration of the search trajectory of HGMRFO.

Algorithm 1
figure a

HGMRFO

In the enhanced MRFO, a hierarchical structure is incorporated to strengthen both the exploratory and exploitative abilities during the somersault foraging phase, thereby forming a hierarchical guidance mechanism. This hierarchical structure consists of four primary layers: the population layer, the personal-best layer, the archive layer, and the rank-based selective pressure layer. These layers play a crucial role in guiding the somersault foraging process.

Specifically, the original somersault foraging equation in MRFO is:

$$x_{i}^{d}(t+1)=x_{i}^{d}(t)+S \cdot \left( r_{2} \cdot x_{\text{ best } }^{d}-r_{3} \cdot x_{i}^{d}(t)\right) , \quad i=1, \ldots , N$$

With the introduction of the hierarchical guidance mechanism, the somersault foraging equation is modified as follows:

$$\begin{aligned} \begin{aligned} x_{i}^{d}(t+1) =&\; x_{i}^{d}(t) + S \cdot \left( r_{2} \cdot ( x_{\text {Pbest}}^{d}(t) - x_{i}^{d}(t) + x_{R1}^{d}(t) - x_{R2}^{d}(t) )\right. \\&\left. - r_{3} \cdot x_{i}^{d}(t) \right) , \quad i = 1, \ldots , N, \quad R1 \ne R2 \ne i \end{aligned} \end{aligned}$$
(9)

In this new formula, the original \(x_{\text {best}}^{d}\) is replaced by \(x_{\text {Pbest}}^{d}(t)-x_{i}^{d}(t)+x_{R1}^{d}(t)-x_{R2}^{d}(t)\), thereby significantly enhancing the population’s ability to converge towards the global optimum. Specifically:

  1. (1)

    Population layer: This foundational layer encompasses all individuals. During the somersault foraging phase, each individual in the population layer adjusts its movement direction based on information provided by the other layers.

  2. (2)

    Personal-best layer: This layer is composed of the top \(P_{N}\) high-quality individuals from the population and serves as a replacement for the global best individual in the original algorithm. During the somersault foraging phase, one individual \(x_{\text {Pbest}}^{d}(t)\) is randomly selected from this layer to provide guidance for individuals within the population, thereby enhancing exploitation capability and effectively preventing convergence to local optima.

  3. (3)

    Archive layer: The archive layer stores individuals that were eliminated during the greedy selection process. These individuals provide diverse search directions, enriching population diversity. By randomly selecting individuals \(x_{R2}^{d}(t)\) from the archive layer, individuals can obtain more diverse movement directions.

  4. (4)

    Rank-based selective pressure layer: In this layer, all individuals are ranked based on their fitness, and one individual \(x_{R1}^{d}(t)\) is randomly selected to participate in the somersault foraging process. High-quality individuals have a higher probability of being selected, which helps accelerate convergence, while the occasional selection of lower-quality individuals promotes solution diversity and prevents premature convergence.

To better illustrate the working principle of the hierarchical guidance mechanism, Fig. 2 visually depicts the movement trajectory during the somersault foraging phase. In the figure, \(\bigstar\) represents the global optimum, and the positions of \(X_{\text {Pbest}}\) and \(X_{R1}\) are closer to the global optimum. The arrows in the figure represent the direction of movement, including vectors \(r_{2}(X_{\text {Pbest}} - X_i)\) and \(r_{2}(X_{R1} - X_{R2})\). The combination of these vectors results in an individual’s final movement direction that tends towards the global optimum, thereby enhancing the overall search capability and convergence performance of the population.

By incorporating the hierarchical layers, the somersault foraging phase is no longer confined to the original global best individual-based update strategy. Instead, it integrates information from the personal-best layer, archive layer, and rank-based selective pressure layer to dynamically adjust the individuals’ trade-off between broad exploration of the solution space and focused exploitation of promising regions. This improvement not only significantly enhances MRFO’s exploration capabilities but also bolsters its exploitation potential, allowing individuals to perform a more comprehensive exploration of the solution space and effectively circumvent local optima. The pseudocode for HGMRFO is provided in Algorithm 1.

Experiments and analysis

Experimental setup

To systematically evaluate the performance of HGMRFO, this study conducted simulations comparing HGMRFO with several state-of-the-art optimization algorithms on the IEEE CEC2017 benchmark function set. It should be noted that the F2 function was excluded from the experiments due to instability in the results. Additionally, to validate the practical applicability of HGMRFO, comparative experiments were carried out with other optimization algorithms on the IEEE CEC2011 real-world optimization problem set. Finally, to assess the performance of HGMRFO in parameter estimation for photovoltaic models, we selected six different photovoltaic models and compared HGMRFO with other methods for parameter estimation. For detailed descriptions of the IEEE CEC2017 benchmark function set, the IEEE CEC2011 real-world optimization problem set, and the photovoltaic models, please refer to references2,61,62.

In this paper, we conducted an experimental comparison on the IEEE CEC2017 benchmark functions and IEEE CEC2011 real-world optimization problems. For all comparison algorithms, the population size \(N=100\), and the maximum number of function evaluations (MFEs) was 10,000D, where \(D\) represents the problem dimension. Each algorithm was run independently 51 times on each function to ensure statistical significance of the results. In the parameter estimation experiments for photovoltaic (PV) models, \(N=100\) and \(MFEs=50,000\). Each algorithm was independently run 30 times for each parameter estimation problem to thoroughly evaluate its stability and effectiveness in practical application scenarios. The experiments were conducted in MATLAB R2021b, utilizing a computer featuring an Intel Core i7-9750H processor running at 2.60 GHz and 16 GB of RAM.

For the evaluation of experimental data, three distinct criteria were employed to comprehensively assess the performance of HGMRFO, as defined below:

1) Mean optimization error and standard deviation: The performance of each algorithm is evaluated by analyzing the optimization error generated in each iteration, which measures the difference between the solution obtained by the algorithm and the known global optimum. Each algorithm is executed independently across multiple trials, and the mean and standard deviation (std) of the optimization error are calculated to assess the robustness and reliability of its performance. The algorithm with the lowest mean error value is considered the most accurate, with this value highlighted in bold to indicate the best performance among all compared algorithms.

2) Wilcoxon rank-sum test63: It represents a non-parametric statistical approach employed to determine if a significant difference exists between the proposed algorithm and the comparison algorithm, using a significance level of \(\alpha = 0.05\). In the resulting test outcomes, the symbol “\(+\)” indicates that the proposed algorithm significantly outperforms its competitor, while the symbol “−” signifies that the proposed algorithm performs significantly worse. The symbol “\(\approx\)” implies that there is no significant difference between the two algorithms. Furthermore, “W/T/L” represents the number of times the proposed algorithm wins, ties, and loses against its competitors, providing a comprehensive assessment of the relative performance of the algorithms.

3) Friedman test64: It is also a non-parametric statistical approach that serves to evaluate the performance of all tested algorithms over the entire problem set and generate a comprehensive ranking. In this test, the mean optimization error of each algorithm is used as the test data, with a lower Friedman rank indicating superior algorithm performance. For clarity and ease of comparison, the smallest Friedman rank is highlighted in bold, emphasizing the best-performing algorithm.

Parameter sensitivity analysis

In the HGMRFO, the number of individuals in the personal-best layer is determined by the parameter \(\text {P}_{\text {rate}}\), calculated as \(P_{N} = \text {P}_{\text {rate}} \times N\). To evaluate the impact of \(\text {P}_{\text {rate}}\) on the algorithm’s performance, this study considers 10 possible values: \(\text {P}_{\text {rate}} \in \{0.02, 0.05, 0.1, 0.15, 0.2, 0.3, 0.4, \text {rand}[0.02, 0.15], \text {rand}[0.02, 0.2], \text {rand}[0.02, 0.3]\}\). In the parameter sensitivity analysis, these values are used to regulate the size of the personal-best layer, enabling an assessment of the algorithm’s performance on the 30-dimensional IEEE CEC2017 benchmark function set under different configurations. By comparing the optimization results across different \(\text {P}_{\text {rate}}\) settings, we can gain insight into how this parameter influences HGMRFO’s convergence performance and global search capability.

The number of individuals in the archive layer is controlled by the parameter \(\text {A}_{\text {rate}}\), calculated as \(A_{N} = \text {A}_{\text {rate}} \times N\). To comprehensively evaluate the effect of archive size on the performance of HGMRFO, \(\text {A}_{\text {rate}}\) is set to 7 different values: \(\text {A}_{\text {rate}} \in \{1, 1.5, 2, 2.5, 3, 3.5, 4\}\). In the experiments, these values are used to adjust the size of the archive layer, and the impact on performance is assessed using the IEEE CEC2017 benchmark function set. Proper configuration of the archive layer is crucial for maintaining population diversity and preventing premature convergence. By comparing the results for different \(\text {A}_{\text {rate}}\) settings, we aim to identify the optimal archive size that best balances exploration and exploitation at various stages of the optimization process.

The sensitivity analysis of parameters \(\text {P}_{\text {rate}}\) and \(\text {A}_{\text {rate}}\) aims to determine the optimal parameter configurations that enhance the performance of HGMRFO on complex optimization problems, improve its global search ability, and increase its local exploitation precision, thereby improving the algorithm’s overall solution efficiency and robustness. Tables 1 and 2 present the experimental results of HGMRFO across different parameter settings on the IEEE CEC2017 benchmark functions. Analysis of the W/T/L metrics in these tables indicates that the algorithm demonstrates limited sensitivity to the parameters \(\text {P}_{\text {rate}}\) and \(\text {A}_{\text {rate}}\). However, according to the Friedman test rankings, HGMRFO achieves the best performance with \(\text {P}_{\text {rate}} = \text {rand}[0.02, 0.2]\) and \(\text {A}_{\text {rate}} = 3.5\). This implies that, although the algorithm is generally robust to variations in these parameters, specific configurations can still lead to statistically significant improvements in performance.

Table 1 Statistical results of HGMRFO with \(\text {P}_{\text {rate}}\) on CEC2017 benchmark functions with 30 dimensions.
Table 2 Statistical results of HGMRFO with \(\text {A}_{\text {rate}}\) on CEC2017 benchmark functions with 30 dimensions.

Analysis of hierarchical interactions

To comprehensively evaluate the efficacy of various hierarchical structures within the HGMRFO, five distinct hierarchical schemes, denoted as L1 through L5, were systematically compared as follows:

  • L1: Personal-best layer only

  • L2: Personal-best layer + Population layer

  • L3: Rank-based selective pressure layer + Archive layer

  • L4: Personal-best layer + Population layer + Rank-based selective pressure layer + Archive layer (i.e., HGMRFO)

  • L5: Current best individual + Population layer + Rank-based selective pressure layer + Archive layer

These five schemes were subjected to experimental evaluation on the CEC2017 suite to thoroughly assess their relative performances, as shown in Table 3.

The experimental results unequivocally demonstrate that L4, the complete version of HGMRFO, which integrates the personal-best layer, population layer, rank-based selective pressure layer, and archive layer, consistently outperformed the other hierarchical schemes in terms of optimization accuracy. Specifically, HGMRFO achieved a more optimal balance between global exploration and local exploitation, enhancing its capability to prevent premature convergence and obtain superior overall solutions.

Notably, the comparison between L4 and L1 underscores the critical importance of incorporating additional layers, such as the rank-based selective pressure layer and archive layer, in enhancing the robustness and search diversity of the algorithm. Similarly, comparisons with L5 reveal that leveraging a dynamically updated personal-best layer, rather than merely relying on the current best individual, provides more effective guidance, ultimately leading to enhanced optimization outcomes.

The comparison between L4 and L2 highlights the impact of adding both the rank-based selective pressure layer and the archive layer. While L2 already incorporates the population layer to complement the personal-best layer, L4 further enhances the search process by utilizing rank-based selective pressure to guide the population more effectively towards promising regions and by incorporating an archive layer to maintain search diversity. This combination enhances resilience against early convergence and optimizes the utilization of the solution space.

Furthermore, the comparison between L4 and L3 reveals the critical role of the personal-best layer in conjunction with the rank-based selective pressure layer and archive layer. L3, which lacks the personal-best layer, showed limited guidance efficiency, leading to suboptimal convergence. By incorporating the personal-best layer, L4 substantially enhanced the algorithm’s capacity to fine-tune the exploration process and attain superior-quality solutions, underscoring the importance of integrating all four layers to achieve optimal performance.

Overall, the experimental analysis substantiates that the complete hierarchical configuration of HGMRFO (L4) significantly improves the algorithm’s performance, emphasizing the importance of each layer in facilitating a comprehensive and effective exploration of the solution space.

Table 3 Statistical results of HGMRFO (i.e., L4) and other hierarchical algorithms on CEC2017 functions with 30 dimensions.

Comparison for competitive algorithms

Table 4 Parameter settings of HGMRFO and various algorithms on IEEE CEC2017, IEEE CEC2011, and PV models.
Table 5 Statistical results of HGMRFO and seven algorithms on CEC2017 functions.

To rigorously evaluate the performance of the HGMRFO, comparative experiments were conducted against several state-of-the-art algorithms, including the original MRFO13, the fractional-order Caputo MRFO (FCMRFO)27, chaotic wind driven optimization with fitness distance balance strategy (CFDBWDO)65, spatial information sampling algorithm (SIS)66, reptile search algorithm (RSA)67, the fractional-order flower pollination algorithm (FOFPA)68, and spherical evolution (SE)69, on the IEEE CEC2017 benchmark functions with dimensions of 10, 30, 50, and 100. The statistical results derived from these experiments are presented in Table 5, with the experimental data provided in Tables 8910, and 11, while the specific parameter settings for each algorithm are detailed in Table 4.

In the statistical results presented in Table 5, HGMRFO demonstrated superior performance across various dimensions of the IEEE CEC2017 benchmark functions. Specifically, in the 10-dimensional benchmark, HGMRFO outperformed MRFO, FCMRFO, CFDBWDO, SIS, RSA, FOFPA, and SE on 15, 11, 27, 22, 29, 23, and 13 problems, respectively, achieving the top rank according to the Friedman test. These results indicate that HGMRFO exhibits remarkable performance in low-dimensional optimization tasks.

In the 30-dimensional benchmark, HGMRFO outperformed the aforementioned algorithms on 21, 19, 28, 21, 28, 24, and 17 problems, respectively, with the highest Friedman ranking. This signifies the strong competitiveness of HGMRFO in medium-dimensional problems.

For the 50-dimensional benchmark, the W/T/L (Wins/Ties/Losses) metric demonstrates that HGMRFO performed better than MRFO, FCMRFO, CFDBWDO, SIS, RSA, FOFPA, and SE on 17, 15, 27, 18, 29, 26, and 17 problems, respectively. It also achieved the smallest Friedman rank, indicating that HGMRFO excels in high-dimensional optimization scenarios.

In the 100-dimensional benchmark, HGMRFO outperformed the competing algorithms on 17, 17, 25, 18, 28, 24, and 18 problems, respectively, while maintaining the smallest rank value, thereby showcasing its effectiveness in ultra-high-dimensional optimization tasks.

The statistical analysis across different dimensions conclusively demonstrates that HGMRFO consistently achieves superior performance, affirming its robustness and high efficiency in handling complex optimization problems of varying dimensionality.

In Table 8, HGMRFO achieved the best solution on 6 out of 10-dimensional functions, while SE obtained the best solution on 12 functions, ranking first in terms of the number of best solutions, with HGMRFO ranking second. Therefore, for low-dimensional functions, the proposed HGMRFO still requires further improvement in future research.

In Tables 910, and 11, HGMRFO achieved the best solution on 15 out of 30-dimensional functions, 10 out of 50-dimensional functions, and 11 out of 100-dimensional functions, respectively, ranking first in terms of the number of best solutions obtained. Thus, HGMRFO demonstrates the best performance on 30-dimensional, 50-dimensional, and 100-dimensional functions.

Fig. 4 shows the convergence curves of the average optimization errors for HGMRFO and other competing algorithms on the F8 and F11 functions across 10-dimensional, 30-dimensional, 50-dimensional, and 100-dimensional spaces. As can be seen from the figure, HGMRFO demonstrates rapid convergence and consistently finds the optimal solution in all dimensions. Compared to the original MRFO algorithm, the figure clearly illustrates the effectiveness of the hierarchical guidance mechanism proposed in this paper.

The computational complexity of each iteration in the HGMRFO algorithm is primarily determined by the following components:

  1. (1)

    Initialization: \(O(N \times D)\).

  2. (2)

    Fitness evaluation: \(O(N)\).

  3. (3)

    Best individual update: \(O(N)\).

  4. (4)

    Position update: \(O(N \times D)\).

  5. (5)

    Boundary detection: \(O(N \times D)\).

  6. (6)

    Archive update: \(O(N \times D)\).

  7. (7)

    Probability calculation and selection: \(O(N)\).

Therefore, the overall computational complexity for a single iteration is: \(O(N \times D)\). If the algorithm runs for T iterations, the total computational complexity becomes: \(O(T \times N \times D)\).

Fig. 3
figure 3

Bar chart depicting CPU running time.

Furthermore, Fig. 3 illustrates a bar chart that shows the computational running times for each algorithm tested on the IEEE CEC2017 benchmark set with dimensions of 10, 30, 50, and 100. As illustrated in Fig. 3, both HGMRFO and MRFO exhibit the shortest CPU running times across all dimensions, with computational complexities that are nearly identical. These results suggest that the introduction of the hierarchical guidance structure in HGMRFO imposes negligible computational overhead on MRFO, thus affirming the scalability and efficiency of HGMRFO. Additionally, the low computational cost of HGMRFO highlights its potential for application in high-dimensional and intricate engineering optimization problems.

Table 6 Analysis and comparative performance of HGMRFO and eight competitive algorithms on IEEE CEC2011 real-world optimization problems.
Table 7 Analysis and comparative performance of HGMRFO and eight competitive algorithms on IEEE CEC2011 real-world optimization problems (continue).
Fig. 4
figure 4

Convergence curves of average optimization error performance by eight algorithms on F8 and F11 (dimensions 10, 30, 50 and 100).

Fig. 5
figure 5

Photovoltaic equivalent circuit: (a) single diode model, (b) double diode model, (c) triple diode model, and (d) photovoltaic module model.

Table 8 Experimental and statistical results of HGMRFO and seven competitive algorithms on IEEE CEC2017 benchmark functions with 10 dimensions, where HGMRFO is the main algorithm in statistical results (W/T/L).
Table 9 Experimental and statistical results of HGMRFO and seven competitive algorithms on IEEE CEC2017 benchmark functions with 30 dimensions, where HGMRFO is the main algorithm in statistical results (W/T/L).
Table 10 Experimental and statistical results of HGMRFO and seven competitive algorithms on IEEE CEC2017 benchmark functions with 50 dimensions, where HGMRFO is the main algorithm in statistical results (W/T/L).
Table 11 Experimental and statistical results of HGMRFO and seven competitive algorithms on IEEE CEC2017 benchmark functions with 100 dimensions, where HGMRFO is the main algorithm in statistical results (W/T/L).

Real-world optimization problems

To comprehensively assess the practicality of the HGMRFO, we conducted extensive simulation experiments comparing HGMRFO with eight state-of-the-art algorithms on real-world optimization problems from the IEEE CEC2011 benchmark suite. The algorithms considered in this comparison include: MRFO13, FCMRFO27, HMRFO12, disGSA70, AHA71, SIS66, DGSA72, and XPSO73. Among them, HMRFO is an hierarchical MRFO with weighted fitness-distance balance selection. disGSA is an adaptive position-guided GSA. AHA denotes the artificial hummingbird algorithm. DGSA is a GSA with hierarchy and distributed framework. XPSO is an expanded PSO based on multi-exemplar and forgetting ability. The specific parameter settings for each algorithm are detailed in Table 4. The experimental data and statistical results are presented in Tables 6 and 7.

From the results summarized in Tables 6 and 7, it can be observed that HGMRFO achieved the best mean values on eight distinct problems, outperforming the competing algorithms by achieving the highest number of optimal solutions. Furthermore, the W/T/L analysis indicates that HGMRFO outperformed MRFO, FCMRFO, HMRFO, disGSA, AHA, SIS, DGSA, and XPSO on 7, 8, 7, 12, 14, 16, 11, and 11 problems, respectively. These results highlight the effectiveness and practical applicability of HGMRFO in addressing real-world optimization challenges.

The demonstrated superior performance, combined with the low computational overhead of HGMRFO, motivated us to apply this algorithm to a practical engineering problem-parameter estimation in solar photovoltaic models.

Comparison on parameter estimation of photovoltaic model

Table 12 Sources of experimental true current-voltage values, measurement conditions, and number of experimental data sets for six models.
Table 13 Parameters ranges for PV models.
Table 14 Analysis and comparative performance of HGMRFO and other competitive algorithms on the PV models.
Fig. 6
figure 6

Box-and-whisker diagrams of errors obtained by eight algorithms on six PV models.

To thoroughly analyze the nonlinear output characteristics of photovoltaic (PV) systems, the adoption of accurate mathematical models for PV cells is crucial. The output characteristics of PV cells are significantly influenced by various factors, including temperature and internal as well as external circuit parameters. Consequently, the development of precise mathematical models is essential for predicting system performance, optimizing system design, and improving energy conversion efficiency. To meet diverse research and application requirements, various PV models have been developed, spanning from simple to more complex structures. These models include the single diode model (SDM)74, the double diode model (DDM)75, the triple diode model (TDM)76, and more complex PV module models (MM)77. The structures of these models are shown in Fig. 5, with a detailed description provided below.

The output current \(I_L\) of the SDM can be determined using the equation below:

$$\begin{aligned} \begin{aligned} I_L = I_{ph} - I_d - I_{sh} = I_{ph} - I_{sd} \left[ \exp \left( \frac{q(V_L + R_s I_L)}{akT}\right) - 1 \right] - \frac{V_L + R_s I_L}{R_{sh}} \end{aligned} \end{aligned}$$
(10)

where, \(I_{ph}\) denotes the photo-generated current, which is the current generated due to illumination, \(I_d\) is the diode current, \(I_{sh}\) is the current through the shunt resistance, \(I_{sd}\) denotes the reverse saturation current of the diode, \(V_L\) is the output voltage, \(R_s\) is the series resistance, \(R_{sh}\) is the shunt resistance, \(a\) is the diode ideality factor, \(q\) is the electron charge (\(1.60217646 \times 10^{-19}\) C), \(k\) is the Boltzmann constant (\(1.3806503 \times 10^{-23}\) J/K), and \(T\) is the cell temperature in Kelvin. In the SDM, five parameters need to be estimated: \(I_{ph}\), \(I_{sd}\), \(R_s\), \(R_{sh}\), and \(a\).

The output current \(I_L\) of the DDM can be determined using the equation below:

$$\begin{aligned} \begin{aligned} I_L = I_{\text {ph}} - I_{d1} - I_{d2} - I_{\text {sh}} = I_{\text {ph}} - I_{\text {sd1}} \left[ \exp \left( \frac{q(V_L + R_s I_L)}{a_1 k T}\right) - 1 \right] - I_{\text {sd2}} \left[ \exp \left( \frac{q(V_L + R_s I_L)}{a_2 k T}\right) - 1 \right] - \frac{V_L + R_s I_L}{R_{\text {sh}}} \end{aligned} \end{aligned}$$
(11)

where \(I_{\text {sd1}}\) is the diffusion current, \(I_{\text {sd2}}\) is the saturation current, and \(a_1\) and \(a_2\) are the ideality factors of the two diodes. In the DDM, seven parameters need to be estimated: \(I_{ph}\), \(I_{sd1}\), \(I_{sd2}\), \(R_s\), \(R_{sh}\), \(a_1\) and \(a_2\).

The output current \(I_{\text {L}}\) of the TDM can be determined using the equation below:

$$\begin{aligned} \begin{aligned} I_L = I_{\text {ph}} - I_{d1} - I_{d2} - I_{d3} - I_{\text {sh}}&= I_{\text {ph}} - I_{\text {sd1}} \left[ \exp \left( \frac{q(V_L + R_s I_L)}{a_1 k T}\right) - 1 \right] - I_{\text {sd2}} \left[ \exp \left( \frac{q(V_L + R_s I_L)}{a_2 k T}\right) - 1 \right] \\ &\quad - I_{\text {sd3}} \left[ \exp \left( \frac{q(V_L + R_s I_L)}{a_3 k T}\right) - 1 \right] - \frac{V_L + R_s I_L}{R_{\text {sh}}} \end{aligned} \end{aligned}$$
(12)

where \(I_{\text {sd3}}\) is leakage current. In the TDM, nine parameters need to be estimated: \(I_{ph}\), \(I_{sd1}\), \(I_{sd2}\), \(I_{sd3}\), \(R_s\), \(R_{sh}\), \(a_1\), \(a_2\) and \(a_3\).

The PV current of MM can be determined by the following equation:

$$\begin{aligned} \begin{aligned} \frac{I_L}{N_p} = I_{\text {ph}} - I_{\text {sd}} \left[ \exp \left( \frac{q\left( \frac{V_L}{N_s} + R_s \frac{I_L}{N_p}\right) }{a k T}\right) - 1 \right] - \frac{\frac{V_L}{N_s} + R_s \frac{I_L}{N_p}}{R_{\text {sh}}} \end{aligned} \end{aligned}$$
(13)

where \(N_{\text {p}}\) is the number of solar cells connected in parallel, and \(N_{\text {s}}\) is the number of solar cells connected in series. In the MM, five parameters need to be estimated: \(I_{ph}\), \(I_{sd}\), \(R_s\), \(R_{sh}\), and \(a\). The objective function for parameter estimation is achieved by minimizing the error between the measured current data and the experimental current data2.

To comprehensively evaluate the performance of the proposed HGMRFO, six groups of experiments were designed to encompass different numbers of series diodes (Ns) and parallel diodes (Np). For the SDM, DDM, and TDM, experimental data were obtained from a commercial silicon solar cell produced by R.T.C. France, with 26 sets of current-voltage data measured at a temperature of \({33}^\circ\)C78. For the photovoltaic MM, the experimental data were sourced from Photowatt-PWP201, STP6-120/36, and STM6-40/36 PV modules, under varying temperature and irradiance conditions78,79. Table 12 presents a comprehensive overview of the true value sources, measurement conditions, and the number of experimental data points for each model. Additionally, Table 13 presents the lower bounds (LB) and upper bounds (UB) for parameter extraction in each model80,81,82. In these six experimental setups, the HGMRFO was systematically compared against seven other advanced optimization algorithms. These algorithms include the original MRFO13, FCMRFO27, HMRFO12, water flow optimizer (WFO)83, fractional-order WFO (FOWFO)84, wind driven optimization (WDO)85, and grey wolf optimizer (GWO)86. The configuration of parameters for all algorithms is summarized in Table 4. The experimental data and statistical results are presented in Table 14.

As presented in Table 14, HGMRFO achieved the best mean values across all photovoltaic (PV) models. Specifically, HGMRFO outperformed MRFO, FCMRFO, HMRFO, FOWFO, WFO, WDO, and GWO in 6, 6, 5, 6, 6, 6, and 6 PV models, respectively. According to the Friedman test ranks, HGMRFO is ranked first, indicating that HGMRFO outperforms other competing algorithms. These results clearly demonstrate the superior performance of HGMRFO in addressing the parameter estimation challenges associated with solar PV models, underscoring its efficacy and robustness in tackling such complex optimization problems.

Figure 6 presents the box-and-whisker diagrams of errors for HGMRFO and seven other algorithms across six photovoltaic models. The median, indicated by the red horizontal line, represents the average performance of the algorithms. From the figure, it can be seen that HGMRFO has the lowest median on the SDM, DDM, TDM, STP6-120/36, and STM6-40/36 models, indicating that HGMRFO performs best on these five models. However, on the Photowatt-PWP201 model, HGMRFO slightly underperforms compared to HMRFO, although its overall performance remains strong. The box-and-whisker diagrams demonstrate that HGMRFO exhibits excellent robustness and accuracy on PV models.

Conclusions

In this study, we address the limitations of the MRFO during the somersault foraging phase, specifically the fixed parameter setting and the exclusive focus on the current best individual during the somersault search. To overcome these limitations, we introduce an adaptive somersault factor and a hierarchical guidance mechanism, enhancing the information exchange among population individuals. Through the synergistic effect of a multi-layered structure, individuals are effectively guided towards the global optimal solution, thereby proposing an enhanced algorithm named hierarchical guided manta ray foraging optimization (HGMRFO). During the experimental phase, we initially identified the optimal parameter configurations for HGMRFO and conducted a systematic analysis of how different hierarchical structures influence the algorithm’s performance. Extensive experiments on the IEEE CEC2017 benchmark functions and IEEE CEC2011 real-world optimization problems confirmed the significant advantages of HGMRFO in terms of optimization performance and practical applicability, demonstrating its lower computational cost and superior efficiency. Finally, HGMRFO was utilized for parameter estimation in solar photovoltaic models with multimodal characteristics, yielding excellent experimental outcomes and further demonstrating the potential and efficacy of this algorithm in addressing complex real-world problems.

In future research, we intend to focus on the following five areas:

  1. 1.

    Further enhancement for low-dimensional optimization problems: Based on the optimization results on the CEC2017 benchmark functions with 10 dimensions, particularly in comparison with SE, it is evident that HGMRFO requires further refinement for better performance in low-dimensional optimization tasks.

  2. 2.

    Improvement in selection methods for HGMRFO: In the boxplot analysis of the PV models, HGMRFO performed weaker than HMRFO on the Photowatt-PWP201 model. In the future, we could draw inspiration from the fitness-distance balance selection strategy in HMRFO12 to improve the selection method in HGMRFO.

  3. 3.

    Extension of hierarchical guidance mechanism: We will explore the integration of the hierarchical guidance mechanism into other optimization algorithms to further enhance their performance and robustness.

  4. 4.

    Investigation of alternative population structures: We aim to study various alternative population structures in an effort to identify more sophisticated configurations that could improve the structural characteristics of the original algorithm, thereby contributing to overall optimization efficiency.

  5. 5.

    Application of HGMRFO to additional real-world problems: We plan to apply HGMRFO to other complex real-world problems, such as the protein structure prediction87,88 and dendritic neural model89,90,91, to validate its versatility and effectiveness in diverse domains.