Introduction

Drones, due to their high flexibility, safety, robustness, and reliability, are widely used in disaster prevention and mitigation, traffic regulation, search and rescue, and other fields. Path planning refers to the process of determining the optimal or near-optimal route for a drone from its starting point to its destination 1,2. Ensuring mission execution and environmental safety is crucial. A path that is highly safe, feasible, computationally efficient, and cost-effective can significantly enhance the completion efficiency of drone missions. The planned path must meet specific criteria to ensure optimality 3. For instance, in applications such as aerial photography, surveying, and surface inspection, the path should be as short as possible to minimize time and fuel consumption 4,5. However, for other tasks such as dynamic target search or surveillance and rescue, the path needs to be formulated based on different criteria, potentially requiring the maximization of detection probability or the minimization of flight time. Additionally, the path planning must consider environmental safety and drone performance constraints to ensure collision avoidance and compliance with requirements related to flight time, altitude, and fuel consumption 6,7,8. Therefore, achieving safe path planning that enables collision-free and feasible movements remains a challenging task.

The core objective of drone path planning is to generate an optimal path from the starting point to the destination while satisfying various constraints, typically involving the comprehensive optimization of metrics such as path length, threat avoidance, energy consumption, and trajectory smoothness 9,10. In recent years, researchers have proposed a variety of path planning methods, which can be broadly categorized into traditional methods and intelligent optimization methods. Traditional methods, such as the A* algorithm 11,12, RRT13,14, and Dijkstra’s algorithm 15,16,17, involve constructing a cost graph for local search. While they can quickly identify feasible paths, they are prone to falling into local optima in high-dimensional complex environments and struggle to balance multi-objective optimization requirements. In contrast, intelligent optimization algorithms achieve global optimization by simulating biological group behaviors or physical processes. They demonstrate significant advantages in handling complex constraints and multi-objective problems, making them a current research focus 18,19.

It is well-known that optimizing the flight path of UAVs typically requires optimization algorithms. These optimization methods generally fall into deterministic mathematical programming methods and stochastic metaheuristic algorithms 20. However, deterministic mathematical programming methods often struggle with nonlinear spaces 21,22. Metaheuristic algorithms can solve complex problems that were previously considered difficult. They generate optimal or near-optimal solutions while reducing computational complexity, making them a good alternative when deterministic algorithms produce inefficient solutions. Path planning is a challenging problem for deterministic techniques, prompting the use of many heuristic-based techniques to address this new problem. Generally, metaheuristic algorithms can be categorized into four types 23,24: Evolutionary Algorithms (EA), Physics-based Algorithms (PhA), Human-based (HB) Algorithms, and Swarm Intelligence (SI) Algorithms. Evolutionary Algorithms typically generate better new populations through combinations and mutations among individuals from previous generations. Examples include Genetic Algorithms (GA) inspired by Darwin’s theory of evolution 25, Differential Evolution (DE) based on natural selection and reproduction processes 26, Genetic Programming (GP), and Evolution Strategies (ES) inspired by biological evolution processes 27. Physics-based methods use rules derived from various physical phenomena in nature to find solutions. Examples include Central Force Optimization 28, Simulated Annealing (SA) based on the annealing process in metallurgy 29, Galaxy-based Optimization (GSO) inspired by galaxy movements 30, and Gravitational Search Algorithm (GSA) based on Newton’s laws of gravity and motion 31. Human-based (HB) algorithms simulate natural human behaviors to solve optimization problems. Examples include Teaching-Learning-Based Optimization (TLBO) which models the influence of a teacher on learners’ outputs 32. Swarm Intelligence (SI) algorithms are inspired by the collective behavior of animals. Examples include Particle Swarm Optimization (PSO) based on the foraging behavior of birds and fish 33, Ant Colony Optimization (ACO) inspired by the social foraging behavior of ants 34, Grey Wolf Optimizer (GWO) based on the social hierarchy and hunting behavior of grey wolves 35, Whale Optimization Algorithm (WOA) inspired by the hunting strategies of humpback whales, which involve search for prey, encircling, and spiral updating 36, Black Widow Optimization (BWO) inspired by the unique reproductive behavior of black widow spiders 37, Red-Billed Blue Magpie Optimizer (RBMO) inspired by the behaviors of searching, chasing, attacking prey, and food storing of red-billed blue magpies 38, Quantum Aviator Navigation Algorithm (QANA) inspired by the extraordinary navigational precision of migratory birds 39, Golden Jackal Optimization (GJO) based on the cooperative hunting behavior of golden jackals 40.

However, stochastic metaheuristic algorithms also have inherent drawbacks. David H. Wolpert and William G. Macready 41 introduced the No Free Lunch (NFL) theorem in 1997, logically proving that no metaheuristic algorithm can best solve all optimization problems. In other words, while intelligent algorithms can achieve desired results in specific optimization problems, they may perform poorly in others. Consequently, researchers are attempting to integrate different intelligent algorithms into UAV path planning to find better solutions 42,43,44,45. For instance, Fu et al. 38 proposed the Red-Billed Blue Magpie Optimizer (RBMO), inspired by the behaviors of red-billed blue magpies such as searching, chasing, attacking prey, and food storing, for 2D and 3D UAV path planning problems. Huang et al. developed a PSO algorithm based on cylindrical vectors with wave and sigmoid functions for UAV path planning 46. Qu et al. combined the Simplified Grey Wolf Optimizer (SGWO) and Modified Symbiotic Organisms Search (MSOS) to propose a novel hybrid Grey Wolf Optimizer for UAV path planning 47. Manh et al. introduced a new algorithm based on Spherical Vector Particle Swarm Optimization (SPSO) to address UAV path planning in complex environments with multiple threats 48. Zhang et al. 3 proposed a new multi-objective evolutionary algorithm with a dual constraint handling mechanism for multi-path UAV planning. Li et al. 49 developed a Fermat Point-based Group Particle Swarm Optimization (FP-GPSO) algorithm for UAV path planning, while Yu et al. 50devised a hybrid GWO and Differential Evolution (HGWODE) algorithm to solve UAV path planning problems.

These efforts reflect the ongoing research to overcome the limitations of single metaheuristic approaches and enhance the effectiveness of UAV path planning through hybrid and innovative algorithmic integrations.

The Secretary Bird Optimization Algorithm (SBOA), recently proposed by Fu et al. 51, is a swarm intelligence algorithm inspired by the hunting and survival behaviors of secretary birds. It is characterized by its simple structure and fast convergence. However, in UAV path planning within complex environments, SBOA faces issues such as imbalance between exploration and exploitation, slow convergence speed, tendency to get trapped in local optima, and low convergence accuracy. To address these problems, this paper proposes a Multi-Strategy Secretary Bird Optimization Algorithm (MSBOA) based on pooling mechanisms, dynamic fitness distance balancing techniques, and centroid reverse learning. This algorithm is applied to UAV path planning in complex environments. The main contributions of this paper are as follows:

  1. (1)

    The successful integration of pooling mechanisms, dynamic fitness distance balancing techniques, Preferential selecting search strategy, and centroid reverse learning into SBOA, enhancing its overall performance.

  2. (2)

    Performance validation of MSBOA through comparison with four classical algorithms and four recent algorithms on CEC2017 benchmark test functions.

  3. (3)

    Successful application of MSBOA to UAV path planning in eight different scenarios generated from two real digital elevation model (DEM) maps with varying complexities.

The remainder of this paper is organized as follows: Section "UAV mathematical model" introduces the UAV mathematical model; Section "Secretary bird optimization algorithm" briefly explains the principles of the SBOA; Section "Multi-strategy secretary bird optimization algorithm" provides a detailed description of the four strategy improvements in the MSBOA; Section "Experimental results and analysis" presents the experiments on CEC2017 benchmark test functions and the UAV path planning simulations; Section "Summary and Prospect" concludes the study and discusses future work.

UAV mathematical model

3. In complex terrains, a UAV follows the path planning results to navigate from the starting point to the target point. During its flight, the UAV may encounter various obstacles such as terrain features, fire threats, and radar scanning areas, in addition to constraints like fuel consumption, maximum climb rate, and maximum turning capability. To ensure the safe and coordinated flight of the UAV, the path planning algorithm must determine the optimal path that connects the start and end points 48,52.

Assuming the drone maintains a predetermined flight speed, the path planning problem is thereby simplified to a static piecewise planning problem. Considering the operational requirements of the drone, the cost function for drone path planning is defined by computing the cost of distance traveled, altitude, threats encountered, and smoothness. Mathematically, it can be represented as follows:

$$\begin{array}{c}{F}_{4}\left({X}_{i}\right)=\sum_{k=1}^{4} {b}_{k}{F}_{k}({X}_{i})\end{array}$$
(1)

Here, \({b}_{k}\) represents the weighting coefficients, with this paper adopting \({b}_{1}={b}_{2}=1,{b}_{3}=10,{b}_{4}=1\), \({F}_{1}\left({X}_{i}\right)\) through \({F}_{4}\left({X}_{i}\right)\) respectively denote the cost functions associated with path length, threat encountered, altitude, and smoothness. The decision variables are denoted by \({X}_{i}\),comprising a list of \(l\) waypoints represented as \({P}_{ij}=({x}_{ij},{y}_{ij},{z}_{ij})\), where \({P}_{ij}\) belongs to the operational space \(O\) of the drone.

Path distance cost

In order to ensure efficient operation of the drone, path planning needs to be optimized according to specific criteria tailored to different application scenarios. Given our primary focus on aerial photography, mapping, and surface inspection, we opt to minimize path length as our objective. The cost function for path length, \({F}_{1}\) is determined by the Euclidean distance between adjacent nodes and is calculated using Equation (2).

$$\begin{array}{c}{F}_{1}\left({X}_{i}\right)=\sum_{j=1}^{l-1} \Vert \overrightarrow{{P}_{ij}{P}_{i,j+1}}\Vert .\end{array}$$
(2)

In this equation, \({P}_{ij}\) and \({P}_{i,j+1}\) represent adjacent nodes, and the sum of \(n-1\) segments between \(n\) nodes constitutes the cost of path length.

Safety and feasibility costs

In addition to optimizing for minimum path length, path planning must also ensure the safe operation of the drone, which may encounter restricted zones known as threat areas (such as radar detection, anti-aircraft equipment attacks, or weather threats) 53. Considering the complexity of threat modeling and the difficulty in obtaining real data, this paper abstracts the threat environment by representing threat areas as cylindrical regions with a fixed radius. The radius of these cylindrical regions defines the effective threat range, and the purpose of defining threat areas is to ensure the drone’s operational safety by enabling it to successfully avoid these regions and complete its flight mission. Let \(K\) denote the set of all threats, assuming each threat is contained within a cylindrical region with its projection center coordinates denoted by \({C}_{k}\) and radius \({R}_{k}\). Figure 1 illustrates the safety constraints. During drone flight, it is essential to avoid tall obstacles on the ground. The safety and feasibility cost of the path \({F}_{2}\) is calculated using Equations (3) and (4).

Fig. 1
figure 1

Safety constraint cost.

$$\begin{array}{c}{F}_{2}({X}_{i})=\sum_{j=1}^{l-1} \sum_{k=1}^{K} {T}_{k}(\overrightarrow{{P}_{ij}{P}_{i,j+1}})\end{array}$$
(3)
$$\begin{array}{c}{T}_{k}(\overrightarrow{{P}_{ij}{P}_{i,j+1}})=\left\{\begin{array}{cc}0,& \text{if }{d}_{k}>S+D+{R}_{k},\\ \left(S+D+{R}_{k}\right)-{d}_{k},& \text{if }D+{R}_{k}<{d}_{k}\le S+D+{R}_{k},\\ \infty ,& \text{if }{d}_{k}\le D+{R}_{k}.\end{array}\right.\end{array}$$
(4)

According to Equation (4) and as depicted in Fig. 1, when the distance \({d}_{k}\) is less than the sum of the danger zone radius \(\left(D+{R}_{k}\right)\), the cost \({F}_{2}\) becomes infinite, indicating that the drone must maintain a safe distance from the danger zone during flight. Within the safety buffer zone, the cost increases as the drone gets closer to obstacles. When the distance \({d}_{k}\) exceeds \(\left(S+D+{R}_{k}\right)\), the safety cost is zero, indicating the most ideal path for the drone.

Flight altitude cost

Subject to natural conditions and application-specific constraints, the altitude of drone flight is restricted. For instance, in tasks such as measurement and inspection, the altitude of flight is limited due to specific resolution and field of view requirements of cameras and inspection equipment 54. As illustrated in Fig. 2,the drone’s flight is constrained by a minimum altitude \({h}_{min}\) and a maximum altitude \({h}_{max}\). The cost associated with the altitude related to waypoint \({P}_{ij}\) is given by:

Fig. 2
figure 2

Highly constrained cost.

$$\begin{array}{c}{H}_{ij}=\left\{\begin{array}{c}|{h}_{ij}-\frac{\left({h}_{max}+{h}_{min}\right)}{2}|,{\text{i}}{\text{f}}{ h}_{min}\le {h}_{ij}\le {h}_{max},\\ \infty ,{\text{o}}{\text{t}}{\text{h}}{\text{e}}{\text{r}}{\text{w}}{\text{i}}{\text{s}}{\text{e}}.\end{array}\right.\end{array}$$
(5)

In Equation (5), \({h}_{ij}\) represents the flight altitude relative to the ground, as depicted in Fig. 2. It is evident that \({H}_{ij}\) maintains an average altitude and penalizes values that exceed the specified range. The summation of \({H}_{ij}\) for all waypoints yields the altitude cost \({F}_{3}\), as shown in Equation (6).

$$\begin{array}{c}{F}_{3}\left({X}_{i}\right)=\sum_{j=1}^{l} {H}_{ij}.\end{array}$$
(6)

Path smoothing cost

The primary control parameters for the drone’s flight angles include the horizontal heading angle and the vertical pitch angle. Both of these variable parameters must adhere to the actual angle constraints of the drone; otherwise, the trajectory planning model cannot generate feasible flight paths. As depicted in Fig. 3, the turning angle \({\phi }_{ij}\) is the angle in the horizontal plane \({O}_{xy}\) between two consecutive path segments \(\overrightarrow {{P_{ij}{\prime} P_{i,j + 1}{\prime} }}\) and \(\overrightarrow {{P_{i,j + 1}{\prime} P_{i,j + 2}{\prime} }}\).Let \(\vec{k}\) be the unit vector in the z-axis direction, and the projection vector is calculated using Equation (7) 54.

$$\begin{array}{*{20}c} {\overrightarrow {{P_{ij}{\prime} P_{i,j + 1}{\prime} }} = \vec{k} \times \left( {\overrightarrow {{P_{ij} P_{i,j + 1} }} \times \vec{k}} \right).} \\ \end{array}$$
(7)
Fig. 3
figure 3

Turning Angle and climbing Angle constraints.

Therefore, the formula for the turning angle can be obtained as shown in Equation (8).

$$\begin{array}{*{20}c} {\phi_{ij} = \arctan \left( {\frac{{\overrightarrow {{P_{ij}{\prime} P_{i,j + 1}{\prime} }} \times \overrightarrow {{P_{i,j + 1}{\prime} P_{i,j + 2}{\prime} }} }}{{\overrightarrow {{P_{ij}{\prime} P_{i,j + 1}{\prime} }} .\overrightarrow {{P_{i,j + 1}{\prime} P_{i,j + 2}{\prime} }} }}} \right),0 < \phi_{ij} < 30^\circ .} \\ \end{array}$$
(8)

\({\psi }_{ij}\) represents the angle between the projection segment \(\left( {\overrightarrow {{P_{ij}{\prime} P_{i,j + 1}{\prime} }} } \right)\) and the path segment \(\left( {\overrightarrow {{P_{ij}{\prime} P_{i,j + 1}{\prime} }} } \right)\) on the horizontal plane. It is calculated using Equation (9) 55.

$$\begin{array}{*{20}c} {\psi_{ij} = \arctan \left( {\frac{{z_{i,j + 1} - z_{ij} }}{{\overrightarrow {{P_{ij}{\prime} P_{i,j + 1}{\prime} }} }}} \right),0 < \psi_{ij} < 15^\circ .} \\ \end{array}$$
(9)

Finally, the smoothness cost \({F}_{4}\) is calculated using Equation (10).

$$\begin{array}{c}{F}_{4}\left({X}_{i}\right)={\varphi }_{1}\sum_{j=1}^{l-2} {\phi }_{ij}+{\varphi }_{2}\sum_{j=1}^{l-1} \mid {\psi }_{ij}-{\psi }_{i,j-1}\mid .\end{array}$$
(10)

where \({\varphi }_{1}\) and \({\varphi }_{2}\) are the penalty coefficients for the turning angle and the climbing angle, respectively, and are taken as 0.5 and 0.7, respectively55.

Secretary bird optimization algorithm

Initialization

The SBOA (Secretary Bird Optimization Algorithm) method belongs to the category of population-based metaheuristic methods, akin to other heuristic algorithms. Like other heuristic algorithms, the SBOA algorithm also randomly generates a set of candidate solutions within the search space.

$$\begin{array}{c}X={\left[\begin{array}{cc}\begin{array}{ccc}{x}_{\text{1,1}}& {x}_{\text{1,2}}& \cdots \\ {x}_{\text{2,1}}& {x}_{\text{2,2}}& \cdots \\ \vdots & \vdots & \ddots \end{array}& \begin{array}{ccc}{x}_{1,j}& \cdots & {x}_{1,Dim}\\ {x}_{2,j}& \cdots & {x}_{2,Dim}\\ \vdots & \ddots & \vdots \end{array}\\ \begin{array}{ccc}{x}_{i,1}& {x}_{i,2}& \cdots \\ \vdots & \vdots & \ddots \\ {x}_{N,1}& {x}_{N,2}& \cdots \end{array}& \begin{array}{ccc}{x}_{i,\text{j}}& \cdots & {x}_{i,Dim}\\ \vdots & \ddots & \vdots \\ {x}_{N,j}& \cdots & {x}_{N,Dim}\end{array}\end{array}\right]}_{N\times Dim }\end{array}$$
(11)

Where \(X\) represents the population of secretary birds, \({X}_{i}\) denotes the position of the \({i}^{th}\) secretary bird, \({x}_{i,j}\) represents the positional information of the \({j}^{th}\) problem variable of the \({i}^{th}\) secretary bird, \(N\) represents the population size, and \(Dim\) represents the dimensionality of the problem variables.

The initial positional information of the secretary birds is randomly determined based on Equation (12):

$$\begin{array}{c}{x}_{i,j}=\left({upper}_{j}-{lower}_{j}\right)\times {r}_{1}+{lower}_{j}\end{array}$$
(12)

In this equation, \({x}_{i,j}\) represents the initial value of the \({j}^{th}\) decision variable for the \({i}^{th}\) candidate solution;

\({upper}_{j}\) and \({lower}_{j}\) are the maximum and minimum bounds; \({r}_{1}\) is a random number in the range (0, 1).

Hunting behavior (exploration)

The hunting behavior of secretary birds is typically divided into three stages: searching for prey \(\left({P}_{1}\right)\), consuming prey \(\left({P}_{2}\right)\), and attacking prey \(\left({P}_{3}\right)\). During the searching for prey stage, the secretary bird seeks potential prey. Once the prey is identified, it enters the consuming prey stage, where it exhausts the prey’s stamina. With keen judgment of its actions towards the prey, the bird hovers, jumps, and provokes near the snake, gradually depleting the prey’s stamina until it launches an attack. This process is modeled using Equations (13) and (14).

$$\begin{array}{c}{x}_{i,j}^{new1}=\left\{\begin{array}{c}{P}_{1}:{ x}_{i,j}+{r}_{2}\times {(x}_{r\_1}-{x}_{{r}_{2}}), if t<\frac{1}{3}T\\ {P}_{2}:{x}_{best}+exp\left({\left(\frac{t}{T}\right)}^{4}\right)\times \left(RB-0.5\right)\times \left({x}_{best}-{x}_{i,j}\right), if \frac{1}{3}T<t<\frac{2}{3}T\\ {P}_{3}:{ x}_{best}+{\left(1-\frac{t}{T}\right)}^{\left(2\times \frac{t}{T}\right)}\times {x}_{i,j}\times RL, else\end{array}\right.\end{array}$$
(13)
$$\begin{array}{c}{X}_{i}=\left\{\begin{array}{c}{ X}_{i}^{new1}, if {F}_{i}^{new1}<{F}_{i}\\ {X}_{i } ,else\end{array}\right.\end{array}$$
(14)

Here, \(t\) represents the current iteration number, \(T\) represents the maximum iteration number, \({x}_{i,j}^{new1}\) represents the new state of the \({i}^{th}\) secretary bird in the first stage, \({x}_{r1}\) and \({x}_{r2}\) are randomly selected candidate solutions for the first stage iteration, \({r}_{2}\) represents \(1\times Dim\) array randomly generated from the interval [0,1], \({x}_{i,j}^{new1}\) represents the positional information for its \({j}^{th}\) dimension, \({F}_{i}^{new1}\) represents its objective function fitness value. \(RB\) represents a \(1\times Dim\) array randomly generated from a standard normal distribution (with mean 0 and standard deviation 1), \({x}_{best}\) represents the best solution obtained so far. \(RL\) represents the Levy flight function, calculated using Equation (15).

$$\begin{array}{c}\left\{\begin{array}{c}RL=0.5\times Levy\left(Dim\right)\\ Levy\left(Dim\right)=0.01\times \frac{u\times \sigma }{|v{|}^{\frac{1}{\eta }}}\\ \sigma ={\left(\frac{\Gamma \left(1+\eta \right)\times \mathit{sin}\left(\frac{\pi \times \eta }{2}\right)}{\Gamma \left(\frac{1+\eta }{2}\right)\times \eta \times 2\left(\frac{\eta -1}{2}\right)}\right)}^{\frac{1}{\eta }}\end{array}\right.\end{array}$$
(15)

In this equation, \(\eta\) is a fixed constant equal to 1.5. \(u\) and \(v\) are random numbers within the interval [0, 1], \(\Gamma\) represents the gamma function, and \(\eta\) also has a value of 1.5.

Escape strategy (exploitation)

Secretary birds may encounter attacks from predators or attempts to steal their food. They are highly intelligent and often employ evasion strategies to protect themselves or their food. These strategies primarily involve two approaches: fleeing by flying or running \(\left({S}_{1}\right)\) and camouflage using the colors or structures in the environment \(\left({S}_{2}\right)\), making it harder for predators to detect them. This process is modeled using Equations (16) and (17).

$$\begin{array}{c}{x}_{i,j}^{new2}=\left\{\begin{array}{c}{{S}_{1}: x}_{best}+\left(2\times RB-1\right)\times {\left(1-\frac{t}{T}\right)}^{2}\times {x}_{i,j} , if q<{r}_{3}\\ {S}_{2}:{ x}_{i,j}+{r}_{4}\times \left({x}_{rand}-l\times {x}_{i,j}\right) , else\end{array}\right.\end{array}$$
(16)
$$\begin{array}{c}{X}_{i}=\left\{\begin{array}{c}{ X}_{i}^{new2}, if {F}_{i}^{new2}<{F}_{i}\\ {X}_{i } ,else\end{array}\right.\end{array}$$
(17)

In these equations, \(q=0.5\), \({r}_{3}\) and \({r}_{4}\) represent a \(1\times Dim\) array randomly generated from a normal distribution, \({x}_{rand}\) represents a randomly selected candidate solution for the current iteration, and \(l\) denotes a random integer, either 1 or 2.

Based on the above description, the flowchart of the Secretary Bird Optimization Algorithm is illustrated in Fig. 4.

Fig. 4
figure 4

Flow chart of SBOA.

Multi-strategy secretary bird optimization algorithm

To address the issues of poor local and global search capabilities, low population diversity, and consequently the inability to obtain high-quality solutions, this section proposes an improved version of the Secretary Bird Optimization Algorithm (SBOA), termed as the Multi-strategy Secretary Bird Optimization Algorithm (MSBOA). This enhancement is based on pooling mechanism, dynamic adaptability distance balancing technique, centroid reverse learning via greedy selection, and Preferential selecting search strategy.

Pooling mechanism

Given a matrix \(Pool=\left({P}_{1},{P}_{2},\dots ,{P}_{k}\right)\) of size \(k\), where each member \({P}_{i}=\left({P}_{i,1},{P}_{i,2},\dots ,{P}_{i,Dim}\right)\) is generated at the end of each iteration using Equation (18), where \({X}_{brand}\) is computed using Equation (19) to generate a random position within the neighborhood of the best secretary bird \({X}_{best}\), and \({X}_{worst}\) is the worst solution obtained in the current iteration \(t\). In this equation, \({B}_{i}\) is a binary random vector, \({\overline{B} }_{i}\) is its reverse vector such that the corresponding values of non-zero elements in \({B}_{i}\) are zero in \({\overline{B} }_{i}\) and vice versa. The pooling mechanism serves as a crossover operator, blending the worst solution with a promising one to generate a new solution, thus increasing diversity. Whenever the size of the \(Pool\) is completed, the new solution will replace the current member of the \(Pool\) 56.

$$\begin{array}{c}{P}_{i}={B}_{i}\times {X}_{brnd}+{\overline{B} }_{i}\times {X}_{worst}\end{array}$$
(18)
$$\begin{array}{c}{X}_{brand}={\xi }_{{best}_{min}}+{r}_{6}\times \left({\xi }_{{best}_{max}}-{\xi }_{{best}_{min}}\right)\end{array}$$
(19)

Where \({\xi }_{{best}_{min}}\) and \({\xi }_{{best}_{max}}\) are the lower and upper bounds of \({X}_{best}\), and \({r}_{6}\) is a uniformly distributed random number between 0 and 1.

Furthermore, in the evasion strategy camouflage \(\left({S}_{1}\right)\) stage of the secretary bird, two individuals are randomly selected from the \(Pool\) for mutation, enhancing population diversity and improving the algorithm’s optimization capability. The new strategy is represented as follows:

$$\begin{array}{c}{{S}_{1}: x}_{best}+\left(2\times RB-1\right)\times {\left(1-\frac{t}{T}\right)}^{2}\times \left(\left({{r}_{6}\times P}_{ran{d}_{1}}+{P}_{ran{d}_{2}}\right)\times 0.5\right)\end{array}$$
(20)

In this equation, \({P}_{ran{d}_{1}}\) and \({P}_{ran{d}_{2}}\) are individuals randomly selected from the Pool matrix in iteration \(t\), and \({r}_{6}=2\times rand\) is a random number in the range [0,2].

Dynamic fitness distance balance strategy

The Dynamic Fitness Distance Balance (DFDB) strategy is an improved version of the Fitness Distance Balance (FDB) method. FDB is a selection method aimed at effectively guiding the exploration and exploitation processes of the MHS algorithm during the search process 57. FDB is also a greedy method, where solution candidates are selected from the population based on their scores. What sets FDB apart from other selection methods is that it calculates scores for candidate solutions and selects based on these scores. In computing the scores, both the fitness value of the candidate solution and its distance from the best solution in the population are considered, ensuring that candidates with high fitness values are chosen 58. On the other hand, this also prevents the selection of a candidate solution very close to the best solution in the population, avoiding local optima. The implementation steps of the DFDB selection method are as follows:

a) When computing the score for each candidate solution in the population, both their fitness value and distance value are considered. Let \({X}_{i}=\left({x}_{i,1},{x}_{i,2},\dots ,{x}_{i,n-1},{x}_{i,n}\right)\) be the \({i}^{th}\) candidate solution in the population, and let \({X}_{best}=\left({x}_{1,best},{x}_{2,best},\dots ,{x}_{n-1,best},{x}_{n,best}\right)\) be the best solution in the population, where \(\left(i=\text{1,2},\dots ,N\right)\). The Euclidean distance \({D}_{{X}_{i}}\) between the \({i}^{th}\) candidate solution and the best solution \({X}_{best}\) in the population is calculated as follows:

$$\begin{array}{c}{}_{i=1}{}^{n}\forall {X}_{i},{D}_{{X}_{i}}=\sqrt{{\left({x}_{i,1}-{X}_{best,1}\right)}^{2}+{\left({x}_{i,2}-{X}_{best,2}\right)}^{2}+\dots +{\left({x}_{i,N}-{X}_{best,n}\right)}^{2}}\end{array}$$
(21)

b) The distances of candidate solutions in the population can be represented by a vector \({D}_{X}\) as follows:

$$\begin{array}{c}{D}_{X}={\left[ \begin{array}{l}{d}_{1}\\ \vdots \\ {d}_{j}\\ \vdots \\ {d}_{m}\end{array}\right]}_{m\times 1}\end{array}$$
(22)

c) The fitness values and distance values of candidate solutions are normalized for calculating the score of each candidate solution. The formula for calculating the score is as follows:

$$\begin{array}{c}{\forall }_{i=1}^{N}{ X}_{i}, { S}_{{X}_{i}}=w\times norm{F}_{{X}_{i}}+(1-w)\times norm{D}_{{X}_{i}}\end{array}$$
(23)

In these equations, \(w\) is the weighting coefficient determining the influence of fitness and distance values (\(norm{F}_{{X}_{i}}\) and \(norm{D}_{{X}_{i}}\)) on the score calculation. In the DFDB method, \(w=0.6\left(1-\frac{mod\left(t,round\left(T/10\right)\right)}{round\left(T/10\right)}\right)\). The DFDB score of solution candidates in the population is represented by the \({S}_{X}\) vector given by Equation (24).

$$\begin{array}{c}{S}_{X}={\left[ \begin{array}{l}{S}_{1}\\ \vdots \\ {S}_{j}\\ \vdots \\ {S}_{m}\end{array}\right]}_{m\times 1}\end{array}$$
(24)

To better balance the exploration and exploitation of the SBOA algorithm and address the issue of premature convergence, this paper integrates the DFDB technique into the SBOA algorithm. By dynamically adjusting the weights of fitness and distance between individuals, this integration aims to increase the diversity of the search, avoid premature convergence, and accelerate convergence speed. Consequently, this helps the algorithm explore the search space more effectively and find high-quality solutions.

Centroid opposition-based learning with greedy selection

Greedy selection is a common optimization algorithm strategy, often used to select local optimal solutions to construct global optimal solutions. In algorithms like K-means clustering, greedy selection is used to choose initial centroids. However, in some cases, this greedy selection may lead to getting stuck in local optima or selecting suboptimal initial centroids. Centroid reverse learning, on the other hand, improves optimization efficiency by introducing reverse solutions of the best solution as feasible solutions, thereby expanding the search space. However, pure reverse learning may pose the risk of falling into local optima. We propose a strategy based on greedy selection of centroid reverse learning. The implementation steps are as follows:

After each iteration, generate a random integer \(K\) from the interval \(\left[1,N\right]\), where \(N\) is the population size. Select the top \(K\) individuals with the best fitness in the current population, denoted as \(X=\left({X}_{1},{X}_{2}\dots {X}_{K-1},{X}_{K}\right)\), and compute the centroid of these \(K\) individuals as follows:

$$\begin{array}{c}Barycenter=\frac{\sum_{i=1}^{K}{K}_{i}}{K},K=\text{1,2},\dots ,N\end{array}$$
(25)

Generate the reverse solutions for each individual at the centroid position as follows:

$$\begin{array}{c}{X}_{i}{\prime}=2\times Barycenter-{X}_{i},i=\text{1,2},\dots ,N-1,N\end{array}$$
(26)

Finally, use greedy selection to choose the top \(N\) individuals with the best fitness values from the merged population \(X\cup {X}{\prime}\) as the new population. Here, \(X\) represents the initial population, and \({X}{\prime}\) represents the reverse solutions of \(X\).

To address the slow convergence speed and low precision of the SBOA, this paper integrates a random centroid reverse learning mechanism into the SBOA algorithm for updating the population. This integration aims to enhance the quality and diversity of the population, ensuring that the algorithm can comprehensively optimize the search space and thus increase the probability of finding the global optimum.

Preferential selecting search strategy

The Preferential Selecting Search (PSS) strategy enhances the exploration capability of the Secretary Bird Optimization Algorithm (SBOA) in searching for prey. This strategy is defined by Equation (27), where \({X}_{i}\) represents the current position of the \({i}^{th}\) secretary bird, \({P}_{ran{d}_{3}}\) and \({P}_{ran{d}_{4}}\) are individuals randomly selected from the Pool matrix in iteration \(t\), and \(Cauchy(i)\) is sampled from a Cauchy distribution computed by Equation (28). Because the Preferential Selecting Search strategy aims to enhance the exploration capability of SBOA, it requires a larger step size. By distributing the secretary birds in different regions of the search space, diverse solutions can be discovered. Therefore, the Cauchy distribution is employed, which has the characteristic of having long and flat tails, resulting in a high probability of generating large values. This distribution is utilized for the hunting attack stage of the secretary bird to enhance exploration 59. In the Chinese context, the utilization of the Cauchy distribution’s heavy-tailed property significantly improves search efficiency by increasing the probability of exploring distant regions in the solution space, thereby effectively preventing premature convergence and enhancing the algorithm’s global optimization capability.

$$\begin{array}{c}{x}_{i,j}^{new1}={x}_{best}+Cauchy\left(i\right)\times \left({P}_{ran{d}_{3}}-{P}_{ran{d}_{4}}\right), while t>\frac{2}{3}T\end{array}$$
(27)
$$\begin{array}{c}Cauchy=0.5+0.1\times tan\left(\pi \times rand\left(N,1\right)-0.5\right)\end{array}$$
(28)

In summary, the MSBOA updates the position information during the hunting and evasion stages using Equation (29) and Equation (30) respectively.

$$\begin{array}{c}{x}_{i,j}^{new1}=\left\{\begin{array}{c}{P}_{1}:{ x}_{i,j}+{r}_{2}\times {(x}_{r1}-{x}_{r2}), if t<\frac{1}{3}T\\ {P}_{2}:{x}_{best}+exp\left({\left(\frac{t}{T}\right)}^{4}\right)\times \left(RB-0.5\right)\times \left({x}_{best}-{x}_{i,j}\right), if \frac{1}{3}T<t<\frac{2}{3}T\\ {P}_{3}:{x}_{best}+Cauchy\left(i\right)\times \left({P}_{ran{d}_{3}}-{P}_{ran{d}_{4}}\right), else\end{array}\right.\end{array}$$
(29)
$$\begin{array}{c}{x}_{i,j}^{new2}=\left\{\begin{array}{c}{{S}_{1}: x}_{best}+\left(2\times RB-1\right)\times {\left(1-\frac{t}{T}\right)}^{2}\times \left(\left({{r}_{6}\times P}_{ran{d}_{1}}+{P}_{ran{d}_{2}}\right)\times 0.5 \right) , if q<{r}_{3}\\ {S}_{2}:{ x}_{i,j}+{r}_{4}\times \left({x}_{rand}-l\times {x}_{i,j}\right) , else\end{array}\right.\end{array}$$
(30)

The pseudocode for MSBOA is depicted as Algorithm 1.

Algorithm 1
figure a

Pseudo-Code of MSBOA

Computational time complexity of MSBOA

The performance of an algorithm is crucial, but it is equally important to evaluate its time complexity. In many optimization tasks, an algorithm must not only demonstrate excellent performance but also exhibit high real-time efficiency. Time complexity reflects how the algorithm’s runtime scales with the size of the input. Analyzing the time complexity of an optimization algorithm helps estimate its computational cost when handling large-scale problems. In the standard SBOA, the computational complexity of the defined control parameters is \(O(N\times D)\), where \(N\) represents the population size and \(D\) denotes the problem dimension. During the initialization phase, the algorithm requires \(O(N\times D)\) time. Furthermore, over \(T\) iterations, the computational complexity for updating individual positions is \(O(T\times N\times D)\).Therefore, the overall computational complexity of the SBOA algorithm can be expressed as \(O(T\times N\times D)\). In the proposed MSBOA, we introduce two enhancement strategies: a pooling mechanism and centroid backpropagation learning. While these strategies do not change the asymptotic complexity order, they introduce additional computational overhead that affects actual runtime. Pooling Mechanism: Maintaining \(k\) individuals and performing crossover operations adds an overhead of \(O(k\times D)\) per iteration. Centroid Backpropagation Learning: Computing the centroid and backpropagated solutions for \(K\) individuals introduces an additional cost of \(O(K\times D)\) per iteration. Therefore, the overall computational complexity of MSBOA can be expressed as \(O(T\times \left(N+k+K\right)\times D)\).

Experimental results and analysis

In this chapter, to validate the effectiveness of MSBOA, its performance is evaluated using the CEC2017 benchmark tests 60 and the unmanned aerial vehicle (UAV) path planning problem described in Section "UAV mathematical model", and compared with eight other advanced comparative algorithms. The CEC2017 benchmark functions are presented in Table 1, classified into four types: unimodal, multimodal, hybrid, and composite functions. Unimodal functions are utilized to assess accuracy and convergence speed on simple optimization problems; multimodal functions evaluate the algorithm’s local exploration capability; and hybrid and composite functions assess the algorithm’s capability to handle complex continuous problems. The experiments were conducted on a computing environment with a Windows 11 operating system, hardware configuration consisting of a 13th Intel(R) Core (TM) i5-13400 2.5GHz processor, 16GB RAM, and MATLAB 2023b software tools.

Table 1 CEC2017 benchmark functions.

Experimental results in CEC2017

In this section, MSBOA is compared with other advanced algorithms on the CEC2017 benchmark functions. All these functions are designed to minimize the objective function. The comparative algorithms include the classical algorithms Grey Wolf Optimizer (GWO) 35, Whale Optimization Algorithm (WOA) 36, Particle Swarm Optimization (PSO) 33, Runge Kutta optimizer (RUN) 61, as well as recently proposed advanced algorithms: Newton-Raphson-based optimizer (NRBO) 62, Crested Porcupine Optimizer (CPO) 63, Black-winged Kite Algorithm (BKA) 64, Dung Beetle Optimizer (DBO) 65, Secretary Bird Optimization Algorithm(SBOA) 51. The parameter settings for each algorithm are provided in Table 2. In the experiments, the population size is set to 30, and the maximum number of iterations is set to 500. To mitigate the influence of randomness and errors, each algorithm is independently run 50 times, calculating the mean and variance to obtain the final comparison results, with the best results highlighted in bold. Partial convergence curves of different algorithms on CEC2017 test functions are depicted in Fig. 5 and experimental statistical results are presented in Table 3. To visualize the ranking of each algorithm on each test function, radar charts as shown in Fig. 6 are plotted.

Table 2 Compare algorithm parameter settings.
Fig. 5
figure 5

Different algorithms in CEC2017 benchmark test the convergence curve of the function.

Table 3 CEC2017 benchmark function experiment results.
Fig. 6
figure 6

Ranking distribution of different algorithms.

According to the test results of Table 3, among the 30 functions in CEC2017, MSBOA performs the best on 25 functions, ranks second on 4 functions, and third on 1 function. In comparison, the performance distribution of the other 9 algorithms across the 30 functions is as follows: CPO and SBOA rank first on 2 and 3 functions respectively, while PSO, RUN, CPO, and SBOA rank second on 1, 3, 3, and 19 functions respectively. GWO, RUN, CPO, BKA, and SBOA rank third on 7, 5, 10, 1, and 6 functions respectively. From the perspective of rankings, the proposed MSBOA maintains its position in the top three on all 30 functions, demonstrating the stability of its performance.

It is noteworthy that the recently proposed CPO and SBOA have Friedman values of 3.83 and 2.30 respectively, higher than the classical algorithms GWO, WOA, PSO, and RUN with Friedman values of 5.37, 9.53, 5.87, and 4.20 respectively. This indicates that the overall performance of these recently proposed algorithms is superior to these classical algorithms on this problem. Moreover, the Friedman value of MSBOA is 1.20, surpassing all other comparative algorithms, indicating the effectiveness of the four integrated innovative strategies, which enhance the algorithm’s exploration performance and consequently improve its overall performance.

Statistical analysis

Wilcoxon rank sum test

In this subsection, the Wilcoxon rank-sum test 66 is employed to assess whether significant differences exist in the performance of the MECOA algorithm, without relying on assumptions of normality. Compared to the traditional t-test, the Wilcoxon test offers greater flexibility, as it remains applicable to data with non-normal distributions or outliers. The test statistic \(W\) for the Wilcoxon rank-sum test is defined by Equation (31).

$$\begin{array}{c}W=\sum_{i=1}^{{n}_{1}}R\left({X}_{i}\right)\end{array}$$
(31)

where \(R\left({X}_{i}\right)\) denotes the rank of \({X}_{i}\) among all observations. The test statistic \(U\) is calculated by Equation (32).

$$\begin{array}{c}U=W-\frac{{n}_{1}\left({n}_{1}+1\right)}{2}\end{array}$$
(32)

For larger sample sizes, \(U\) is approximately normally distributed by Equation (33)

$$\begin{array}{c}\genfrac{}{}{0pt}{}{{\mu }_{U}=\frac{{n}_{1}{n}_{2}}{2}}{{\sigma }_{U}=\sqrt{\frac{{n}_{1}{n}_{2}\left({n}_{1}+{n}_{2}+1\right)}{12}}}\end{array}$$
(33)

and the standardized statistic Z is calculated by Equation (34).

$$\begin{array}{c}Z=\frac{U-{\mu }_{U}}{{\sigma }_{U}}\end{array}$$
(34)

A significance level of 0.05 was adopted to determine whether the results of each MSBOA run exhibited a statistically significant difference from those of other algorithms. Under the null hypothesis (\({H}_{0}\)), it is assumed that no significant difference exists between the two algorithms. If the \(p\)-value is less than 0.05, \({H}_{0}\) is rejected, indicating a significant performance difference; otherwise, it is retained.

To further validate the superior performance of MSBOA compared to other benchmark algorithms, the Wilcoxon rank-sum test was employed to conduct a statistical significance analysis of the algorithms’ performance on the CEC2017 benchmark functions. The test results are presented in Table 4, where the "(+/=/-)" indicator represents the number of benchmark functions (out of 30) in which MSBOA performed better than, equal to, or worse than the comparison algorithms, respectively. The data in Table 4 shows that MSBOA exhibits strong performance competitiveness. Specifically, when compared with WOA, NRBO, BKA, and DBO, MSBOA outperformed all four algorithms across all 30 benchmark functions, with no instances of equal or inferior performance, fully demonstrating its comprehensive advantage in complex function optimization scenarios. In comparisons with GWO, PSO, and RUN, MSBOA performed slightly worse on only one benchmark function but maintained a performance lead on the remaining 29 functions, indicating that while it has minor limitations in a few specific function structures, its overall optimization capability is significantly superior to traditional classical algorithms. Even when compared with the relatively high-performing CPO and the original SBOA, MSBOA still outperformed them on 26 and 24 functions, respectively, while underperforming on only 4 and 6 functions, further confirming the performance enhancement achieved by the proposed multi-strategy improvements to the SBOA algorithm.

Table 4 Results for various algorithms on the CEC 2017.

Based on a significance level of 0.05, all the aforementioned performance differences passed the statistical significance test, indicating that MSBOA’s performance advantages are not due to random factors but are stable improvements resulting from its algorithmic structural enhancements. This lays a reliable theoretical foundation for its subsequent application to UAV path planning problems in complex environments.

Friedman mean rank test

In this subsection, the Friedman test 67 is used to determine the overall ranking of the MECOA algorithm relative to other methods. As a nonparametric approach, the Friedman test is suitable for comparing median performance differences across three or more matched groups. It is particularly well-suited for repeated measures or block designs, and is often employed as a robust alternative to ANOVA when the assumption of normality is violated. The Friedman test statistic is calculated according to Equation (35).

$$\begin{array}{c}Q=\frac{12}{nk\left(k+1\right)}\sum_{j=1}^{k}{R}_{j}^{2}-3n\left(k+1\right)\end{array}$$
(35)

where \(n\) is the number of blocks, \(k\) is the number of groups, and \({R}_{j}\) is the rank sum for \(j\)-th group. When \(n\) and \(k\) are large, \(Q\) follows approximately a \({\chi }^{2}\) distribution with \(k-1\) degrees of freedom.

According to the Friedman mean rank test results in Table 5, MSBOA achieved an average rank (M.R) of 1.20 on the CEC2017 benchmark functions, with a total rank (T.R) of first place, demonstrating significant superiority over other comparison algorithms. Traditional algorithms such as GWO (M.R=5.37, T.R=5) and WOA (M.R=9.53, T.R=10), as well as recent advanced algorithms like CPO (M.R=3.83, T.R=3) and the original SBOA (M.R=2.30, T.R=2), all exhibited lower average and total rankings compared to MSBOA. These results fully confirm that MSBOA delivers superior comprehensive performance when handling unimodal, multimodal, hybrid, and composition optimization problems covered by the CEC2017 benchmark suite. The multi-strategy improvements effectively enhance the algorithm’s overall optimization capability, providing reliable support for its subsequent application to UAV path planning in complex environments.

Table 5 Friedman mean rank test result.

Experiment in UAV path planning

Scenario Settings

The scenarios used for evaluation are based on real Digital Elevation Model (DEM) (terrain resolution:\(1045\times 879\))maps derived from LiDAR sensors 68. Initially, two terrains from Christmas Island, Australia, representing both land and ocean topographies, are selected. When the altitude is below sea level, it is uniformly set to 0. Subsequently, different regions with varying complexities are enhanced to generate eight benchmark scenarios. In these eight scenarios, the number and positions of threatening objects (represented by red cylinders) vary, implying different levels of complexity in solving, thus effectively assessing the algorithm’s ability to find the optimal path.

In this study, the number of waypoints is set to \(n = 12\), corresponding to 10 line segments55. In each comparison, the parameter settings for all algorithms are as shown in Table 2, and each algorithm is run 50 times to obtain the mean and standard deviation. Additionally, we employ the non-parametric Wilcoxon rank-sum (P-value) test to compare differences between algorithms 69. "\(N/A\)" indicates that the results of two algorithms are too similar to make a significant judgment. If the result exceeds 0.05, we consider there is no significant difference between the algorithms and highlight accordingly. Conversely, significant differences exist between the two algorithms.

Compared with the classical algorithm

The top-down view of the paths generated by MSBOA and classical algorithms is depicted in Fig. 7. It can be observed that all algorithms are capable of generating feasible paths that satisfy requirements for path length, threats, turning angles, climb/descent angles, and altitude. However, their optimality varies depending on the scenario. For simple scenarios 1, 2, 5, and 6, all algorithms exhibit good convergence, albeit with slightly different fitness values. The test values in Table 6 also indicate that GWO is not statistically significant in scenarios 1, 2, 5, and 6. However, for complex scenarios 3, 4, 7, and 8, their performances differ significantly. MSBOA is able to achieve near-optimal solutions, while GWO, PSO, and SBOA converge to relatively good solutions only. WOA fails to find high-quality solutions. Table 6 displays the mean, standard deviation, and non-parametric Wilcoxon rank-sum test values, further confirming the fitness values of the tests. It suggests that MSBOA statistically achieves the best fitness across most tested scenarios, while GWO and PSO are effective only in simple scenarios. RUN introduces relatively good results across all scenarios, reflecting stable convergence with minor deviations.

Fig. 7
figure 7

Top view of the paths generated by the classical algorithms.

Table 6 Experimental results of MSBOA and classical algorithms.

Figure 8 provides a closer observation of the behavior of these classical algorithms by displaying their best fitness values during iterations. From the graph, it can be observed that in simple scenarios 1 and 5, GWO obtains good fitness values with fast convergence speed. However, for other complex scenarios, it is prone to getting trapped in local optima, failing to achieve high-quality solutions, similar to other algorithms. On the other hand, MSBOA consistently obtains high-quality solutions across all eight scenarios. Although in scenarios 2, 4, and 6, its convergence speed in the early iterations is slower, it does not get trapped in local optima like SBOA, GWO, and WOA, but continues to search for higher-quality solutions. This demonstrates the effectiveness of the four embedded improvement strategies, enhancing the overall performance of MSBOA, capable of obtaining relatively smooth and shorter paths in complex UAV path planning scenarios.

Fig. 8
figure 8

Best fitness values over iterations of the classical algorithms.

Compare with recent algorithms

To further evaluate the performance of MSBOA, we compare its efficacy with other recently proposed advanced metaheuristic algorithms, namely Newton-Raphson-based optimizer (NRBO), Crested Porcupine Optimizer (CPO), Black-winged Kite Algorithm (BKA), Dung Beetle Optimizer (DBO), and Secretary Bird Optimization Algorithm (SBOA). The parameter settings for each algorithm adhere to the standards specified in their respective original literature, as illustrated in Table 2 above.

Table 7 presents the fitness results. It is evident that the performance of MSBOA surpasses that of other algorithms. Specifically, across all scenarios, except for the standard deviation in scenarios 1 and 3 where SBOA slightly outperforms, both the mean and standard deviation of MSBOA excel those of the comparative algorithms. Particularly noteworthy is its significantly superior mean performance in complex scenarios 3, 4, 5, and 8 compared to other algorithms.

Table 7 Experimental results of MSBOA and recent algorithms.

Figure 9 provides a top-down view of the generated paths. It is evident that all algorithms are capable of generating collision-free paths. However, MSBOA consistently produces the smoothest and shortest paths in most cases. In contrast, the original SBOA only achieves relatively smooth and short paths in simple environments; its performance significantly deteriorates in complex environments, failing to generate the smoothest and shortest paths.

Fig. 9
figure 9

Top view of the paths generated by the recent algorithms.

Figure 10 illustrates the variation of the best fitness value during iterations. It is apparent that DBO quickly converges to premature solutions as the number of iterations increases. Due to its limited exploration capability, CPO exhibits slow convergence and is prone to getting trapped in local optima during later iterations. SBOA demonstrates stable convergence but also falls into local optima towards the end of iterations, failing to attain higher-quality solutions. In contrast, MSBOA exhibits sufficiently rapid convergence, attributed to the four effective improvement strategies introduced, thereby enhancing its convergence performance.

Fig. 10
figure 10

Best fitness values over iterations of the recent algorithms.

Summary and prospect

This paper proposes an improved Secretary Bird Optimization Algorithm – the Multi-Strategy Secretary Bird Optimization Algorithm (MSBOA) – for solving unmanned aerial vehicle (UAV) path planning problems, with a focus on ensuring path safety and feasibility. The design of the cost function incorporates constraints related to optimality, safety, and feasibility simultaneously. MSBOA is devised by integrating four innovative strategies: the Pooling mechanism, Dynamic Fitness Distance Balance Strategy, Centroid Opposition-based Learning with Greedy Selection, and Preferential Selecting Search Strategy.

To validate the effectiveness of MSBOA, comparisons are conducted with nine other algorithms on the CEC2017 benchmark functions. Results indicate that MSBOA exhibits superior convergence performance. Comparative analyses on eight benchmark scenarios generated from DEM maps demonstrate that MSBOA consistently achieves optimal-quality paths in most cases, along with high robustness. Comparisons with other metaheuristic algorithms also confirm the outstanding performance of MSBOA.

As the No Free Lunch theorem posits, no metaheuristic algorithm can outperform all others across all problem domains, and MSBOA is no exception, with its limitations particularly evident in handling certain CEC2017 functions. Its inferior performance compared to SBOA and CPO on F10, F27, and F30 stems from mismatches between its core mechanisms and the specific characteristics of these functions, which also points to directions for targeted improvements:

For F10, which features a rugged landscape with the global optimum located at the search space boundary, MSBOA’s underperformance is primarily attributed to its centroid reverse learning strategy. This strategy tends to generate reverse solutions concentrated in the central region of the search space, resulting in insufficient exploration of boundary areas. In contrast, SBOA’s Levy flight (which enables occasional long jumps to boundary regions) and CPO’s adaptive local search (which intensifies exploitation near promising boundary areas) are better suited to this function’s trait. To address this, future iterations could modify the centroid reverse learning mechanism—for instance, by adjusting the centroid calculation to incorporate boundary information or introducing a probability weight that biases reverse solution generation toward boundary regions, thereby enhancing boundary exploration.

For composite functions F27 and F30, which consist of fragmented sub-regions with distinct optima, MSBOA’s pooling mechanism leads to computational redundancy when adapting to these dispersed sub-regions. This is because the pooling strategy aggregates information across the entire population, diluting the specificity required to respond to fragmented landscapes. In comparison, SBOA and CPO’s simpler update rules (which prioritize local information exchange) allow faster adaptation to sub-region characteristics. Potential improvements here could involve refining the pooling mechanism—such as dynamically partitioning the population into sub-groups corresponding to detected sub-regions, or introducing adaptive weights to the pooling process that emphasize local sub-region information, thus reducing redundancy and accelerating adaptation to composite function structures.

In some UAV path planning scenarios, MSBOA lags behind GWO, BKA, and SBOA in early convergence speed. This is because it employs centroid reverse learning to avoid premature convergence—a critical design choice for complex environments with dense threats—but this comes at the cost of slower early exploration. In contrast, GWO, BKA, and SBOA rely on fast local exploitation or stage-wise hunting strategies, which enable rapid early convergence in simple, sparse-threat scenarios. To balance this, future work could explore a dynamic strategy adjustment mechanism: adopting more aggressive local exploitation in the early stages for sparse-threat environments while retaining the current anti-premature convergence mechanisms for complex environments, thus improving adaptability across scenario types.

To further expand the theoretical depth and practical application value of the Multi-Strategy Secretary Bird Optimization Algorithm (MSBOA), subsequent research will be conducted across multiple dimensions. On one hand, for multi-UAV cooperative operation scenarios, MSBOA will be integrated with a distributed optimization framework to design a multi-objective cost function that incorporates task completion time, fuel consumption, and cooperative collision avoidance distance. Simultaneously, a communication constraint model will be introduced to adapt to the data transmission characteristics between UAVs, enabling efficient multi-UAV cooperative path planning. On the other hand, focusing on the demands of dynamic online environments, a dynamic threat prediction model based on Kalman filtering or LSTM neural networks will be constructed. An incremental population update strategy will be designed to reduce computational redundancy during environmental changes, ensuring that UAVs can quickly adjust their paths in scenarios involving moving threats and sudden obstacles. Additionally, a Hardware-in-the-Loop (HIL) test platform integrating UAV flight control systems, simulation software, and environmental sensors will be developed. This platform will validate the path tracking accuracy and robustness of MSBOA in an environment simulating real-flight physical characteristics. Through outdoor physical experiments using small consumer-grade UAVs, the algorithm will be advanced from software simulation to practical application. Furthermore, the step-size control mechanism of the priority selection search strategy will be optimized to enhance adaptability to high-dimensional composite problems. The algorithm will also be extended to cross-domain scenarios such as unmanned vehicle path planning and photovoltaic system optimization. Relevant source code and experimental toolkits will be open-sourced to promote academic exchange and industry application.