1 Introduction

In recent years, with the continuous development of various industries, the complexity of mathematical problems and the volume of data have significantly increased. Consequently, optimization algorithms have gained increasing importance. Optimization aims to find the most favorable value of an objective function while adhering to a set of constraints (Zhao et al. 2022; Hamad et al. 2022). It involves seeking the optimal solution within specific constraints, which often include real-world laws and practical considerations such as physics laws, geometric constraints, and material properties. Constraints can be broadly classified into two types: equality constraints and inequality constraints. These constraints are expressed as mathematical equations in the form of equalities and inequalities, respectively. The decision variables involved in the problem must satisfy both types of constraints. In the past, gradient-based information of the involved functions was commonly used to search for the optimal solution (Sulaiman et al. 2020). While this approach is effective in certain situations, it has limitations when dealing with problems that involve complex mathematical functions. In such scenarios, satisfactory results may not be achieved using this approach, necessitating the exploration of alternative methods.

Metaheuristic algorithms (MHs) are a type of algorithm that simulates mechanisms found in nature or biological evolution. These algorithms have gained increasing popularity among researchers. One notable advantage of metaheuristic algorithms is their ability to address challenging problems without relying on gradients, making them suitable for finding solutions that are close to optimal (Hashim et al. 2024). Furthermore, metaheuristic methods possess attributes such as simplicity, adaptability, a derivation-free framework, and the capability to overcome local optima (Yıldız et al. 2023). These features enable them to tackle complex mathematical problems, providing valuable approximations even if they may not achieve optimality. Over the past years, there has been a proliferation of metaheuristic algorithms, which can be broadly categorized into four main groups based on their sources of inspiration: evolution-based, physics-inspired, human-mimetic, and swarm intelligence optimization algorithms. By categorizing metaheuristic algorithms into these distinct groups based on their sources of inspiration, researchers have been able to develop a diverse range of algorithms with unique characteristics and capabilities.

  1. 1.

    Evolution-based heuristic algorithms simulate the evolutionary law of “survival of the fittest” observed in nature to achieve overall progress in a population. These algorithms utilize concepts such as genetic operators and selection mechanisms to guide the search for optimal solutions, making them effective in exploring diverse solution spaces and adapting to changing environments. Examples of such algorithms include genetic algorithms (GA) (Holland 1992), differential evolution (DE) (Storn and Price 1997), Biogeography-Based Optimizer (BBO) (Simon 2008), evolutionary programming (EP) (Yao et al. 1999), and evolutionary strategies (ES) (Amoretti 2014).

  2. 2.

    Physics-based algorithms draw inspiration from the laws of physics in the real world. These algorithms employ physical concepts like energy, force, temperature, etc. to guide the search process, often leveraging analogies from thermodynamics or celestial mechanics. This enables them to effectively navigate complex solution landscapes and escape local optima. Examples of these algorithms include simulated annealing (Kirkpatrick et al. 1983), big-bang big-crunch (BBBC) (Erol and Eksin 2006), gravitational search algorithm (GSA) (Rashedi et al. 2009), charged system search (CSS) (Kaveh and Talatahari 2010), central force optimization (CFO) (Formato 2007), and artificial chemical reaction optimization algorithm (ACROA) (Alatas 2011).

  3. 3.

    Human-mimetic algorithms are based on human behavior. These algorithms take inspiration from various aspects of human society, such as learning, cooperation, or competition, to guide the optimization process. They often incorporate social or cognitive mechanisms to balance exploration and exploitation, allowing them to effectively adapt to dynamic environments. Notable examples in this category include the volleyball premier league algorithm (VPL) (Moghdani and Salimifard 2018), meta-optimization inspired by the FBI’s methodologies (FBI) (Chou and Nguyen 2020), algorithm for optimization based on the teaching-learning paradigm (TLBO) (Rao et al. 2011), and harmony search (HS) (Geem et al. 2001).

  4. 4.

    Swarm intelligence optimization algorithms mimic the social behavior observed in social groups. These algorithms imitate the collective intelligence and cooperation seen in natural systems, where multiple agents interact and share information to collectively optimize the search process. They often feature communication and coordination mechanisms among the individuals of the swarm, enabling efficient exploration and exploitation of the search space. The most prominent optimization algorithm in this class is particle swarm optimization (PSO). Other examples include artificial bee colony (ABC) (Karaboga and Akay 2009), dandelion optimizer (DO), ant-lion optimizer (ALO) (Mirjalili 2015), cuckoo search algorithm (CSA) (Rakhshani and Rahati 2017), African vultures optimization algorithm (AVOA) (Abdollahzadeh et al. 2021), nutcracker optimizer algorithm (NOA) (Abdel-Basset et al. 2023a), Hybrid Moth Flame Optimizer (HMFO) (Sahoo and Saha 2022), and snake optimizer (SO) (Hashim and Hussien 2022).

In addition to these categories, metaheuristic algorithms can also be classified based on the number of search agents involved and the number of objectives being optimized. Single-agent optimization algorithms, such as simulated annealing and tabu search, rely on a single agent to explore the search space and find optimal solutions. In contrast, population-based optimization algorithms, like genetic algorithms and particle swarm optimization, utilize a population of agents that collaborate and share information to effectively navigate the solution landscape (Sahoo et al. 2023a, b).

Furthermore, metaheuristic algorithms can be designed to solve either single-objective optimization problems, where the goal is to optimize a single performance measure, or multi-objective optimization problems, where multiple, often conflicting, objectives need to be considered simultaneously (Sahoo et al. 2024a, b).

Swarm intelligence optimization algorithm as a kind of methods that ignore gradient changes, it main utilize particles swarm continuously explores the feasible space, and updates and iterates through certain strategies to achieve better fitness values. This process is somewhat analogous to a process of continuous trial and error, what we need to do is find a best feasible solution under limited amount of computation. Therefore, metaheuristic algorithms conduct a selective exhaustive search, which gives birth to the two stages of exploration and exploitation. The significant numerical disparities among individuals within a dimension suggest their widespread distribution in the search environment. This widespread exploration is termed ”Exploration” in the realm of swarm intelligence optimization algorithms. On the contrary, as the population converges, the disparities diminish, resulting in the individuals converging towards a specific region. This is known as ”Exploitation”. Exploration can be understood as a process of global search, whereas Exploitation is akin to a local search. Both these search paradigms play a pivotal role in determining the effectiveness of the algorithm. Different swarm intelligence optimization algorithms seem to be doing the same thing, that is balance the exploration and exploitation. Empirical experimentation has unequivocally shown that there exists a profound correlation between the exploration-exploitation capacity of a particular search method and its rate of convergence (Sahoo et al. 2022, 2023c; Morales-Castañeda et al. 2020; Sahoo et al. 2024c).

The execution and transition between the exploration and exploitation phases differ across various optimization algorithms, highlighting the diversity in their approaches. Each algorithm employs distinct strategies to conduct these phases, allowing for effective navigation of the solution space. The specific methods utilized during these phases significantly impact the algorithm’s overall performance and its ability to achieve optimal solutions. Examples of such algorithms include sine cosine algorithm (SCA) (Mirjalili 2016), use the parameter r to control the two stage. When r is less than 1, the optimization algorithms is in the exploration stage, when r is more than 1 the optimization algorithms is in the exploitation stage. Similarly, the grey wolf optimizer (GWO) (Mirjalili et al. 2014) used the parameter A to control the two stage. The arithmetic optimization algorithm (AOA) (Abualigah et al. 2021), which mainly consists of four operators: addition, subtraction, multiplication and division.In addition, subtraction operator the AOA mainly going the exploration stage, in the multiplication and division operator the AOA mainly going the exploitation stage. The dandelion optimizer (DO) (Zhao et al. 2022), which mainly consists of three operators: rising stage, descending stage, and landing stage in the rising and descending stage the algorithm mainly going to exploration in the landing stage the algorithm mainly going to exploitation. In the African vultures optimization algorithm (AVOA), when \(|F| \ge 1\), the vulture is in the exploration stage, and when \(|F| \le 1\), it is in the exploration stage. Swarm intelligence optimization algorithm optimization problems abound in our lives. Hussain et al. used artificial bee colony (ABC) algorithm to optimize co-clustering (Hussain et al. 2020). Tharwat et al. use Chaotic antlion algorithm to optimize parameter of support vector machine (Tharwat and Hassanien 2018). Braik et al. used cuckoo search algorithm (CSA) to model industrial coiling process (Braik et al. 2023). Swain et al. used African vultures optimization algorithm to stabilize frequency response of a microgrid (Swain et al. 2023). Chen et al. used the grey wolf optimization (GWO) to conduct research on face recognition (Chen et al. 2023). Wang et al. used the adaptive parallel arithmetic optimization algorithm to plan robot path (Wang et al. 2021).

Although there are so many different optimization algorithms, based on the ”No Free Lunch” theorem (Wolpert and Macready 1997), it is hardly possible that an optimization algorithm can solve all optimization problems very well. Different algorithms have different effects in solving different problems. Perhaps an algorithm is very efficient in handling one type of the problems, but has certain flaws in dealing with another type of problem. Concerning this, it is imperative to propose either novel optimization techniques or modified versions of the existing ones in order to address subsets of tasks across diverse fields. Li et al. used Levy flight and nonlinear sine function to enhance the solution quality of northern goshawk optimization algorithm (Li et al. 2023). Jiang et al. (2023) proposed an adaptive tree seed algorithm (ATSA) by using the tree migration and seed intelligent generation mechanisms. Alsattar et al. 2019 proposed a novel swarm intelligence algorithm bald eagle search (BES) for solving engineering optimization problems inspired by the hunting strategy and intelligent social behaviours of bald eagles as they search for prey. The algorithm mainly consist of three stage: select stage, search stage and swoop stage. The select stage is exploration stage of the algorithm. In this stage, bald eagles extensively search for the location of the prey. The search stage and swoop stage are exploitation stages of the algorithms. In this stage, bald eagles precisely locate the prey’s position. The superiority and distinction of BES are clearly demonstrated by mean, standard deviation, best point and Wilcoxon signed-rank test on CEC2014, CEC2017 test functions. From the function convergence curve, it can be clearly seen that BES has a faster convergence speed than other compared algorithms, such as DO (Zhao et al. 2022), GWO (Mirjalili et al. 2014), AOA (Abualigah et al. 2021), WOA (Mirjalili and Lewis 2016), SCA (Mirjalili 2016), and the optimal solution found is better. BES also has the advantage of simple structure few control parameters.

Since its inception, the BES algorithm has garnered significant attention from researchers due to its superior performance in various applications. It has been extensively studied and enhanced to optimize diverse domains, such as fuel cell design (Alsaidan et al. 2022), pixel selection (Bahaddad et al. 2023), battery parameter extraction (Ferahtia et al. 2023), path planning for unmanned vessels (Chen et al. 2023), battery capacity enhancement (El Marghichi et al. 2023), feature selection (Chhabra et al. 2023), controller parameter tuning (Liu et al. 2023), speaker verification (Nirmal et al. 2024), energy scheduling optimization (Zhang et al. 2022), and dam safety monitoring (Yu et al. 2023). Notably, the BES exhibits faster convergence, higher accuracy, and outperforms many well-known algorithms. Its simplicity, comprehensibility, and ability to achieve desirable results with smaller populations make it an appealing choice.

Despite its favorable optimization effects, BES faces challenges in finding the global optimal solution. After a large number of iterations, the difference in performance between BES and its early iterations becomes negligible, leading to inefficient utilization of computational resources. Moreover, BES exhibits limitations in terms of low convergence accuracy and inadequate handling of high-dimensional and complex problems. Previous attempts at improving BES have either been overly complex or effective only for specific problem domains.

As a result, researchers have been motivated to further explore and extend the BES model, aiming to unlock even more remarkable capabilities and address its limitations. The ongoing efforts in developing modified versions of BES seek to enhance its overall performance, overcome convergence issues, and improve its applicability to high-dimensional and complex problem scenarios. For that reason, (Ramadan et al. 2021) introduced IBES, a novel variant of the BES algorithm, designed for precise estimation of photovoltaic (PV) model parameters. IBES improves upon the original BES by dynamically adjusting the learning parameter, which controls the position changes in each iteration. The value of this parameter is updated using a decay equation to enhance exploration and exploitation. IBES exhibits superior performance compared to the original BES and other algorithms across 23 classical benchmark tests. Real data from a commercial silicon solar cell is employed to highlight IBES’ exceptional accuracy in parameter estimation for both original and modified PV models. In another work, (Chen et al. 2023) proposed mBES, a modified version of the BES algorithm that addresses its limitations by incorporating opposition-based learning (OBL), chaotic local search (CLS), and transition & pharsor operators. Evaluation on 29 CEC2017 and 10 CEC2020 benchmark functions, engineering design problems, and a feature selection problem demonstrates the superiority of mBES over classical metaheuristic algorithms, highlighting its effectiveness in solving diverse optimization problems.

Ferahtia et al. (2023) proposed a modified bald eagle search (mBES) algorithm. The mBES algorithm introduces adaptive parameters based on the current and maximum number of iterations to enhance diversity and address issues such as sluggish convergence and an imbalance between exploitation and exploration. The algorithm’s performance is evaluated using benchmark functions CEC2020 and CEC2022. Additionally, the efficacy of mBES is demonstrated in the complex problem of parameter extraction for lithium-ion batteries. The results of this evaluation are compared to other state-of-the-art algorithms, and statistical tests are conducted to validate the significance of the findings. Overall, the mBES algorithm exhibits promising results in real-world scenarios.

In Alsaidan et al. (2022), Alsaidan et al. introduced a novel optimization algorithm, the enhanced bald eagle search (EBES) optimization algorithm, to estimate the design variables of proton exchange membrane fuel cells. The main objective is to minimize the sum of squared errors (SSE) by improving the accuracy of the calculated data compared to the measured data. The EBES algorithm builds upon the original BES by incorporating enhancements for increased search efficiency and overcoming local optima via Levy function. To evaluate the proposed algorithm, three tested cases are utilized, including the BCS 500 W, 250 W, and Horizon H-12 stacks. Furthermore, the impact of pressure and temperature variations on the algorithm’s performance is investigated. The results are compared with existing optimization algorithms to highlight the effectiveness and superiority of the newly developed EBES algorithm in solving complex optimization problems.

To address the challenges encountered in the path planning of unmanned ships in complex waters. Chen et al. (2023) proposed a novel optimization algorithm called the self-adaptive hybrid bald eagle search (SAHBES) algorithm, which enhances the traditional BES algorithm by incorporating adaptive factors and pigeon-inspired optimization (PIO). The improved fitness function incorporates a distance metric between the ships’ path corners and factors in obstacle-based path length calculations. Furthermore, a curve optimization module is employed to generate optimal and collision-free path planning results. The proposed SAHBES algorithm is evaluated through 4 selected multimodal surface functions from 23 standard test functions (CEC2005) and path planning under different obstacle scenarios. The results demonstrate that the SAHBES algorithm can effectively generate the shortest and smoothest paths in complex water environments, while considering the limitations of ship maneuvering operations.

To address issues related to local optima stagnation and premature convergence in the BES algorithm, (Sharma et al. 2023) introduced the self adaptive bald eagle search (SABES) algorithm. The SABES algorithm incorporates the dynamic-opposite learning (DOL) method during the initialization process to improve population diversity and convergence speed. To enhance the exploitation capability and discover better global solutions, dynamic-opposite solutions are considered. Additionally, the algorithmic parameter values of the BES algorithm are determined using a linear and non-linear time-varying adaption strategy, striking a balance between search abilities and overall performance. The performance of the SABES algorithm is evaluated by CEC2017 and CEC2020 test functions, in comparison to previous algorithms. The SABES algorithm achieves optimal results for the highest number of functions compared to state-of-the-art algorithms.

A novel approach called the polar coordinate bald eagle search algorithm (PBES) is proposed for curve approximation tasks by Zhang et al. (2022). Inspired by the spiral predation mechanism of the bald eagle, the PBES algorithm utilizes a polar coordinate representation to optimize problems in polar coordinates more effectively. The initialization stage of the PBES algorithm is modified to ensure a more uniform distribution of initialized individuals, while additional parameters are introduced to enhance exploration and exploitation capabilities. The performance of the PBES algorithm is evaluated through experimentation on various benchmark functions, including polar coordinate transcendental equations, curve approximation, and robotic manipulator problems. The results demonstrate the superiority of the PBES algorithm over well-known metaheuristic algorithms in effectively addressing curve approximation problems.

A variant known as the cauchy adaptive BES algorithm (CABES) is proposed by Wang et al. (2023). CABES integrates Cauchy mutation and adaptive optimization techniques to enhance the performance of BES in mitigating local optima. In CABES, the Cauchy mutation strategy is introduced to adjust the step size during the selection stage, enabling the selection of a more favorable search range. Furthermore, in the search stage, the search position update formula is updated using an adaptive weight factor, further improving the local optimization capability of BES. To evaluate the performance of CABES, simulations are conducted using the benchmark functions from CEC2017. Comparative analysis is performed against three other algorithms, namely PSO, whale optimization algorithm (WOA), and Archimedes optimization algorithm (AOA). The experimental results demonstrate that CABES exhibits strong exploration and exploitation capabilities and competes favorably with other tested algorithms. Moreover, CABES is applied to four constrained engineering problems and a groundwater engineering model to validate its effectiveness and efficiency in practical engineering scenarios. These applications further confirm the robustness and efficacy of CABES.

In the domain of vehicular networks, the focus is on reducing computing delay and energy consumption. To address this challenge, a task offloading model that combines local vehicle computing, mobile edge computing server computing, and cloud computing is proposed by Shen et al. (2022). This model takes into account task prioritization, as well as system delay and energy consumption. In their work, an improved variant of the BES, named IBES, is introduced to make computational of floading decisions. The IBES algorithm incorporates several enhancements to the original BES, including the integration of tent chaotic mapping, Levy flight mechanism, and adaptive weights. These modifications aim to increase the diversity of the initial population, enhance local search capabilities, and improve global convergence. To evaluate the proposed IBES algorithm, simulations are conducted and compared against two other algorithms, namely PSO and the original BES. The results reveal that the total cost of IBES is found to be 33.07% and 22.73% lower than that of PSO and BES, respectively. These findings demonstrate the superior performance of the IBES in this task.

Table 1 offers an overview of the referenced studies, encompassing five primary criteria: author names and references, the designation of the BES version, the method of enhancement employed, the utilization of the IEEE CEC benchmark functions, and the identification of real-world applications.

Table 1 Summary of proposed enhanced BES algorithms and their applications

Despite the extensive development of various BES variants by numerous researchers, the underlying limitations of the algorithm persist. These limitations serve as the impetus for our endeavor to propose a novel variant that aims to overcome these challenges and enhance BES’s efficacy in addressing complex optimization problems. Introducing the multi-strategy boosted bald eagle search algorithm (MBBES), our proposed variant integrates multiple strategies to effectively address the shortcomings of original BES. MBBES incorporates several enhancements, including combines two mutations that are DE/Current-to-best/1 mutation and DE/BEST/1 mutation, one adaptive weight parameters \(\alpha\), and the fall stage. Firstly, the adaptive weight parameters \(\alpha\) is applied at the select space stage to increase the randomness of the algorithm and improve the global search capability. Secondly, the DE/Current-to-best/1 mutation is used in the select space stage to increase the possibility of algorithm for finding the global optimal. And then the DE/BEST/1 mutation is added to the search space stage, which helps to conduct in-depth search within the neighborhood of the current individual optimal value. Finally, the fall stage is used to instead the swoop stage. With the increasing of the number of iterations, the fall step size increases gradually, which helps the algorithm to jump out of the local optimal solution. The main contributions of this study are as follows:

  1. 1.

    MBBES introduces enhancements to the BES algorithm, including the incorporation of adaptive parameters, two DE mutation strategies, and the replacement of the swoop stage with a fall stage. These modifications aim to enhance the exploration and exploitation capabilities of the optimizer.

  2. 2.

    The performance of MBBES is evaluated by comparing it to other advanced algorithms on the CEC2014 and CEC2017 test sets. Statistical indicators such as mean, standard deviation, convergence curve, Wilcoxon signed-rank test, and box plots are utilized to demonstrate the superior performance of MBBES.

  3. 3.

    The practical evaluation conducted in this research demonstrates the effectiveness of MBBES in real-world scenarios. The practical applicability of MBBES is showcased in two ways:

    1. (a)

      By successfully solving a variety of engineering problems, including Welded Beam Design, Pressure Vessel Design, Reducer Design, Three-Bar Truss Design, and String Design. MBBES proves its capability to provide optimized solutions for these engineering challenges.

    2. (b)

      By effectively solving multi-layer Perceptron problems. MBBES demonstrates its ability to optimize the performance of multi-layer Perceptron models, showcasing its applicability in this domain.

The organization of the paper is structured as follows: Sect. 2 introduces the basic BES algorithm. Section 3 demonstrates the improvements for the BES algorithm. Section 4 discussed the optimization performance of MBBES on two sets of functions, which are the CEC2014 and CEC2017. In Sect. 5, five engineering design optimization problems are utilized to verify the superiority of proposed MBBES. Section 6 is about the application of MBBES in three multi-layer perceptron (MLP) classification problems. Finally, conclusions and look forward to the research content of this article are given in Sect. 7.

2 The bald eagle search algorithm

In this section, the bald eagle search algorithm (BES) is described in detail. It is raised by Hassan et al. (2019) in 2019, which is a population-based and nature-inspired metaheuristic algorithm. The BES mimics the hunting strategies and intelligent social behaviours of bald eagles when they search for fish. The hunting processes of bald eagles are divided into three stages: select space stage, search space stage and swoop stage. The mathematical models of BES are described in the following.

2.1 Initialization

The initialization operation is used to determine the initial positions of bald eagle group. The initial positions of bald eagles are an important factor for the global optimal solution, which may affect the convergence speed and accuracy of BES. In the BES, the bald eagles are randomly distributed in the search space. The initial positions are generated by Eq. (1):

$$\begin{aligned} X_{i,j} = rand \times (ub_j - lb_j) + lb_j,~i = 1,~2,~\dots ,~N,~j = 1,~2,~\dots ,~D \end{aligned}$$
(1)

where \(X_{i,j}\) is the position of individual within the defined area. \(ub_j\) and \(lb_j\) indicate the j-th upper bound and nether bound of the feasible domain. N is the size of bald eagle group. D is the dimension of variables. And rand is a random number within [0, 1].

2.2 Select space stage

During the hunting process of bald eagles, the first step is to identify the area where prey may appear. At this stage, all bald eagles randomly search the space based on the previously selected search space. In other words, the bald eagles utilize the location information obtained during the previous hunting to search other areas near it, where may exist better optimal solution. The position is calculated by Eq. (2):

$$\begin{aligned} P_{i,new} = P_{best} + \alpha \times r \times (P_{mean} - P_i) \end{aligned}$$
(2)

where \(P_{i,new}\) is the new generated position of the i-th individual. \(P_{best}\), \(P_{mean}\) and \(P_{i}\) are the prey position, average position and original i-th individual position, respectively. \(\alpha\) is a constant in the interval [1.5, 2], and r is a random number lying in [0, 1].

2.3 Search space stage

After identifying the area where the prey may appear, the bald eagles accelerate to their prey within the chosen search space using the spiral line. This spiral search not only improves the convergence speed of BES, but also increases the likelihood of finding the global optimum. The trajectory of spiral motion is mathematically modeled by Eqs. (3)–(6):

$$\begin{aligned} P_{i,new}= & P_{i} + y(i) \times (P_i - P_{i + 1}) + x(i) \times (P_i - P_{mean}) \end{aligned}$$
(3)
$$\begin{aligned} x(i)= & \frac{xr(i)}{max(|xr|)},~y(i) = \frac{yr(i)}{max(|yr|)} \end{aligned}$$
(4)
$$\begin{aligned} xr(i)= & r(i) \times sin(\theta (i)),~y(i) = r(i) \times cos(\theta (i)) \end{aligned}$$
(5)
$$\begin{aligned} \theta (i)= & a \times \pi \times rand,~r(i) =\theta (i) + R \times rand \end{aligned}$$
(6)

where \(P_{i + 1}\) is the adjacent position of the i-th individual. a is a constant value between [5, 10]. And R takes the constant within [0.5, 2].

2.4 Swoop stage

In the swoop stage, all bald eagles swing from the best position to their target prey position in search space. The polar equation is employed to model this hunting behavior. The formulas are shown below:

$$\begin{aligned} P_{i,new}= & rand \times P_{best} + x_1(i) \times (P_i - c_1 \times P_{mean}) + y_1(i) \times (P_i - c_2 \times P_{best}) \end{aligned}$$
(7)
$$\begin{aligned} x_1(i)= & \frac{xr(i)}{max(|xr|)},~y_1(i) = \frac{yr(i)}{max(|yr|)} \end{aligned}$$
(8)
$$\begin{aligned} xr(i)= & r(i) \times sinh[\theta (i)],~y(i) = r(i) \times cosh[\theta (i)] \end{aligned}$$
(9)
$$\begin{aligned} \theta (i)= & a \times \pi \times rand,~r(i) =\theta (i) \end{aligned}$$
(10)

where the \(c_1\) and \(c_2\) are in the range of [1, 2].

Algorithm 1 gives the pseudo-code of BES.

Algorithm 1
figure a

The pseudo-code of BES

3 The multi-strategy boosted BES (MBBES)

In this study, three different strategies are used to enhance the optimization ability of basic BES, including a adaptive weight parameter, two mutation operators (DE/Current-to-best/1 mutation and DE/BEST/1 mutation), and the fall stage. The fall stage is used to replace the swoop stage, which is inspired from the whale falls in nature (Changting et al. 2022).

3.1 Motivation

The original BES has an outstanding performance in seeking capability and rapid convergence. It also has a wide range of applications in the field of practical engineering. Discovered through our experiments that the ability to optimization of BES is better than many algorithms like WOA (Mirjalili and Lewis 2016), SCA (Mirjalili 2016), GWO (Mirjalili et al. 2014), DO (Zhao et al. 2022), ALO (Mirjalili 2015), MFO (Mirjalili 2015) on most test functions in CEC2014 and CEC2017. BES has the advantage of converges quickly but it hard to avoid falling into the local optimal solution. The inability to efficiently jump out the global optimal and accurately convergence emphasizes the importance to improve BES (Gharehchopogh 2024). The key is BES has potential to be promoted. We get a result through a lot of experiments that each of the three phases of BES has play its own role, the select stage help BES algorithm identify an area where an optimal solution is likely to occur, which is like the algorithm’s exploration, at this stage bald eagle group extensive search within the feasible domains. However, randomness of the select stage is not strong enough, the region where the optimal solution occurs may be missed. At search phase the bald eagle group hover at a very high altitude to search their prey in their selected area,this means that the algorithm exploitation phase has begun. However, hovering in the air may mistake a location near the prey for the prey’s location. After bald eagle group find a area where prey may appear, the hunt enters the swoop phase. At the swoop stage, bald eagle swoop down in a spiral from high to catch their prey, it is obvious that the distance traveled in a spiral is longer than that of a straight line.

3.2 Adaptive parameter

A crucial aspect that significantly impacts the performance and precision of an optimization algorithm is the equilibrium achieved between the exploration phase and the exploitation phase. The balance between these two phases will directly affect the performance of the algorithm. As the most commonly used balance parameter, inertia weight has been widely used in various optimization algorithms. In order to balance the exploration phase and exploitation phase, an adaptive parameter is applied as the new \(\alpha\), which is defined in the following formula.

$$\begin{aligned} \alpha _{new} = cos(0.5 \times pi \times (t/T)) + 1.5 \end{aligned}$$
(11)

The cos function is one of the most well-know mathematical functions used in the MHs, such as the SCA algorithm (Mirjalili 2016). The cos function has the advantage of high degree of flexibility and volatility, which enables the rate of parameter \(\alpha _{new}\) to gradually decrease with the number of iterations, as shown in Fig. 1. The original \(\alpha\) is a constant value set to 2 in the basic BES. However, The \(\alpha _{new}\) is non-linearly decreased from 2.5 to 1.5

Fig. 1
figure 1

The comparison of \(\alpha\) and \(\alpha _{new}\)

3.3 Mutation operator

Mutation operators are commonly improvement strategies with the advantages of use varied information currently obtained. Mutation strategies are equivalent to interacting location information of the present with others, which can efficient help the individuals escape the local optimal condition. Many different mutation operators were developed to meet different demand. Hu et al. (2023) used the DE/best/1/bin mutation to improve the dandelion optimizer, and achieved significant better optimization results. Zheng et al. (2022) adopted multiple mutation strategies in the ESMA to increase the diversity of basic SMA. In our proposed MBBES, two mutation operators are applied to help the original BES escape the local optimal.

3.3.1 DE/Current-to-best/1 mutation

DE/Current-to-best/1 mutation is an efficient mutation operator. The individuals will use the distance between the best position and the position of the current individual, and the distance between two random individuals to update the position. In the MBBES, the DE/Current-to-best/1 mutation is applied in the select space stage to help the bald eagle group select a better area. The equation is given in follows:

$$\begin{aligned} P_{i, DE} (t + 1) = P_i(t) + F_{1} \times (P_{best}(t) - P_{i}(t)) + F_{2} \times (P_{a}(t) - P_{b}(t)) \end{aligned}$$
(12)

The \(P_i\) is the location of the current iteration. \(P_a\) and \(P_b\) are random bald eagle individuals selected from the group. Scaling parameters \(F_1\) and \(F_2\) are the constants that control the specific gravity, which generally take the values between 0 and 1. In the MBBES, some common values of scaling parameters, such as 1.0, 0.9, 0.8 (Zhang et al. 2021), are evaluated and it is found that 0.8 is the best value for the performance of MBBES.

\(F_1\) and \(F_2\) are set to 0.8.

3.3.2 DE/BEST/1 mutation

DE/BEST/1 mutation is another popular mutation operator, the individuals will appear randomly around the best individual. The best position and the distance between two randomly selected individuals are considered in this method. In the MBBES, the DE/BEST/1 mutation is applied in the search stage to help to find prey with precision while noticing more areas where prey may appear in search stage, and increasing randomness of the algorithm. The fomula is expressed as follows.

$$\begin{aligned} P_{i, DE} (t + 1) = P_{best}(t) + F \times (P_{a} - P_{b}) \end{aligned}$$
(13)

where \(P_i\) is the location of the current iteration. \(P_{best}\) is the best position so far. \(P_a\) and \(P_b\) are the randomly selected bald eagle individuals. Scaling parameter F is a constant value which is also set to 0.8.

3.4 Fall stage

The fall stage is inspired by the whale fall in nature (Xinguang et al. 2023). Shortly after a whale’s demise, the colossal creature of the ocean commences its descent, plummeting kilometers upon kilometers until it ultimately settles on the ocean floor. Fall stage has the advantages of high randomness and mobility, that is why we use the fall stage to instead swoop stage. The mathematical model of fall stage is shown below:

$$\begin{aligned} P_{i} (t + 1) = r_5 \times P_{i} (t) - r_6 \times P_{rand} (t) + r_5 \times P_{step} \end{aligned}$$
(14)

where \(P_i(t + 1)\) represents the position of the \(i-th\) individual in the next iteration, \(P_i(t)\) represent the position of the \(i-th\) individual, \(P_{rand}\) is a random individual in this iteration. \(r_5\), \(r_6\), \(r_7\) are random number between 0 and 1. \(P_{step}\) is shown below:

$$\begin{aligned} P_{step} = (ub - lb) \times exp(-C_2 \times t/T) \end{aligned}$$
(15)

where t is the current number of iterations. T is the max number of iteration. \(C_2\) is a constant that controls the change in step size, which is related to the number of the bald eagle group.

$$\begin{aligned} C_2 = 2W_f \times N \end{aligned}$$
(16)

where \(W_f\) is calculated as below:

$$\begin{aligned} W_f = 0.12 - 0.12 \times t/T \end{aligned}$$
(17)

The process of changing the value of \(W_f\) is a process that decreases with the number of iterations. The \(C_2\) is proportional to \(W_f\) and the size of bald eagle group. so within the upper and lower bounds, \(X_{step}\) increases with the number of iterations. The step size of the algorithm increases as iteration to help jump out of the local optimal solution. \(r_5\), \(r_6\) and \(r_7\) are random numbers that can increase the searching randomness.

The flowchart of MBBES is illustrated in Fig. 2. And Algorithm 2 gives the pseudo-code of MBBES.

Fig. 2
figure 2

The flowchart of proposed MBBES

Algorithm 2
figure b

The pseudo-code of proposed MBBES

3.5 Computational time complexity of proposed MBBES

The computational time complexity is crucial for the optimization algorithms. There are three stages in the proposed MBBES, select space stage, search space stage and fall stage. The main factors of computational time complexity for the optimization algorithms are the population size N, dimension of optimization problem D, the maximum number of iterations T, and the objective function calculation time f. In the proposed MBBES, the time complexity of initialization process is \(O(N \times D)\). In the iterative process, the time complexity of selected space stage, search space stage and fall stage is \(O(N \times D \times T)\) + \(O(N \times f \times T)\). Therefore, the total time complexity for the proposed MBBES is \(O(N \times D + N \times D \times T + N \times f \times T)\), which is the same as the original BES.

4 Experimental results of the MBBES

In this section, the feasibility and effectiveness of the improvements applied on basic BES are demonstrated. The optimization performance of MBBES were evaluated using two test sets, i.e., the CEC2014 and CEC2017 (Liang et al. 2013; Wu edt al. 2017). To illustrate the optimization ability of MBBES compared with other methods, eight different and well-known optimization techniques are employed for the comparison of performance, including four basic algorithms, the BES (Hassan et al. 2019), GWO (Mirjalili et al. 2014), BO (Das and Pratihar 2022) and DO (Zhao et al. 2022), and four modified algorithms, the QRHHO (Fan et al. 2020), DETDO (Hu et al. 2023), MPSO (Bao et al. 2009) and TACPSO (Ziyu and Dingxue 2009). The setting of each algorithm can be founded in Table 2.

For a fair comparison, the number of individuals in the population and the iterations are set to 30 and 500, respectively. To obtain statistically valid results, all algorithms have been performed 30 times for each test function. In addition, all experiments were performed using the same computer. The operating system is Windows 11 with 13th Gen Intel(R) Core(TM) i5-13500 H @ 2.60 GHz and 16 GB RAM. And the code is run by using MATLAB R2021b.

Table 2 Parameter values for compared algorithms

4.1 The detailed information of the CEC2014 and CEC2017 test sets

CEC2014 and CEC2017 are two commonly used sets of test functions for performance evaluation of optimization algorithms. The detailed information also can be found in study (Abdel-Basset et al. 2023b). The primary objective of these functions is to comprehensively evaluate the convergence speed, solution accuracy, ability to escape local optima, and exploration capability of various optimization algorithms. By running optimization algorithms on these test functions, the strength and weakness of optimization algorithms for solving optimization problems can be obtained.

The details of CEC2014 and CEC2017 are given in Tables 3 and 4. The CEC2014 test suite comprises 30 test functions, encompassing various types of functions, including unimodal functions (F1–F3), simple multimodal functions (F4–F16), hybrid functions (F17–F22), and composite functions (F22–F30). For the CEC2017 test suite, it contains 29 test functions, including unimodal functions(F1, F3), multimodal functions (F4–F10), hybrid functions (F11–F20) and composite functions (F21–F30).

Table 3 The detail information of CEC2014
Table 4 The detail information of CEC2017

4.2 Comparison between MBBES and other algorithms on CEC2014

In this section, the experimental results of MBBES and other eight advanced algorithms on the CEC2014 test set are discussed. It is known that the figures have the advantages of accuracy, credibility, comparability, quantifiability and other advantages, which can be used to describe the tendency of data concentration and to compare the levels of different data. The mean value, standard deviation and ranking of the 30 test function are shown in Table 5. In the Table 5, the best figures of the average values in these nine compared algorithms under the same test function are marked in bold to make it look more intuitive. From the Table 5, it can be observed that the mean values of MBBES are smaller than BES on almost all of test functions in CEC2014, especially the F1, F3, F11, F17, F18, F21, F23, F28, F29. The average of MBBES on these test functions is significantly lower than that of BES. Moreover, according to the results of Friedman rankings (Friedman 1937), the proposed MBBES obtains first rank in CEC2014 with the smallest average rank of 1.97, which is significantly lower than those of BES and BO.

Wilcoxon signed-rank test (R. A., Nonparametric statistics for non-statisticians 2010) results are shown in Table 6. According to the 5% level of significance and combining the ranking results, MBBES is better than other algorithms on most of the test functions. The comparison results for BES, BO, DO, and GWO are 15/14/1, 16/8/6, 21/8/1 and 22/8/0. And for the QRHHO, DETDO, MPSO and TACPSO, the results are 29/1/0, 26/5/2, 20/10/0, 12/16/2.

Figure 3 shows the convergence curves of MBBES and other compared algorithms. These convergence curves clearly show that the convergence curve’s descent speed of MBBES is faster than BES and other prominent ones are used to contrast algorithms. In a certain iteration interval, the convergence curves of MBBES even has plummeting effects during the iterations, such as those of F5, F11, F16. On simple multimodal functions, MBBES performs closer to the ideal value compared to BES, with the most significant improvement observed on the F11 test function. Moreover, on most of the test functions, MBBES converges faster than other algorithms in the first 50 iterations.

Boxplots can visually show the dispersion characteristics of data. The bottom and top of the box are represented by the lower and upper quartiles, respectively. The median is located in the center of the box. The length of the box reflects the range of most of the data in the dataset, and the two whiskers (also known as the box whiskers) extend to the minimum and maximum values of the data, respectively. The boxplots of these tested algorithms on CEC2014 are presented in Fig. 4. It can be seen that the MBBES has narrower distribution ranges or smaller minimum values than those of other comparison algorithms on most of functions. In addition, significant improvements can be observed on functions F16, F22, F25, F27, F28, indicating the effectiveness of improvement strategies applied on the basic BES.

4.3 Ablation test on some functions from CEC2014

To demonstrate the effectiveness of the MBBES approach, we conducted an ablation study by testing the BES algorithm with each of our proposed strategies independently. Table 7 presents the results of BES combined with each strategy separately. BES1 refers to BES with an adaptive weight parameter, BES2 represents BES enhanced with two mutation operators (DE/Current-to-best/1 and DE/BEST/1), and BES3 involves BES where the fall stage is replaced by the swoop stage. We evaluated these variations on a selection of random functions from different categories. Specifically, F1 and F3 were chosen from unimodal functions; F4, F6, and F9 from multimodal functions; F13, F15, and F18 from hybrid functions; and F21, F22, and F23 from composition functions. Table 7 provides a detailed comparison of the results, showcasing the contribution of each strategy to the overall performance of the BES algorithm.

Table 5 Numerical results of MBBES and comparison algorithms on CEC2014
Table 6 The Wilcoxon signed-rank test results between MBBES and other algorithms on CEC2014
Table 7 Ablation experiments on some functions from CEC2014
Fig. 3
figure 3

Convergence curves of MBBES and comparison algorithms on CEC2014

Fig. 4
figure 4

Boxplots of MBBES and comparison algorithms on CEC2014

4.4 Comparison between MBBES and other algorithms on CEC2017

To further validate the superiority and competitiveness of MBBES, 29 challenging test functions of CEC2017 are selected for the optimization performance evaluation. Table 8 gives the results of mean value, standard deviation and ranking. The best results are marked in bold. It is observed that MBBES has obtained the best optimal solutions on most of the test functions, including F1, F4, F6, F8, F10, F12, F13, F16, F18, F23, F25, F27, F29. In addition, MBBES also gets good performance on other functions. And MBBES obtains better or equivalent solutions than basic BES on CEC2017 only except test function F24. In addition,the results of Friedman rankings show that the proposed MBBES obtains first rank in CEC2017 with the smallest average rank of 1.86.

The Wilcoxon signed– rank test results of CEC2017 are given in Table 9. It can be obsearved that MBBES has significant difference with other algorithms on most of the test functions. The comparison results for BES, BO, DO, and GWO are 15/14/0, 16/12/1, 21/8/0 and 23/6/0. And for the QRHHO, DETDO, MPSO and TACPSO, the results are 29/0/0, 26/6/1, 21/8/0, 14/15/0.

The Figs. 5 and 6 show the convergence curves and boxplots of CEC2017. In Fig. 5, MBBES is able to find the best solutions within fewer iterations than other comparative algorithms. From the Fig. 6, it can be seen that MBBES has the most compact boxes, indicating the stability of MBBES among all these algorithms.

Table 8 Numerical results of MBBES with comparison algorithms on CEC2017
Table 9 The Wilcoxon signed-rank test results between MBBES and other algorithms on CEC2017
Fig. 5
figure 5

Convergence curves of MBBES and comparison algorithms on CEC2017

Fig. 6
figure 6

Boxplots of MBBES and comparison algorithms on CEC2017

5 Application of MBBES on real-world engineering issues

Engineering design is a process that aims to fulfill the requirements involved in the construction of a product. The number of design variables can be extremely numerous, and their influences on the objective function seeking optimization result can be extremely complex and nonlinear. In this section, five real-world engineering issues are employed to verify the ability of MBBES, including welded beam, pressure vessel, speed reducer, three-bar truss design issue, and tubular column. And the size of population and maximum number of iterations are set to 30 and 500, respectively.

5.1 Welded beam design issue

The welded beam design problem is to find optimal variables to minimize the fabrication cost of welding materials. There are seven constraints and four variables in this problem, as shown in Fig. 7. The mathematical model is expressed in Eq. (18).

Fig. 7
figure 7

3D schematic (left) and structural parameters (right) of welded beam design problem

$$\begin{aligned} \begin{aligned}&\text {Consider variable} & \vec {z}=[z_{1},z_{2},z_{3},z_{4}]=[h,l,t,b]. \\&\underset{\vec {z}}{\text {Minimize}} & f(\vec {z})=1.10471 z_{1}^{2} z_{2} + 0.04811 z_{3} z_{4}\left( 14 + z_{2}\right) . \\&\text {Subject to} & g_{1}(\vec {z})=\tau (z)-\tau _{\max } \leqslant 0. \\ & &g_{2}(\vec {z})=\sigma (z)-\sigma _{\max } \leqslant 0. \\ & &g_{3}(\vec {z})=z_{1}-z_{4} \leqslant 0. \\ & &g_{4}(\vec {z})=0.10471 z_{1}^{2} + 0.04811 z_{3} z_{4}\left( 14 + z_{2}\right) -5 \leqslant 0. \\ & &g_{5}(\vec {z})=0.125-z_{1} \leqslant 0. \\ & &g_{6}(\vec {z})=\delta (z)-\delta _{\max } \leqslant 0. \\ & &g_{7}(\vec {z})=P-P_{c}(z) \leqslant 0. \\&\text {Variable range} & 0.1 \leqslant z_{1},z_{4} \leqslant 2. \\ & &0.1 \leqslant z_{2},z_{3} \leqslant 10. \\&\text {Where} & \tau (z)=\sqrt{\left( \tau ^{\prime }\right) ^{2} + 2 \tau ^{\prime } \tau ^{\prime \prime } \frac{z_{2}}{2 R} + \left( \tau ^{\prime \prime }\right) ^{2}}. \\ & &\tau ^{\prime }=\frac{P}{\sqrt{2} z_{1} z_{2}},\quad \tau ^{\prime \prime }=\frac{M R}{J}. \\ & &M=P\left( L + \frac{z_{2}}{2}\right) . \\ & &R=\sqrt{\frac{z_{2}^{2}}{4} + \left( \frac{z_{1} + z_{3}}{2}\right) ^{2}}. \\ & &J=2\left\{ \sqrt{2} z_{1} z_{2}\left[ \frac{z_{2}^{2}}{12} + \left( \frac{z_{1} + z_{3}}{2}\right) ^{2}\right] \right\} . \\ & &\sigma (z)=\frac{6 P L}{z_{4} \chi _{3}^{2}}, \quad \delta (z)=\frac{4 P L^{3}}{E z_{3}^{3} z_{4}}. \\ & &P_{c}(z)=\frac{4.013 E \sqrt{\frac{z^{2} z^{5}}{36}}}{L^{2}}\left( 1-\frac{z_{3}}{2 L} \sqrt{\frac{E}{4 G}}\right) . \\ & &P=6000 lb, \quad L=14 in, \quad E=30 \times 10^{6} \textrm{psi}, \quad G=12 \times 10^{6} \textrm{psi}, \\ & &\tau _{\max }=13,600 ps, \quad \sigma _{\max }=30,000 \textrm{psi}, \quad \delta _{\max }=0.25 in. \\ \end{aligned} \end{aligned}$$
(18)
Table 10 Comparison of optimization results for welded beam design problem

Table 10 gives the optimal solutions of MBBES and other algorithms when solving the welded beam design problem. MBBES outperforms other algorithms and achieves the lowest cost with 1.6952. It shows that MBBES has the merits in dealing with this problem.

5.2 Pressure vessel design issue

Pressure vessel design issue is one of the most common engineering design problems, with the objection of minimizing the manufacturing cost of cylindrical pressure vessels while meeting the required pressure standards. As shown in Fig. 8, there are four key structural parameters: shell thickness (\(T_s\)), head thickness (\(T_h\)), inner radius (R), and length of the headless cylindrical section (L). The mathematical formulas are shown in Eq. (19).

Table 11 lists the results of pressure vessel design issue. MBBES obtains the lowest optimal value of 5734.9131, which reveals that MBBES is superior to other competing algorithms.

Fig. 8
figure 8

3D schematic (left) and structural parameters (right) of pressure vessel problem

$$\begin{aligned} \begin{aligned}&\text {Consider variable} & \vec {x}=[y_{1},y_{2},y_{3},y_{4}]=[T_{s},T_{h},R,L].\\&\underset{\vec {y}}{\text {Minimize}} & f(\vec {y})=0.6224 y_{1} y_{3} y_{4} + 1.7781 y_{2} y_{3}^{2} + \\ & &3.1661 y_{1}^{2} y_{4} + 19.84 y_{1}^{2} y_{3}.\\&\text {Subject to} & g_{1}(\vec {y})=-y_{1} + 0.0193 y. \\ & &g_{2}(\vec {y})=-y_{2} + 0.00954 y_{3} \leqslant 0. \\ & &g_{3}(\vec {y})=-\pi y_{3}^{2} y_{4}-4 / 3 \pi y_{3}^{3} + 1296,000 \leqslant 0. \\ & &g_{4}(\vec {y})=y_{4}-240 \leqslant 0. \\&\text {Variable range} & 0 \leqslant y_{1},y_{2} \leqslant 99.\\ & &10 \leqslant y_{3},y_{4} \leqslant 200.\\ \end{aligned} \end{aligned}$$
(19)
Table 11 Comparison of optimization results for pressure vessel design problem

5.3 Speed reducer design issue

The target of speed reducer design issue is to obtain the lowest weight of mechanical device with eleven constraints. As shown in Fig. 9, there are seven variables that need to be optimized. The mathematical model is given in Eq. (20).

As shown in Table 12, MBBES obtains the minimum weight of 2995.4373. According to the results, it can be seen that MBBES has good ability to tackle this problem.

Fig. 9
figure 9

3D schematic (left) and structural parameters (right) of reducer design problem

$$\begin{aligned} \begin{aligned}&\text {Consider variable} & \vec {y}=[y_{1},y_{2},y_{3},y_{4},y_{5},y_{6},y_{7}]=\\ & &[b,m,z,l_{1},l_{2},d_{1},d_{2}].\\&\underset{\vec {y}}{\text {Minimize}} & f(\vec {y}) = \\ & &0.7854 y_{1} y_{2}^{2}\left( 3.3333 y_{3}^{2} + 14.9334 y_{3}-43.0934\right) \\ & &-1.508 y_{1}\left( y_{6}^{2} + y_{7}^{2}\right) + 7.4777\left( y_{6}^{3} + y_{7}^{3}\right) \\ & & + 0.7854\left( y_{4} y_{6}^{2} + y_{5} y_{7}^{2}\right) . \\&\text {Subject to} & g_{1}(\vec {y})=\frac{27}{y_{1} y_{2}^{2} y_{3}}-1 \leqslant 0, \quad \\ & &g_{2}(y)=\frac{397.5}{y_{1} y_{2}^{2} y_{2}^{2}}-1 \leqslant 0\\ & &g_{3}(\vec {y})=\frac{1.93 y_{4}^{3}}{y_{2} y_{6}^{4} y_{3}}-1 \leqslant 0, \quad \\ & &g_{4}(y)=\frac{1.93 y_{5}^{3}}{y_{2} y_{7}^{4} y_{3}}-1 \leqslant 0\\ & &g_{5}(\vec {y})=\frac{\left[ \left( 745 y_{4} / y_{2} y_{3}\right) ^{2} + 16.9 \times 10^{6}\right] ^{1 / 2}}{110 y^{3}}-1 \leqslant 0. \\ & &g_{6}(\vec {y})=\frac{\left[ \left( 745 y_{5} / y_{2} y_{3}\right) ^{2} + 157.5 \times 10^{6}\right] ^{1 / 2}}{85 y_{7}^{3}}-1 \leqslant 0. \\ & &g_{7}(\vec {y})=\frac{y_{2} y_{3}}{40}-1 \leqslant 0, \quad \\ & &g_{8}(y)=\frac{5 y_{2}}{y_{1}}-1 \leqslant 0\\ & &g_{9}(\vec {y})=\frac{y_{1}}{12 y_{2}}-1 \leqslant 0, \quad \\ & &g_{10}(y)=\frac{1.5 y_{6} + 1.9}{y_{4}}-1 \leqslant 0\\ & &g_{11}(\vec {y})=\frac{1.1 y_{7} + 1.9}{y_{5}}-1 \leqslant 0. \\&\text {Where} & 2.6 \leqslant y_{1} \leqslant 3.6, \quad 0.7 \leqslant y_{2} \leqslant 0.8, \quad 17 \leqslant y_{3} \leqslant 28,\\ & &7.3 \leqslant y_{4} \leqslant 8.3, \quad 7.3 \leqslant y_{5} \leqslant 8.3, \quad 2.9 \leqslant y_{6} \leqslant 3.9,\\ & &5.0 \leqslant y_{7} \leqslant 5.5. \\ \end{aligned} \end{aligned}$$
(20)
Table 12 Comparison of optimization results for reducer design problem

5.4 Three-bar truss design issue

The goal of three-bar truss design issue is to minimize its weight (Sadollah et al. 2013), as shown in Fig. 10. Two variables with three constraints need to be optimized. The detailed formulas of this problem is given in Eq. (21).

Table 13 shows the optimal results obtained by MBBES and other algorithms. According to the results, MBBES obtains the minimum weight with 263.8523, indicating that MBBES is superior to other comparison algorithms.

Fig. 10
figure 10

3D schematic (left) and structural parameters (right) of three-bar truss design problem

$$\begin{aligned} \begin{aligned}&\text {Consider variable} & \vec {y}=[y_{1},y_{2}]=[A_{1},A_{2}].\\&\underset{\vec {y}}{\text {Minimize}} & f(\vec {y})= \left( 2\sqrt{2}y_{2} + y_{2}\right) \times l.\\&\text {Subject to} & g_{1}(\vec {y})=\frac{\sqrt{2}y_{1} + y_{2}}{\sqrt{2} y_{1}^{2} + 2y_{1} y_{2}}P-\sigma \leqslant 0. \\ & &g_{2}(\vec {y})=\frac{y_{2}}{\sqrt{2} y_{1}^{2} + 2y_{1} y_{2}}P-\sigma \leqslant 0. \\ & &g_{3}(\vec {y})=\frac{y_{2}}{y_{1} + \sqrt{2} y_{2}}P-\sigma \leqslant 0. \\&\text {Variable range} & 0 \leqslant y_{1},~y_{2} \leqslant 1.\\&\text {Where} & l=10 cm,\quad P=2 KN/cm^{2},\quad \sigma =2 KN/cm^{2}. \\ \end{aligned} \end{aligned}$$
(21)
Table 13 Comparison of optimization results for three-bar truss design problem

5.5 Tubular column design issue

The tubular column design problem requires to design uniform columns withstand compressive load while minimizing both material and construction costs (Hu et al. 2023), as shown in Fig. 11. There are two variables in this problem: average diameter of column d and thickness of tube T. The model is given in Eq. (22).

As shown in Table 14, the cost obtained by MBBES is the lowest, which is 26.5313. Therefore, MBBES has better optimization performance for solving this problem.

Fig. 11
figure 11

Structural parameters of tubular column design problem

$$\begin{aligned} \begin{aligned}&\text {Minimize} & f(d, t)= 9.8dt + 2d\\&\text {Subject to} & g_{1} =\frac{P}{\pi dt \sigma _y} - 1 \leqslant 0 \\ & &g_{2}=\frac{9PL^{2}}{\pi ^{3}Edt(d^2 + t^2)} - 1 \leqslant 0 \\ & &g_{3}=\frac{2.0}{d} - 1 \leqslant 0 \\ & &g_{4}=\frac{d}{14} - 1 \leqslant 0 \\ & &g_{5}=\frac{0.2}{t} - 1 \leqslant 0 \\ & &g_{6}=\frac{t}{0.8} - 1 \leqslant 0 \\&\text {Variable range} & 0.01 \leqslant d, T \leqslant 100\\ \end{aligned} \end{aligned}$$
(22)
Table 14 Comparison of optimization results for tubular column design problem

6 MLP classification problems

Multi-layer perceptron (MLP) is a particular type of feedforward neural networks that contains only one hidden layer (Bansal et al. 2019). The core idea of MLP is to learn different complex patterns by adjusting the number of neurons in the hidden layer and how they connect. Neurons in the hidden layer perform nonlinear transformations of the input data through activation functions, thereby increasing the representation power of the model. MLP is a supervised learning model that is typically trained using a backpropagation algorithm. Through continuous iterative training, MLP can automatically learn the complex relationships between input features and make predictions on new data.

To examine the wide advantage of MBBES, in this section, we show the three results for MLP classification problems, including MLP_XOR dataset problem, MLP_Iris dataset problem and MLP_Cancer dataset problem.

6.1 MLP_XOR dataset

The MBBES and the comparison algorithms are used to solve the MLP_XOR classification problem. The results obtained are shown in Table 15, which is obviously shown that MBBES has the best classification rate, classify 100% correctly. Figure 12 shows the convergence curves of these algorithms on this MLP_XOR classification problem. It is observed that MBBES has the lowest fitness compared with other algorithms. In the Fig. 13, we also can learn that the MBBES has very good robustness when solving this classification problem.

Table 15 Comparison optimization results of MLP_XOR classification problem
Fig. 12
figure 12

The convergence curves of MLP_XOR dataset problem

Fig. 13
figure 13

The variance diagram of MLP_XOR dataset problem

6.2 MLP_Iris dataset

As shown in Table 16, the proposed MBBES gets a 100% classification rate. In the aspect of best and average values, MBBES obtains the first rank with 1.16E– 03 and 8.65E– 03, respectively. As shown in Fig. 14, the MBBES exhibits excellent optimization ability to find the global solution. Figure 15 illustrates the good stability of MBBES when solving the MLP_Iris classification problem.

Table 16 Comparison optimization results of MLP_Iris classification problem
Fig. 14
figure 14

The convergence curves of MLP_Iris dataset problem

Fig. 15
figure 15

The variance diagram of MLP_Iris dataset problem

6.3 MLP_Cancer problem

Table 17 presents the outcomes of MBBES and compared algorithms for the MLP_Cancer classification problem. It can be seen that MBBES has the best classification rate with 99%, which is also obtained by the TACPSO. It’s also noted that MBBES has the lowest mean value. Figure 16 shows the optimizing process for these methods. It can be seen that MBBES performs the best with lowest final fitness value. Figure 17 shows the optimization stability of these algorithms. It can be observed that the MBBES algorithm also has the merits on robustness when solving this problem.

Table 17 Comparison optimization results of MLP_Cancer classification problem
Fig. 16
figure 16

The convergence curves of MLP_Cancer dataset problem

Fig. 17
figure 17

The variance diagram of MLP_Cancer dataset problem

7 Conclusions and future outlook

BES has been utilized to solve complex optimization problems across a diverse range of fields. However, the basic BES still has a likelihood of getting stuck in a local optimal stagnation and failing to attain the satisfied global optimal solution. To handle this problem, we proposed a novel multi-strategy boosted BES (MBBES), which includes the modifications of parameter \(\alpha\), two mutation operators and fall stage strategy. The modified adaptive parameter \(\alpha\) strikes a good balance between the exploration and exploitation stages of the algorithm. Two mutation operators cooperate well, helping the algorithm escape from local optima as well as enhancing its convergence. Moreover, a fall stage is introduced instead of the swoop stage, which can accelerate the convergence speed of the basic BES. The performance of proposed MBBES is demonstrated using two sets of test functions CEC2014 and CEC2017. The numerical results, convergence curves, boxplots, Friedman ranking test and Wilcoxon signed-rank test show that the proposed MBBES is better than the other compared algorithms, including the BO, DO, GWO, QRHHO, DETDO, MPSO, and TACPSO. Moreover, the MBBES also has been employed to tackle five realistic constrained issues, including the welded beam design, pressure vessel design, reducer design, three-bar truss design and tubular column design, as well as three MLP classification problems, namely MLP_XOR, MLP_Iris and MLP_Cancer. According to the optimization results, the proposed MBBES has demonstrated its superior performance for solving these optimization problems in practise.

However, it also should be noted that the proposed MBBES is not good at solving some test functions in CEC2014 and CEC2017, such as F29 in CEC2014 and F30 in CEC2017.

In the future, the MBBES can be further strengthened by introducing more effective strategies to avoid falling into the local optimal and get closer to the theoretical results. It can be also compared with some recent algorithms such as GAEFA-HK (Chauhan et al. 2024), iAEFA (Chauhan and Yadav 2023), and others. Meanwhile, the MBBES can be used to solve other real-world complex optimization problems as well as develop a parallel computing algorithm that leverages MBBES and parallel theory to tackle intricate optimization tasks.