iconOpen Access

ARTICLE

crossmark

Enhanced Harris Hawks Optimization Integrated with Coot Bird Optimization for Solving Continuous Numerical Optimization Problems

Hao Cui, Yanling Guo*, Yaning Xiao, Yangwei Wang*, Jian Li, Yapeng Zhang, Haoyu Zhang

College of Mechanical and Electrical Engineering, Northeast Forestry University, Harbin, 150040, China

* Corresponding Authors: Yanling Guo. Email: email; Yangwei Wang. Email: email

(This article belongs to the Special Issue: Bio-inspired Computer Modelling: Theories and Applications in Engineering and Sciences)

Computer Modeling in Engineering & Sciences 2023, 137(2), 1635-1675. https://doi.org/10.32604/cmes.2023.026019

Abstract

Harris Hawks Optimization (HHO) is a novel meta-heuristic algorithm that imitates the predation characteristics of Harris Hawk and combines Lévy flight to solve complex multidimensional problems. Nevertheless, the basic HHO algorithm still has certain limitations, including the tendency to fall into the local optima and poor convergence accuracy. Coot Bird Optimization (CBO) is another new swarm-based optimization algorithm. CBO originates from the regular and irregular motion of a bird called Coot on the water’s surface. Although the framework of CBO is slightly complicated, it has outstanding exploration potential and excellent capability to avoid falling into local optimal solutions. This paper proposes a novel enhanced hybrid algorithm based on the basic HHO and CBO named Enhanced Harris Hawks Optimization Integrated with Coot Bird Optimization (EHHOCBO). EHHOCBO can provide higher-quality solutions for numerical optimization problems. It first embeds the leadership mechanism of CBO into the population initialization process of HHO. This way can take full advantage of the valuable solution information to provide a good foundation for the global search of the hybrid algorithm. Secondly, the Ensemble Mutation Strategy (EMS) is introduced to generate the mutant candidate positions for consideration, further improving the hybrid algorithm’s exploration trend and population diversity. To further reduce the likelihood of falling into the local optima and speed up the convergence, Refracted Opposition-Based Learning (ROBL) is adopted to update the current optimal solution in the swarm. Using 23 classical benchmark functions and the IEEE CEC2017 test suite, the performance of the proposed EHHOCBO is comprehensively evaluated and compared with eight other basic meta-heuristic algorithms and six improved variants. Experimental results show that EHHOCBO can achieve better solution accuracy, faster convergence speed, and a more robust ability to jump out of local optima than other advanced optimizers in most test cases. Finally, EHHOCBO is applied to address four engineering design problems. Our findings indicate that the proposed method also provides satisfactory performance regarding the convergence accuracy of the optimal global solution.

Graphic Abstract

Enhanced Harris Hawks Optimization Integrated with Coot Bird Optimization for Solving Continuous Numerical Optimization Problems

Keywords


1  Introduction

The optimization process refers to determining the best solution for a given problem. Optimization problems are widespread in the contemporary military, engineering, and management fields. Optimization methods have become an essential computational tool for technicians in different specialties [14]. With the evolution of human society and technology, more complex optimization problems are emerging. However, mathematical-programming optimization algorithms cannot effectively solve these problems, such as the Gradient Method and Newton’s Method. These conventional methods need help to locate the optimal global solution in large-scale, high-dimensional, and sub-optimal search domains [46]. To accomplish more powerful optimization tools, in recent decades many scholars have embarked on the research of meta-heuristic algorithms (MAs) [7]. MAs are a combination of stochastic algorithms and local search algorithms. The design of MAs is mainly inspired by a series of random phenomena in nature [8,9]. For MAs, they are concept-simple, require few parameter settings, and do not have any special requirements for the objective function. Consequently, MAs are not limited to specific problems and can be used in many applications [10,11]. MAs usually includes four categories [1214]: evolutionary-based algorithms, physics-based algorithms, swarm-based algorithms (social behaviors of organisms in nature), and human-based algorithms. Some famous algorithms are Genetic Algorithm (GA) [15], Differential Evolution (DE) [16], Evolution Strategy (ES) [17], Simulated Annealing (SA) [18], Multi-Verse Optimizer (MVO) [19], Atom Search Optimization (ASO) [20], Arithmetic Optimization Algorithm (AOA) [21], Particle Swarm Optimization (PSO) [22], Sine Cosine Algorithm (SCA) [23], Sooty Tern Optimization Algorithm (STOA) [24], Chimp Optimization Algorithm (ChOA) [25], Aquila Optimizer (AO) [26], Dragonfly Algorithm (DA) [27], Butterfly Optimization Algorithm (BOA) [28], Slime Mould Algorithm (SMA) [29], Whale Optimization Algorithm (WOA) [30], Grey Wolf Optimizer (GWO) [31], Seagull Optimization Algorithm (SOA) [32], Gorilla Troops Optimizer (GTO) [33], Search Group Algorithm (SGA) [34], etc. Although hundreds of MAs have been proposed since the Artificial Intelligence (AI) age, the No Free Lunch (NFL) [35] theorem states that there is no single algorithm capable of tackling all optimization problems. Hence, further innovation in MAs is indispensable. At the same time, most MAs still need help with slow convergence, difficulty getting rid of local optimization, and imbalance between exploration and exploitation stages [36]. In addition to developing new MAs, several researchers try to boost the performance of existing algorithms by incorporating various search operators and achieving promising results. Nguyen et al. [13] proposed an improved Slime Mould Algorithm and applied it to control stepped hydropower plants. Debnath et al. [37] hybridized the Dragonfly Algorithm with Differential Evolution to better solve the problem of maximizing secondary user throughput in energy-harvesting cognitive radio networks. Ziyu et al. [38] proposed an improved Particle Swarm Optimization algorithm named TACPSO by introducing acceleration factors and random speeds. Li et al. [39] combined Differential Evolution and Hunger Games Search into a new DECEHGS algorithm, validated on the CEC2014 test suite, CEC2017 test suite, and four engineering problems. Jia et al. [40] constructed an improved Condor Search Algorithm based on the Lévy flight and simulated annealing mechanisms. All the above-improved variants are proven to outperform the basic algorithm to some extent, which indicates that it is feasible to enhance the performance of algorithms by introducing some additional search strategies.

The Harris Hawks Optimization (HHO) simulates the cooperative behavior and chasing style of Harris hawks in nature [41]. HHO is easy to implement with few parameters and performs well in solving many optimization problems. However, the basic HHO algorithm has several limitations, including the tendency to fall into the local optima and poor convergence accuracy. Its insufficient exploration phase causes this. Research on the improvement of HHO has been in full swing. For example, Fan et al. [42] proposed a quasi-reflective Harris Hawks Algorithm (QRHHO), enhancing the native algorithm’s exploration capability and convergence speed. Dokeroglu et al. [43] presented an island parallel HHO (IP-HHO) version of the algorithm for optimizing continuous multi-dimensional problems. Wang et al. [44] hybridized the exploration phase of Aquila Optimizer and the exploitation phase of Harris Hawks Optimization to preserve the powerful search abilities of each in both algorithms. They presented a novel improved optimizer, namely IHAOHHO. Another algorithm discussed in this paper is Coot Bird Optimization (CBO), proposed by Naruei et al. [45] in 2021. The CBO algorithm mimics four movement patterns of coot birds on the water surface, including random movement, chain movement, location adjustment according to the group leaders, and leader movement. In this algorithm, the Coot leaders continuously lead the population toward the goal. The leaders sometimes have to leave their current positions to find better target areas, which gives the algorithm have strong global ability to explore different parts of the search space. But the overall framework of CBO is relatively complex compared to other algorithms, and its exploitation stage needs to be revised, often leading to premature convergence. Considering this, mixing HHO and CBO can compensate for the lack of HHO exploration capability and thus improve its global search performance. The hybrid algorithm can effectively use the solution information in the search space to maintain population diversity later in the search. Currently, some improved variants of the CBO-based algorithm have achieved promising performance. Huang et al. [46] developed a COOTCLCO algorithm that provided satisfactory solutions for solving the uneven distribution and low coverage problems of randomly deployed Wireless Sensor Network (WSN) nodes.

Motivated by the NFL theory and fully considering the characteristics of the basic HHO and CBO algorithms: HHO has good exploitation properties, but its exploration stage is insufficient, and conversely, CBO owns excellent exploration capability benefiting from the leader mechanism, but the algorithm itself is complex and lacks exploitation ability. Therefore, this paper attempts to combine the basic HHO and CBO algorithms by taking advantage of each and proposes a novel hybrid meta-heuristic algorithm with better all-around search performance for global optimization, called EHHOCBO. First, integrate the leadership mechanism of CBO into the population initialization stage of HHO. In the leader movement, the optimal individual guides the whole population to explore all regions of the search space as much as possible in a circular envelope so that the algorithm has good randomness and rich population diversity. It can make up for the lack of helpful information exchange between the optimal individual and other individuals in the swarm, thus laying a good foundation for the global search of HHO. Then, during the iterative computation, Ensemble Mutation Strategy (EMS) [47] is introduced to generate the variant candidate positions of the current individual further to enhance the algorithm exploration capability and convergence accuracy. Moreover, to extend the search range and avoid the algorithm from falling into the local optima in the later search phase, Refracted Opposition-Based Learning (ROBL) [48] is used to evaluate the fitness value of the inverse position of the current optimal solution and perform the update. To assess the performance of the proposed EHHOCBO, we used 23 classical benchmark functions and the IEEE CEC2017 test suite for testing and applied the proposed algorithm to address four engineering problems. Experimental results demonstrate that EHHOCBO performs better than competitors concerning solution accuracy, convergence speed, stability, and local optima avoidance.

This paper is organized as follows: Section 2 briefly introduces the basic HHO and CBO algorithm and two improved search strategies. In Section 3, the proposed hybrid EHHOCBO optimizer is described in detail. In Section 4, we evaluate the performance of EHHOCBO through a series of simulation experiments and analyze the results obtained. Based on this, the proposed method addresses four real-world engineering design problems to highlight its practicality in Section 5. Finally, a conclusion of this paper is provided in Section 6.

2  Related Works

2.1 Harris Hawks Optimization (HHO)

Harris Hawks Optimization (HHO) was proposed by Heidari et al. [41] in 2019, which mimics the predatory behavior of the Harris Hawk, a species of raptor living in southern Arizona, U.S.A. The HHO algorithm consists of three main components: the exploration phase, the exploitation phase, and the transition from exploration to exploitation. Fig. 1 shows the different search processes of the basic HHO algorithm.

images

Figure 1: Different phases of HHO

2.1.1 Exploration Phase

During this phase, the Harris hawks randomly perch in certain places. They detect the prey with keen vision and then choose between two strategies with the same probability of undertaking hunting activities. The formula for the position update of Harris hawk in this phase is as follows.

X(t+1)={Xrand(t)r1|Xrand(t)2r2X(t)|,q0.5(Xprey(t)Xm(t))r3(LB+r4(UBLB)),q<0.5(1)

where X(t) and X(t+1) represent the position vector of the Harris hawk in the current and next iterations. t represents the current number of iterations. Xprey is the prey position, which is also regarded as the optimal solution. Xrand(t) is the position vector of the random individual in the current population. r1, r2, r3, r4, and q are random numbers between [0,1]. LB and UB are the lower and upper bounds of variables. Xm(t) represents the average position of all hawks in the population, which is calculated as follows:

Xm(t)=1Ni=1NXi(t)(2)

where N is the population size. Xi(t) is the current position vector of the i-th hawk.

2.1.2 Transition from Exploration to Exploitation

In the HHO algorithm, the prey’s escape energy E drives the algorithm to switch between the global exploration and local exploitation phases. The prey’s energy decreases gradually throughout the escaping process, which can be simulated as in Eq. (3).

E=2E0(TtT)(3)

where E0 is a random number between [1,1], representing the initial state of the escaping energy of the prey. T is the maximum number of iterations. When |E|1, the hawk will continue searching for the location of prey in the target area, defined as the exploration phase. In the case of |E|<1, the hawk will start to hunt the prey found in the earlier step and enter the exploitation stage.

2.1.3 Exploitation Phase

Four possible strategies exist in the exploitation phase, including soft besiege, hard besiege, soft besiege with progressive rapid dives, and hard besiege with progressive rapid dives, to simulate the attack process of Harris hawk on its prey. r represents the probability of whether the prey can escape the danger before the Harris hawk attack, which is a random number between [0,1]. If r<0.5 means that the prey has successfully run through the the dangerous situation, r0.5 denotes the case of unsuccessful escape. Different combinations of r-value and escape energy E correspond to different predation strategies. When |E|<0.5, a hard siege is performed. Otherwise, a soft siege is conducted.

•   Soft Besiege

The soft besiege is performed when r0.5 and |E|0.5. At this stage, the position of Harris hawk is updated as follows:

X(t+1)=ΔX(t)E|JXprey(t)X(t)|(4)

ΔX(t)=Xprey(t)X(t)(5)

J=2(1r5)(6)

where ΔX(t) denotes the distance between the hawk’s position and the prey. r5 is the random number between [0,1]. J indicates the random jump intensity of the prey.

•   Hard Besiege

The hawk will take a hard besiege when r0.5 and |E|<0.5. The mathematical formula for this behavior is as follows:

X(t+1)=Xprey(t)E|ΔX(t)|(7)

•   Soft Besiege with Progressive Rapid Dives

When r<0.5 and |E|0.5, the hawk will take a soft besiege with progressive rapid dives. Lévy flight is integrated into the HHO algorithm, and the mathematical model of the above behavior is as follows:

Y=Xprey(t)E|JXprey(t)X(t)|(8)

Z=Y+S×LF(D)(9)

X(t+1)={Y,if F(Y)<F(X(t))Z,if F(Z)<F(X(t))(10)

where D is the dimension of problem. S is a random vector whose size is 1×D. F() is the objective function. Only the better position between Y and Z is selected as the next position. LF() is the Lévy flight function, which is calculated as follows:

LF(x)=0.01×u×σ|v|1β,σ=(Γ(1+β)×sin(πβ2)Γ(1+β)×β×2(β12))1/β(11)

where u and v are two random numbers between [0,1]. β is a constant with a fixed value of 1.5. Γ() is the gamma function.

•   Hard Besiege with Progressive Rapid Dives

When r<0.5 and |E|<0.5, the Harris hawk will perform a hard siege to get close to the prey and then make a surprise attack. The mathematical model of this behavior is simulated as follows:

Y=Xprey(t)E|JXprey(t)Xm(t)|(12)

Z=Y+S×LF(D)(13)

X(t+1)={Y,if F(Y)<F(X(t))Z,if F(Z)<F(X(t))(14)

where Xm(t) is calculated by Eq. (2). Note that only the better position between Y and Z is selected as the next position. The pseudo-code of the basic HHO algorithm is shown in Algorithm 1.

images

2.2 Coot Bird Optimization(CBO)

Coot Bird Optimization (CBO) is a bio-inspired, population-based, and gradient-free optimization technique developed by Naruei et al. [45] in 2021, which mimics the collective behaviors of American Coots (a small water bird) on the water surface. In the CBO, there are four different irregular and regular movements implemented: (1) Random movement, (2) Chain movement, (3) Adjusting the position based on the group leaders, and (4) Leader movement. The mathematical model of the algorithm is presented below.

Usually, the Coots live in a group and make a chain structure to move toward the target area (food). In front of the group are a few coots, also known as group leaders, who guide the direction and take charge of the whole flock. Therefore, according to the habits of Coots, the initial population is divided into two parts: the leader Coots and the follower Coots. If N is the population size, the number of Coot leaders is calculated as a percentage of the total population equal to L, and the remaining members (NL) are considered Coot followers. It is noted that the leaders are all selected from the population randomly. Then the mentioned four movements are implemented.

2.2.1 Random Movement

In this stage, the random position Q is defined using Eq. (15). The Coot followers are moved towards this random position to explore various parts of the search domain.

Q=rand(1,D).(UBLB)+LB(15)

where D is the dimension of problem. LB and UB are the lower and upper bounds of variables. The random movement gives the algorithm better global search performance and strengthens the algorithm’s capability to escape from the local optima. The Coot’s new position is updated as follows:

Xi(t+1)=Xi(t)+A×r6×(QXi(t))(16)

where Xi(t+1) denotes the position of the i-th follower in the next iteration t, r6 is a random number in the range of [0,1], and the parameter A is calculated according to Eq. (17).

A=1tT(17)

where t represents the number of current iterations and T is the maximum iterations.

2.2.2 Chain Movement

In the Salp Swarm Algorithm (SSA), the average position of two individuals is used to perform chain movements. The same approach is employed in CBO. The new position of the Coot follower is calculated as follows:

Xi(t+1)=12×(Xi1(t)+Xi(t))(18)

where Xi1(t) is the position of the (i1)-th follower in the current iteration t.

2.2.3 Adjust the Position Based on the Group Leaders

Generally, the whole group is led by some of the group leaders in the front, and all the remaining coot followers need to adjust their position based on the leaders and move towards them. However, a severe issue that must be addressed is that each Coot should update its position according to which leader. Eq. (19) is designed to select the leader as follows:

k=1+(i MOD L)(19)

where i is the index number of the current follower, L denotes the number of leaders, and k stands for the leader’s index number.

The next position of the Coot follower based on the selected leader k is calculated using Eq. (20).

Xi(t+1)=LeaderXk(t)+2×r7×cos(2Rπ)×(LeaderXk(t)Xi(t))(20)

where LeaderXk(t) is the position of the selected leader, r7 is a random number within the interval [0,1], and R denotes a random number within the interval [1,1].

2.2.4 Leader Movement

The group must be oriented towards the optimal area, so in some instances, the leaders have to leave the current optimal position to search for a better one. The formula for updating the leader position is given as follows:

LeaderXi(t+1)={B×r8×cos(2Rπ)×(gBest(t)LeaderXi(t))+gBest(t),r9<0.5B×r8×cos(2Rπ)×(gBest(t)LeaderXi(t))gBest(t),r90.5(21)

In Eq. (21), gBest denotes the current optimal position, r8 and r9 are random numbers within the interval [0,1], and R is a random number within the interval [1,1]. B×r8 generates a more significant stochastic movement to help the algorithm eliminate the local optimal solutions. And cos(2Rπ) is designed to search for the best individual with different radius to obtain a superior position. The value of B is calculated using Eq. (22).

B=2t×(1T)(22)

where t represents the number of current iterations, and T denotes the maximum. The pseudo-code of Coot Bird Optimization (CBO) is summarized in Algorithm 2.

images

2.3 Ensemble Mutation Strategy (EMS)

The set variation strategy is a new improvement mechanism proposed by Zheng et al. [8,47], which can generate diverse individuals to improve the global search capability of the hybrid algorithm. The mathematical formula of EMS is as follows:

Vi1={XR1+F1×(XR2XR3),r10<C1Xi,r10C1(23)

Vi2={XR4+F2×(XR5XR6)+F2×(XR7XR8),r11<C2Xi,r11C2(24)

Vi3={Xi+F3×(XR9Xi)+F3×(XR10XR11),r12<C3Xi,r12C3(25)

where Vi1, Vi2, and Vi3 are the newly generated mutant positions of the ith search agent. R1R11 are different integer exponents in the range [1,N]. F1, F2, and F3 represent scale factors with values of 1.0, 0.8, and 1.0, respectively. r10r12 are random numbers in the range of [0,1]. In addition, the parameters C1, C2, and C3 are equal to 0.1, 0.2, and 0.9, which denote the crossover rate.

After the mutant candidate positions Vi1, Vi2, and Vi3 are generated, the best position Vi with the lowest fitness value will be selected to compare with the fitness of the original post Xi, and then the better one will be saved as the new Xi to participate in the following iterative calculation. These processes can be described using Eq. (26).

Xi={Vi,if F(Vi)<F(Xi)Xi,otherwise(26)

where F() is the cost function.

2.4 Refracted Opposition-Based Learning

Refracted Opposition-based Learning [48] is an improved variant of Opposition-based Learning (OBL) [49]. ROBL introduces the principle of refraction of light to generate dynamic inverse solutions based on OBL, which solves the limitation that OBL can easily fall into local optimum in later iterations. Currently, ROBL has been used to effectively improve the optimization performance of the Grey Wolf Optimizer (GWO) [31]. The schematic diagram of ROBL is shown in Fig. 2, and its mathematical model is formulated as follows:

Xi,j=aj+bj2+aj+bj2zηXi,jzη,j=1,2,,D(27)

images

Figure 2: Refracted opposition-based learning

where Xi,j is the location of the ith search agent in the jth dimension. Xi,j is the inverse solution of Xi,j. aj and bj are the upper and lower limits in the jth dimension. D deontes the dimension of problem. η is the refraction index, which can be calculated using Eq. (28). z indicates the scale factor, which can be calculated using Eq. (29).

z=hh(28)

η=sinθ1sinθ2(29)

where h and h are the lengths of the incident light and the refracted light. θ1 and θ2 are the angle of incidence and refraction, respectively.

3  The Proposed EHHOCBO Algorithm

This section integrates the basic HHO and CBO algorithms into a new swarm intelligence algorithm called EHHOCBO. EHHOCBO takes HHO as the core framework and is hybridized with the powerful exploratory leadership mechanism of hybrid CBO. In addition, ensemble mutation strategy (EMS) and refracted opposition-based learning (ROBL) are introduced into the preliminary hybrid algorithm to enhance its search performance further.

According to Section 2, the basic HHO algorithm consists of three main steps: the exploration phase, the transition from exploration to exploitation, and the exploitation phase. In the exploration phase, Harris hawk detects prey in the search space with two equal probabilities. After detecting prey, the hawk transitions between global and local search based on the decay of prey energy. In the exploitation phase, the energy of the prey and the probability of the prey escaping determine the four different predation methods that the Harris hawk will adopt, namely soft besiege, hard besiege, soft besiege with progressive rapid dives, and soft besiege with progressive rapid dives. When the prey escapes successfully, the Lévy flight is introduced into the algorithm to ensure that the algorithm can jump out of the local optima. In the basic algorithm, the Lévy flight is only used when the position has a better fitness value, which leads to the algorithm being unable to jump out of the local optimum well. The low population diversity and simple search method in the exploration phase weaken the global search capability of HHO. This leads to a longer time to obtain the optimal global solution and increases the possibility of getting stuck in local optima. These can be improved to explore better and exploit the HHO algorithm [50]. For the CBO algorithm, the population is divided into a few randomly chosen Coot leaders and Coot followers. The leaders can make full use of the information in the search space and have a strong tendency to explore, thus effectively leading the followers to the target area. CBO mainly consists of individual random movement, chain movement, and optimal individual-guided movement. The CBO algorithm will choose a random position in the search space as a reference, which is also a an excellent way to help the algorithm jump out of the local optimum. Chain movement allows the algorithm to quickly improve the accuracy of the algorithm when the population is concentrated. In the leader movement of the optimal individual, the position of the follower will change with the position of the corresponding leader. This allows the leader to lead the followers closer and closer to the optimal region. In this paper, the optimal individual movement mechanism of CBO is integrated into the population initialization of HHO to take full advantage of the strengths of both algorithms. In the CBO optimal individual-led movement, B×r8 generates a large amount of randomness. This process can be used to initialize the population to increase the diversity of the position and make up for the lack of information exchange with other individuals, thus laying a good foundation for global search. First, the population initialization formula Eq. (30) of the hybrid algorithm is derived from Eq. (21). After the positions of all search agents are generated by using Eq. (30), the fitness value of each new position is compared with that of the original position. The candicate position with the highest fitness value is selected as the updated position for the hybrid algorithm.

Xnewi(t+1)={B×r8×cos(2Rπ)×(Xprey(t)Xi(t))+Xprey(t),r9<0.5B×r8×cos(2Rπ)×(Xprey(t)Xi(t))Xprey(t),r90.5(30)

where Xnewi(t+1) is the position of the ith search agent in the (t+1)th iteration. Xi(t) denotes the position in the iteration. Xprey(t) is the current optimal solution, and the remaining parameters are defined in Eq. (21).

Secondly, although the basic HHO algorithm introduces the Lévy flight to ensure that it can jump out of the local optimum, it still has limitations that can cause the algorithm to fall into a local optimum in later iterations. The EMS has multiple mutation operators, which can generate three differentcandidate positions according to Eqs. (23)(25). Then, the mutant positions with the lowest fitness is selected for comparison with the fitness of the original positions. The position with the highest fitness is selected as the new position. In this way, the global search ability of the algorithm can be improved. At the end of the algorithm, ROBL is integrated into the HHO. The inverse solution of the current optimum is generated by Eq. (27). This can expand the search range, improve the convergence speed, and reduce the possibility of falling into the local optima. The introduction of EMS and ROBL strategies makes up for the shortcomings of the basic HHO algorithm in the exploitation stage. The flowchart and pseudo-code of the proposed EHHOCBO are shown in Fig. 3 and Algorithm 3, respectively.

images

Figure 3: Flow chart of the proposed EHHOCBO algorithm

images

4  Experimental Results and Discussions

In this section, the effectiveness and feasibility of the proposed EHHOCBO is validated on two sets of optimization problems. Firstly, the performance of the algorithm in solving simple numerical problems is verified by 23 classical benchmark functions [51]. The algorithm’s performance in solving complex numerical optimization problems is evaluated using the IEEE CEC2017 benchmark functions [52]. All experiments were conducted on Windows 10 operating system with a computer hardware configuration of Intel(R) Core(TM) i5-7300HQ CPU@2.50 GHz and 8 GB RAM, and the tool used was MATLAB2020a software.

4.1 Experiment 1: Classical Benchmark Functions

In this section, 23 classical benchmark functions are used to evaluate the performance of EHHOCBO. The functions F1–F7 are unimodal and are used to assess the development capability of the algorithm. F8–F13 are multimodal, which are used to test the exploration capability of the algorithm and its ability to escape from local optimum solutions. F14–F23 are fixed-dimension multimodalities, a combination of unimodal and multimodal, used to evaluate the stability of an algorithm between exploration and exploitation [51]. The population size and maximum iterations for all algorithms were set to 30 and 500, respectively, for a fair comparison. In the proposed EHHOCBO, we set the refraction index z=100 and k=1000. All other parameter settings of the algorithm were kept the same as in the original paper, as shown in Table 1. All experiments should be tested independently 30 to reduce randomness. Detailed descriptions of the 23 functions are listed in Tables 24, and Fig. 4 visualizes the search space for the 23 benchmark functions. The following algorithms are adopted for comparison tests:

•   Harris Hawks Optimization (HHO) [41]

•   Coot Bird Optimization(CBO) [45]

•   Arithmetic Optimization Algorithm (AOA) [21]

•   Whale Optimization Algorithm (WOA) [30]

•   Grey Wolf Optimizer (GWO) [31]

•   Particle Swarm Optimization (PSO) [22]

•   Sooty Tern Optimization Algorithm (STOA) [24]

•   Chimp Optimization Algorithm (ChOA) [25]

images

images

images

images

imagesimages

Figure 4: 3D view of the search space for 23 benchmark functions

In order to evaluate the performance of the algorithm more intuitively, this paper introduces the Mean fitness (Mean) and Standard deviation (Std) evaluation indicators. The difference between the mean fitness and the target value represents the convergence accuracy of the algorithm. The standard deviation evaluates the degree of fluctuation of the experimental results. The size of the Std depends on how stable the algorithm is. The following formula calculates them:

Mean=1timesi=1timesAi(31)

Std=1times1i=1times(AiMean)2(32)

where times is the number of experiments. Ai is the result for each experiment.

4.1.1 Effectiveness of the Introduced Strategy

This paper first integrates the original HHO and CBO algorithms into a new intelligent algorithm named EHHOCBO. The algorithm takes HHO as the core framework, introduces the strong exploratory leadership mechanism of CBO, and then improves the algorithm with EMS and ROBL strategies. To investigate the impact of each improved component on the algorithm, this paper designs a comparison experiment between EHHOCBO and three other derived algorithms (EHHOCBO1, EHHOCBO2, EHHOCBO3). EHHOCBO1 only introduced the leadership mechanism of CBO into the algorithm. EHHOCBO2 introduces the leadership mechanism of the CBO algorithm and the EMS strategy into the algorithm. EHHOCBO3 introduces the leadership mechanism of the CBO algorithm and ROBL strategy. The performance of these four algorithms is tested using 23 standard test functions under the same parameter settings, and the sizes of Mean and Std are listed in Table 5.

images

The experimental results show that in comparing between EHHOCBO and the remaining three derived algorithms, EHHOCBO has better simulation results on the test functions F1–F23. This verifies the effectiveness of the improved method in this paper. The theoretical minimum of the test functions was achieved for the test functions F1–F4, F8, F9, F11, F14, F16–F19, and F23. The remaining three derived algorithms also reached theoretical minima in some test functions, but the overall performance shows a different advantage than the EHHOCBO algorithm. In terms of standard deviation, the simulation results of EHHOCBO also obtained better results. This shows that EHHOCBO has strong stability in solving all test functions. The results show that the improvements in this paper significantly improve the performance of the original HHO algorithm. The improved method in this paper enables the EHHOCBO algorithm to obtain strong exploration and utilization capabilities, laying a foundation for the ability to provide high-quality solutions. After validation, EHHOCBO was selected as the final version for further comparison and discussion.

4.1.2 Comparison between EHHOCBO and Other Optimization Algorithms

This section evaluates the proposed algorithm’s exploration and development capabilities using standard test functions. The above eight well-known algorithms are compared mainly from numerical analysis, box plot, convergence curve, and Wilcoxon test. After 30 independent runs of the experiment. The Mean and Std of the obtained results are recorded in Table 6 below.

images

The proposed EHHOCBO experiment showed better mean fitness and standard deviation. The EHHOCBO has an overwhelming advantage in solving the unimodal test function problem. In the test functions F1–F4, the EHHOCBO can acquire the theoretically global optimal solution compared with the original HHO algorithm and CBO algorithm. In the multi-dimensional test functions F8–F13, the EHHOCBO algorithm achieves better simulation results than the other algorithms. It is worth mentioning that in functions F9 and F11, EHHOCBO obtains the ideal optimal value. The experimental results of the unimodal and multimodal functions fully demonstrate that the EHHOCBO has a better exploitation and exploration capability and a higher possibility of jumping out of the local optimum solution. This shows that the introduced improvement strategy greatly improves the relationship between exploration and exploitation in the HHO algorithm. The EHHOCBO integrates the leadership mechanism of the CBO algorithm into the population initialization of the HHO algorithm. In this way, the stable exploration and development capabilities of the two algorithms are extracted so that the algorithm can easily jump out of the local optimum. For the fixed-dimension benchmark functions F14–F23, the EHHOCBO algorithm obtains better solutions and smaller deviations. Especially in the functions F14, F16–F19, EHHOCBO obtained the theoretically optimal Mean and Std. The smaller Std indicates that the EHHOCBO has higher stability, which can ensure that the algorithm can obtain more accurate results in the application process. The fixed dimension test function results show that the proposed EHHOCBO has good stability between exploration and exploitation due to the introduction of the EMS improvement strategy.

Boxplots are drawn in this paper to visualize the distribution of the data. It can well describe the consistency between data. Fig. 5 depicts the boxplot of EHHOCBO against eight other comparison algorithms over 15 representative benchmark functions. In this figure, serial numbers 1–9 correspond to the algorithms in the table in order, specifically 1-GWO, 2-WOA, 3-PSO, 4-AOA, 5-HHO, 6-ChOA, 7-STOA, 8-CBO, and 9-EHHOCBO. As can be seen from Fig. 5, the EHHOCBO algorithm has better consistency than other original algorithms. No outliers were generated during the iteration. The obtained median, maximum and minimum values are more concentrated than other comparison algorithms and only slightly worse than the HHO algorithm in function F6. The above confirms the high stability of EHHOCBO.

images

Figure 5: Boxplot analysis on 23 benchmark functions

In order to study the convergence behavior of the EHHOCBO algorithm, this paper compares the convergence curve of the EHHOCBO algorithm with the convergence curves of the other eight comparison algorithms on some representative test functions. The results are shown in Fig. 6. According to the convergence curve graph, we can see that the EHHOCBO algorithm has a better convergence performance. In the test functions F1, F3, and F4, the EHHOCBO algorithm can quickly converge to the global optimum, and its convergence curve shows the fastest decay trend. At the same time, other algorithms have obvious lag and slow convergence speed. This is because the algorithm introduces an improved strategy so that the algorithm have better randomness and population diversity. Among the test functions F5 and F6, the convergence of the EHHOCBO algorithm is also the best, and its convergence curve is consistent with that of HHO at the beginning of the iteration. But in the later stage, the convergence trend of HHO could be clearer, and the EHHOCBO algorithm still maintains good convergence until the end. This shows that the introduction of the improved strategy effectively reduces the possibility of the algorithm falling into the local optimum. In the multi-dimensional test functions F8–F12, the EHHOCBO algorithm also maintains good convergence. In test functions F9–F11, EHHOCBO converges to the optimal global solution only after several iterations. In F8, the convergence of EHHOCBO is slightly worse than that of HHO in the early stage. But in the later iterations, the HHO algorithm gradually falls into the local optimum, while the EHHOCBO algorithm still maintains good convergence performance. In F10–F12, although the EHHOCBO algorithm does not reach the theoretical optimum, its convergence performance and final accuracy are the best among all algorithms. The EHHOCBO algorithm also shows strong performance in solving fixed-dimensional testing problems. EHHOCBO converges rapidly in the early stages of testing functions F17, F18, F20, F22, and F23. And make a quick transitions between exploration and exploitation. Finally, the optimal value was determined. The convergence accuracy and operational efficiency of the EHHOCBO algorithm are also improved to some extent compared to HHO and CBO.

images

Figure 6: Convergence curves of different algorithms on fifteen benchmark functions

Table 7 reports the average computation time for each algorithm. The total running time for each algorithm is calculated and ranked as follows: ChOA(51 s) > EHHOCBO(22.8 s) > HHO(9.75 s) > GWO(7.29 s) > CBO(5.98 s) > STOA(5.92 s) > PSO(5.90 s) > AOA(5.43 s) > WOA(4.07 s). Table 7 shows that the EHHOCBO algorithm takes more time than the original HHO algorithm. The main reason is that the improved approach adds steps and extra time to the HHO algorithm. Overall, our proposed algorithm is acceptable due to the performance improvements.

images

To better evaluate the correlation, the Wilcoxon rank sum test was designed, and the significance level was set at 0.05 [53]. The p-values are recorded in Table 8. In the table, the parts where EHHOCBO performs better are highlighted in bold. NaN indicates the same performance as the comparison algorithm. The rest is the part where EHHOCBO has poor performance. As can be seen from Table 8, EHHOCBO performs worse than some algorithms only on individual functions. EHHOCBO performs slightly worse than the PSO algorithm on F18 and slightly worse than the HHO algorithm on F7. In other cases, EHHOCBO achieved the same or better results. Based on statistical theory, EHHOCBO has better performance.

images

In addition, the performance of some optimization algorithms gradually deteriorates as the dimension of the problem expands. In this paper, EHHOCBO and HHO algorithms are simulated in different dimensions to evaluate the impact of scalability on the EHHOCBO algorithm. The experimental tools are F1–F13 of the standard test functions, and the results are recorded in Table 9. The experimental results show that the simulation accuracy of both the original HHO and EHHOCBO decreases with the increase in the number of iterations. However, the simulation results of EHHOCBO are consistently more accurate than HHO. The accuracy of EHHOCBO will remain relatively high and will always be kept in a relatively precise state. It is worth mentioning that EHHOCBO reaches the global optimal solution in F1–F4, F9, and F11. In conclusion, the EHHOCBO algorithm also performs well in solving high-dimensional problems.

images

4.2 Experiment 2: IEEE CEC2017 Test Functions

The benchmark function test proves that EHHOCBO performs well in solving simple problems, but is insufficient to prove its superior performance. To further exploit its performance, this paper applies the EHHOCBO algorithm to solve 29 test functions in IEEE CEC2017 [50,51] to evaluate its performance in solving complex numerical problems. In this section, comparative experiments are conducted between EHHOCBO and six well-known hybrid algorithms based on IEEE CEC2017 test functions. The six algorithms are the Differential Squirrel Search Algorithm (DSSA) [54], Equilibrium Slime Mould Algorithm (ESMA) [55], Grey Wolf Optimizer Based on Aquila Exploration Method (AGWO) [56], Teaching-Learning-Based Optimization Algorithm with Reinforcement Learning Strategy (RLTLBO) [57], Leader Harris Hawks optimization (LHHO) [58] and Hybrid Sine-Cosine Harris Hawks Optimization (SCHHO) [59]. IEEE CEC2017 test functions include 29 functions, which are widely used for performance testing and evaluation of intelligent bee swarm algorithms. Among them, CEC-1 and CEC-3 are single-peaked functions, F4–F10 are simple multi-peaked functions, CEC-11 to CEC-20 are hybrid functions, and CEC-21 to CEC-30 are combined functions. Table 10 lists these functions’ names, dimensions, target values, and search ranges. The proposed EHHOCBO algorithm was run independently of the other comparison algorithms 30 times. The population size and maximum iterations were set to 30 and 500. The mean and standard deviation obtained were recorded in Table 11.

images

images

The results show that the performance of the EHHOCBO algorithm is significantly better than that of DSSA, ESMA, AGWO, LHHO, and SCHHO algorithms and is comparable to the performance of the RLTLBO algorithm. For unimodal functions CEC-1 and CEC-3, EHHOCBO gives better simulation results, but in CEC-1, the standard deviation of the algorithm is slightly worse than RLTLBO. For multimodal functions, EHHOCBO achieves the best standard deviation in the test functions CEC-4, CEC-7, CEC8, and CEC-10, and the simulation results in other multimodal functions are also slightly worse than the best results. The EHHOCBO algorithm also performs well in mixed and composite functions. Among the 20 test functions from CEC-11 to CEC-30, the EHHOCBO algorithm performs better than or equal to other algorithms in 15 of them. In the remaining six functions, the simulation results of EHHOCBO are slightly worse than the optimal results. The above experimental results show that EHHOCBO also has better advantages in solving various complex optimization problems. Fig. 7 is the radar ranking chart of the five algorithms on the CEC2017 test function. The size of the range enclosed by each curve in the figure represents algorithm’s performance. The smaller the range, the better the performance. The figure shows that the EHHOCBO has better performance.

images

Figure 7: Ranking on IEEE CEC2017 test function

5  EHHOCBO for Addressing Engineering Problems

This section uses EHHOCBO to solve four common engineering problems in the structure field. Engineering problem testing validates the applicability and black-box nature of meta-heuristic algorithms in real-world constrained optimization. The four problems used are the cantilever beam design problem, speed reducer design problem, welded beam design problem, and rolling element bearing design problem. This paper introduces the death penalty function [60] to deal with those non-feasible candidate solutions under equality and inequality constraints. The maximum number of iterations and population size are still set to 500 and 30, respectively. In this experiment, each algorithm runs independently 30 times. The detailed experimental results and discussions are as follows.

5.1 Cantilever Beam Design Problem

The cantilever beam design problem is a structural engineering design problem that is related to the weight optimization of a cantilever beam with square cross-section. One end of the cantilever beam is rigidly supported, and the vertical force acts on the free node of the cantilever. The beam comprises five hollow blocks with constant thickness, and its height is a decision variable. The structure of the cantilever beam is shown in Fig. 8, and the mathematical formula of the problem is as follows:

images

Figure 8: Schematic diagram of cantilever beam design problem

Consider: x=[x1,x2,x3,x4,x5]

Minimize: f(x)=0.0624(x1+x2+x3+x4+x5)

Subject to: g(x)=61x13+37x23+19x33+7x43+1x5310

Variable range: 0.01xi100,i=1,2,,5

The optimization results of EHHOCBO and other algorithms for the cantilever beam design problem are presented in Table 12. The results show that EHHOCBO proposed in this paper achieves the best design. Its performance is significantly improved compared with the basic CBO and HHO algorithms. Therefore, we believe that the EHHOCBO algorithm has good potential for solving such problems.

images

5.2 Speed Reducer Design Problem

The objective of this optimization problem is to minimize the weight of the reduction gear subject to 11 constraints. The decision variables in this problem are the width of the tooth face x1, the modulus of the gear x2, the number of teeth x3 in the pinion, the length of the first shaft x4, and the diameter x5, the length of the second shaft x6, and the diameter x7. Fig. 9 is the schematic diagram of the reducer, and the mathematical formula is shown as follows:

images

Figure 9: Schematic diagram of speed reducer design problem

Consider: x=[x1,x2,x3,x4,x5,x6,x7]

Minimize: f(x)=0.7854x1x22(3.3333x32+14.9334x343.0934)1.508x1(x62+x72)+7.4777(x63+x73)

Subject to: g1(x)=27x1x22x310,g2(x)=397.5x1x22x3210g3(x)=1.93x43x2x3x6410,g4(x)=1.93x53x2x3x7410g5(x)=(745x4x2x3)2+16.9×106110.0x6310g6(x)=(745x4x2x3)2+157.5×10685.0x6310g7(x)=x2x34010,g8(x)=5x2x110

g9(x)=x112x210,g10(x)=1.5x6+1.9x410g11(x)=1.1x7+1.9x510

Variable range: 2.6x13.6,0.7x20.8,17x328,7.3x48.3,7.8x58.3,2.9x63.9,5.0x75.5

The results of this experiment are recorded in Table 13. The results show that the minimum weight of EHHOCBO is slightly worse than that of RLTLBO and ESMA algorithms. Although EHHOCBO could have better design results, it achieves relatively good performance. Therefore, it is reasonable to believe that the proposed hybrid technique is suitable for solving the speed reducer design problem.

images

5.3 Welded Beam Design Problem

The purpose of the design problem of the welded beam is to reduce the manufacturing cost of the design as much as possible under the constraint of meeting the beam’s shear stress, bending stress, bending load, end deviation, and boundary conditions. The problem can be described as an optimization design problem with four decision variables. The four decision variables are length l, height t, thickness b, and weld thickness h of the beam bar. Fig. 10 is the schematic diagram of the welded beam. The mathematical representation of the problem is as follows:

images

Figure 10: Schematic diagram of the welded beam design problem

Consider: x=[x1,x2,x3,x4]=[h,l,t,b]

Minimize: f(x)=1.10471x12x2+0.04811x3x4(14+x2)

Subject to: g1(x)=τ(x)τmax0g2(x)=σσmax0g3(x)=δδmax0g4(x)=x1x40g5(x)=PPC(x)0g6(x)=0.125x10g7(x)=1.10471x12+0.04811x3x4(14+x2)50

Variable range: 0.1x1,x42,0.1x2,x310

where f(x) represents the cost of manufacturing welded beams, and other variables are defined as follows:

τ(x)=(τ)2+2ττx22R+(τ)2,τ=P2x1x2,τ=MRJ,M=P(L+x22)R=x224+(x1+x32)2,J=2{2x1x2[x224+(x1+x32)2]},σ(x)=6PLEx32x4

δ(x)=6PL3Ex32x4,PC(x)=4.013Ex32x4636L2(1x32LE4G)

P=6000lb,L=14in,E=30×106psi,G=12×106psi,δmax=0.25in,τmax=13600psi,σmax=30000psi.

The results of the experiment are shown in Table 14. EHHOCBO obtaines the lowest manufacturing cost of 1.7256 under the corresponding parameters h=0.2057, l=3.4698, t=9.0438, and b=0.2057. Since EHHOCBO has achieved the most satisfactory results in all competitive algorithms, it has a good application prospect in designing welded beams.

images

5.4 Rolling Element Bearing Design Problem

The objective of the rolling element bearing design problem is to find the maximum dynamic load capacity of a bearing subject to more than ten decision parameters. The decision parameters are pitch diameter Dm, ball diameter Db, number of balls Z, inner and outer raceway radius of curvature coefficients fi and fo, Kdmin, Kdmax, δ, e, and ζ. Fig. 11 shows schematic diagram of the rolling element bearing. The mathematical expression of the problem is as follows:

images

Figure 11: Schematic diagram of rolling element bearing design problem

Maximize: Cd={fcZ2/3Db1.8,if Db25.4mm3.647fcZ2/3Db1.4,else

Subject to:

g1(x)=ϕ02sin1(Db/Dm)Z+10g2(x)=2DbKdmin(Dd)>0g3(x)=Kdmax(Dd)2Db0g4(x)=ζBwDb0g5(x)=Dm0.5(D+d)0g6(x)=(0.5+e)(D+d)Dm0g7(x)=0.5(DDmDb)δDb0g8(x)=fi0.515g9(x)=fo0.515

where

fc=37.91[1+{1.04(1γ1+γ)1.72(fi(2fo1)fo(2fi1))0.41}10/3]0.3×[γ0.3(1γ)1.39(1+γ)1/3][2fi2fi1]0.41x=[{(Dd)/23(T/4)}2+{D/2T/4Db}2{d/2+T/4}2]y=2{(Dd)/23(T/4)}{D/2T/4Db}ϕ0=2cos1(xy),γ=DbDm,fi=riDb,fo=roDb,T=Dd2DbD=160,d=90,Bw=30,ri=ro=11.033,0.5(D+d)Dm0.6(D+d)0.15(Dd)Db0.45(Dd),4Z50,0.515fi and fo0.60.4Kdmin0.5,0.6Kdmax0.7,0.3δ0.4,0.02e0.1,0.6ζ0.85.

The results of this experiment are reported in Table 15. It can be seen from this table that compared with other classical algorithms, the proposed EHHOCBO provides the best fitness value of 85549.

images

The above experiments show that EHHOCBO has advantages in solving practical engineering problems. This is attributed to the combination of HHO and CBO, as well as the introduction of the ensemble mutation strategy and refracted opposition-based learning, which improves the proposed method’s searchability significantly.

6  Conclusion and Future Directions

Considering the respective characteristics of the HHO algorithm and CBO algorithm, a new improved hybrid algorithm named EHHOCBO is proposed in this paper. First, the leader mechanism of CBO is introduced in the initialization phase of HHO to provide a good basis for global search and enhance the exploration ability of the algorithm. Then, an ensemble mutation strategy is employed to increase the population diversity and further boost the exploration trend. Finally, refracted opposition-based learning is used to update the optimal solution to expand the search range and avoid the algorithm falling into the local optima. In order to fully evaluate the performance of the proposed algorithm, EHHOCBO was compared with the basic HHO, CBO algorithm, and other meta-heuristic algorithms based on classical benchmark functions and the IEEE CEC2017 test suite. Wilcoxon’s rank sum test verified the significance of the experimental results. Through a series of numerical statistics, it is verified that the EHHOCBO algorithm has significantly improved in terms of accuracy, convergence speed, stability, and avoidance of falling into local optimum solutions. In addition, four engineering problems were used to verify the applicability of the EHHOCBO algorithm. Experiments show that the algorithm can effectively provide competitive solutions for these real-life engineering problems.

Although the EHHOCBO algorithm proposed in this paper has been significantly improved over the basic HHO and CBO algorithms, there is still room for further improvement in its performance on IEEE CEC2017 test functions. The EHHOCBO algorithm still suffers from the major limitation of excessive computation time and needs to be improved. It is believed that this situation can be alleviated by introducing several parallel mechanisms, such as master-slave models, cellular models and coordination strategies. In future work, we aim to further improve the performance of the EHHOCBO algorithm by developing new improvement strategies and parallelism mechanisms while reducing the total time consumption. And we hope to make the parameter settings more reasonable by incorporating parameter adaptive control. In addition, we believe it would also be interesting to implement EHHOCBO to solve more practical optimization problems such as feature selection, path planning, predictive modeling, image segmentation, etc. [61]. We will enhance our work in this area.

Acknowledgement: The authors are grateful to the editor and reviewers for their constructive comments and suggestions, which have improved the presentation.

Funding Statement: This work is financially supported by the National Natural Science Foundation of China under Grant 52075090, Key Research and Development Program Projects of Heilongjiang Province under Grant GA21A403, the Fundamental Research Funds for the Central Universities under Grant 2572021BF01, and Natural Science Foundation of Heilongjiang Province under Grant YQ2021E002.

Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.

References

1. Xiao, Y., Sun, X., Guo, Y., Li, S., Zhang, Y. et al. (2022). An improved gorilla troops optimizer based on lens opposition-based learning and adaptive beta-Hill climbing for global optimization. Computer Modeling in Engineering & Sciences, 131(2), 815–850. https://doi.org/10.32604/cmes.2022.019198 [Google Scholar] [CrossRef]

2. Xiao, Y., Sun, X., Guo, Y., Cui, H., Wang, Y. et al. (2022). An enhanced honey badger algorithm based on Levy flight and refraction opposition-based learning for engineering design problems. Journal of Intelligent & Fuzzy Systems, 43(4), 4517–4540. https://doi.org/10.3233/JIFS-213206 [Google Scholar] [CrossRef]

3. Zhang, X. C., Zhao, K., Niu, Y. (2020). Improved Harris hawks optimization based on adaptive cooperative foraging and dispersed foraging strategies. IEEE Access, 8, 160297–160314. https://doi.org/10.1109/ACCESS.2020.3013332 [Google Scholar] [CrossRef]

4. Boussaid, I., Lepagnot, J., Siarry, P. (2013). A survey on optimization metaheuristics. Information Sciences, 237, 82–117. https://doi.org/10.1016/j.ins.2013.02.041 [Google Scholar] [CrossRef]

5. Jia, H. M., Li, Y., Sun, K. J., Cao, N., Zhou, H. M. (2021). Hybrid sooty tern optimization and differential evolution for feature selection. Computer Systems Science and Engineering, 39(3), 321–335. https://doi.org/10.32604/csse.2021.017536 [Google Scholar] [CrossRef]

6. Fan, Q. S., Huang, H. S., Yang, K., Zhang, S. S., Yao, L. G. et al. (2021). A modified equilibrium optimizer using opposition-based learning and novel update rules. Expert Systems with Applications, 170(9), 114575. https://doi.org/10.1016/j.eswa.2021.114575 [Google Scholar] [CrossRef]

7. Hussain, K., Mohd Salleh, M. N., Cheng, S., Shi, Y. (2019). Metaheuristic research: A comprehensive survey. Artificial Intelligence Review, 52(4), 2191–2233. https://doi.org/10.1007/s10462-017-9605-z [Google Scholar] [CrossRef]

8. Zheng, R., Jia, H. M., Wang, S., Liu, Q. X. (2022). Enhanced slime mould algorithm with multiple mutation strategy and restart mechanism for global optimization. Journal of Intelligent & Fuzzy Systems, 42(6), 5069–5083. https://doi.org/10.3233/JIFS-211408 [Google Scholar] [CrossRef]

9. Xiao, Y., Guo, Y., Cui, H., Wang, Y., Li, J. et al. (2022). IHAOAVOA: An improved hybrid aquila optimizer and African vultures optimization algorithm for global optimization problems. Mathematical Biosciences and Engineering, 19(11), 10963–11017. https://doi.org/10.3934/mbe.2022512 [Google Scholar] [PubMed] [CrossRef]

10. Zhong, K. Y., Zhou, G., Deng, W., Zhou, Y. Q., Luo, Q. F. (2021). MOMPA: Multi-objective marine predator algorithm. Computer Methods in Applied Mechanics and Engineering, 385(1), 114029. https://doi.org/10.1016/j.cma.2021.114029 [Google Scholar] [CrossRef]

11. Yu, G., Wang, H., Zhou, H. Z., Zhao, S. S., Wang, Y. (2021). An efficient firefly algorithm based on modified search strategy and neighborhood attraction. International Journal of Intelligent Systems, 36(8), 4346–4363. https://doi.org/10.1002/int.22462 [Google Scholar] [CrossRef]

12. Xiao, Y., Sun, X., Zhang, Y., Guo, Y., Wang, Y. et al. (2021). An improved slime mould algorithm based on tent chaotic mapping and nonlinear inertia weight. International Journal of Innovative Computing, Information and Control, 17(6), 2151–2176. https://doi.org/10.24507/ijicic.17.06.2151 [Google Scholar] [CrossRef]

13. Nguyen, T. T., Wang, H. J., Dao, T. K., Pan, J. S., Liu, J. H. et al. (2020). An improved slime mold algorithm and its application for optimal operation of cascade hydropower stations. IEEE Access, 8, 226754–226772. https://doi.org/10.1109/ACCESS.2020.3045975 [Google Scholar] [CrossRef]

14. Dehghani, M., Montazeri, Z., Givi, H., Guerrero, J., Dhiman, G. (2020). Darts game optimizer: A new optimization technique based on darts game. International Journal of Intelligent Engineering and Systems, 13(5), 286–294. https://doi.org/10.22266/ijies2020.1031.26 [Google Scholar] [CrossRef]

15. Hamed, A. Y., Alkinani, M. H., Hassan, M. R. (2020). A genetic algorithm optimization for multi-objective multicast routing. Intelligent Automation and Soft Computing, 26(6), 1201–1216. https://doi.org/10.32604/iasc.2020.012663 [Google Scholar] [CrossRef]

16. Storn., R., Price., K. (1997). Differential evolution–A simple and efficient heuristic for global optimization over continuous spaces. Journal of Global Optimization, 11(4), 341–359. https://doi.org/10.1023/A:1008202821328 [Google Scholar] [CrossRef]

17. Beyer, H. G., Schwefel, H. P. (2002). Evolution strategies. A comprehensive introduction. Natural Computing, 1(1), 3–52. https://doi.org/10.1023/A:1015059928466 [Google Scholar] [CrossRef]

18. Kirkpatrick, S., Gelatt, C. D., Vecchi, M. P. (1983). Optimization by simulated annealing. Science, 220(4598), 671–680. https://doi.org/10.1126/science.220.4598.671 [Google Scholar] [PubMed] [CrossRef]

19. Mirjalili, S., Mirjalili, S. M., Hatamlou, A. (2016). Multi-verse optimizer: A nature-inspired algorithm for global optimization. Neural Computing & Applications, 27(2), 495–513. https://doi.org/10.1007/s00521-015-1870-7 [Google Scholar] [CrossRef]

20. Zhao, W. G., Wang, L. Y., Zhang, Z. X. (2019). Atom search optimization and its application to solve a hydrogeologic parameter estimation problem. Knowledge-Based Systems, 163(4598), 283–304. https://doi.org/10.1016/j.knosys.2018.08.030 [Google Scholar] [CrossRef]

21. Abualigah, L., Diabat, A., Mirjalili, S., Abd Elaziz, M., Gandomi, A. H. (2021). The arithmetic optimization algorithm. Computer Methods in Applied Mechanics and Engineering, 376(2), 113609. https://doi.org/10.1016/j.cma.2020.113609 [Google Scholar] [CrossRef]

22. Kennedy, J., Eberhart, R. (1995). Particle swarm optimization. Proceedings of ICNN’95—International Conference on Neural Networks, pp. 1942–1948. Perth, WA, Australia. https://doi.org/10.1109/ICNN.1995.488968 [Google Scholar] [CrossRef]

23. Mirjalili, S. (2016). SCA: A sine cosine algorithm for solving optimization problems. Knowledge-Based Systems, 96(63), 120–133. https://doi.org/10.1016/j.knosys.2015.12.022 [Google Scholar] [CrossRef]

24. Dhiman, G., Kaur, A. (2019). STOA: A bio-inspired based optimization algorithm for industrial engineering problems. Engineering Applications of Artificial Intelligence, 82(2), 148–174. https://doi.org/10.1016/j.engappai.2019.03.021 [Google Scholar] [CrossRef]

25. Khishe, M., Mosavi, M. R. (2020). Chimp optimization algorithm. Expert Systems with Applications, 149(1), 113338. https://doi.org/10.1016/j.eswa.2020.113338 [Google Scholar] [CrossRef]

26. Abualigah, L., Yousri, D., Abd Elaziz, M., Ewees, A. A., Al-qaness, M. A. A. et al. (2021). Aquila optimizer: A novel meta-heuristic optimization algorithm. Computers & Industrial Engineering, 157(11), 107250. https://doi.org/10.1016/j.cie.2021.107250 [Google Scholar] [CrossRef]

27. Mirjalili, S. (2016). Dragonfly algorithm: A new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems. Neural Computing & Applications, 27(4), 1053–1073. https://doi.org/10.1007/s00521-015-1920-1 [Google Scholar] [CrossRef]

28. Arora, S., Singh, S. (2019). Butterfly optimization algorithm: A novel approach for global optimization. Soft Computing, 23(3), 715–734. https://doi.org/10.1007/s00500-018-3102-4 [Google Scholar] [CrossRef]

29. Li, S. M., Chen, H. L., Wang, M. J., Heidari, A. A., Mirjalili, S. (2020). Slime mould algorithm: A new method for stochastic optimization. Future Generation Computer Systems, 111(Supplement C), 300–323. https://doi.org/10.1016/j.future.2020.03.055 [Google Scholar] [CrossRef]

30. Mirjalili, S., Lewis, A. (2016). The whale optimization algorithm. Advances in Engineering Software, 95(12), 51–67. https://doi.org/10.1016/j.advengsoft.2016.01.008 [Google Scholar] [CrossRef]

31. Mirjalili, S., Mirjalili, S. M., Lewis, A. (2014). Grey wolf optimizer. Advances in Engineering Software, 69, 46–61. https://doi.org/10.1016/j.advengsoft.2013.12.007 [Google Scholar] [CrossRef]

32. Dhiman, G., Kumar, V. (2019). Seagull optimization algorithm: Theory and its applications for large-scale industrial engineering problems. Knowledge-Based Systems, 165(25), 169–196. https://doi.org/10.1016/j.knosys.2018.11.024 [Google Scholar] [CrossRef]

33. Abdollahzadeh, B., Gharehchopogh, F. S., Mirjalili, S. (2021). Artificial gorilla troops optimizer: A new nature-inspired metaheuristic algorithm for global optimization problems. International Journal of Intelligent Systems, 36(10), 5887–5958. https://doi.org/10.1002/int.22535 [Google Scholar] [CrossRef]

34. Goncalves, M. S., Lopez, R. H., Miguel, L. F. F. (2015). Search group algorithm: A new metaheuristic method for the optimization of truss structures. Computers & Structures, 153(12), 165–184. https://doi.org/10.1016/j.compstruc.2015.03.003 [Google Scholar] [CrossRef]

35. Wolpert, D. H., Macready, W. G. (1997). No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation, 1(1), 67–82. https://doi.org/10.1109/4235.585893 [Google Scholar] [CrossRef]

36. Jia, H. M., Sun, K. J., Zhang, W. Y., Leng, X. (2022). An enhanced chimp optimization algorithm for continuous optimization domains. Complex & Intelligent Systems, 8(1), 65–82. https://doi.org/10.1007/s40747-021-00346-5 [Google Scholar] [CrossRef]

37. Debnath, S., Baishya, S., Sen, D., Arif, W. (2021). A hybrid memory-based dragonfly algorithm with differential evolution for engineering application. Engineering with Computers, 37(4), 2775–2802. https://doi.org/10.1007/s00366-020-00958-4 [Google Scholar] [CrossRef]

38. Ziyu, T., Dingxue, Z. (2009). A modified particle swarm optimization with an adaptive acceleration coefficients. 2009 Asia-Pacific Conference on Information Processing, pp. 330–332. Shenzhen, China. https://doi.org/10.1109/APCIP.2009.217 [Google Scholar] [CrossRef]

39. Li, S. L., Li, X. B., Chen, H., Zhao, Y. X., Dong, J. W. (2021). A novel hybrid hunger games search algorithm with differential evolution for improving the behaviors of non-cooperative animals. IEEE Access, 9, 164188–164205. https://doi.org/10.1109/ACCESS.2021.3132617 [Google Scholar] [CrossRef]

40. Jia, H., Jiang, Z., Li, Y. (2022). Simultaneous feature selection optimization based on improved bald eagle search algorithm. Control and Decision, 37(2), 445–454. https://doi.org/10.13195/j.kzyjc.2020.1025 [Google Scholar] [CrossRef]

41. Heidari, A. A., Mirjalili, S., Faris, H., Aljarah, I., Mafarja, M. et al. (2019). Harris hawks optimization: Algorithm and applications. Future Generation Computer Systems, 97, 849–872. https://doi.org/10.1016/j.future.2019.02.028 [Google Scholar] [CrossRef]

42. Fan, Q., Chen, Z. J., Xia, Z. H. (2020). A novel quasi-reflected Harris hawks optimization algorithm for global optimization problems. Soft Computing, 24(19), 14825–14843. https://doi.org/10.1007/s00500-020-04834-7 [Google Scholar] [CrossRef]

43. Dokeroglu, T., Sevinc, E. (2022). An island parallel Harris hawks optimization algorithm. Neural Computing and Applications, 34(21), 18341–18368. https://doi.org/10.1007/s00521-022-07367-2 [Google Scholar] [CrossRef]

44. Wang, S. A., Jia, H. M., Abualigah, L., Liu, Q. X., Zheng, R. (2021). An improved hybrid aquila optimizer and Harris hawks algorithm for solving industrial engineering optimization problems. Processes, 9(9), 1551. https://doi.org/10.3390/pr9091551 [Google Scholar] [CrossRef]

45. Naruei, I., Keynia, F. (2021). A new optimization method based on COOT bird natural life model. Expert Systems with Applications, 183(2), 115352. https://doi.org/10.1016/j.eswa.2021.115352 [Google Scholar] [CrossRef]

46. Huang, Y. H., Zhang, J., Wei, W., Qin, T., Fan, Y. C. et al. (2022). Research on coverage optimization in a WSN based on an improved COOT bird algorithm. Sensors, 22(9), 3383. https://doi.org/10.3390/s22093383 [Google Scholar] [PubMed] [CrossRef]

47. Wang, Y., Cai, Z. X., Zhang, Q. F. (2011). Differential evolution with composite trial vector generation strategies and control parameters. IEEE Transactions on Evolutionary Computation, 15(1), 55–66. https://doi.org/10.1109/TEVC.2010.2087271 [Google Scholar] [CrossRef]

48. Long, W., Wu, T. B., Jiao, J. J., Tang, M. Z., Xu, M. (2020). Refraction-learning-based whale optimization algorithm for high-dimensional problems and parameter estimation of PV model. Engineering Applications of Artificial Intelligence, 89(1), 103457. https://doi.org/10.1016/j.engappai.2019.103457 [Google Scholar] [CrossRef]

49. Tizhoosh, H. R. (2005). Opposition-based learning: A new scheme for machine intelligence. International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce, pp. 695–701. Vienna, Austria. https://doi.org/10.1109/CIMCA.2005.1631345 [Google Scholar] [CrossRef]

50. Sihwail, R., Omar, K., Ariffin, K. A. Z., Tubishat, M. (2020). Improved harris hawks optimization using elite opposition-based learning and novel search mechanism for feature selection. IEEE Access, 8, 121127–121145. https://doi.org/10.1109/ACCESS.2020.3006473 [Google Scholar] [CrossRef]

51. Digalakis, J. G., Margaritis, K. G. (2001). On benchmarking functions for genetic algorithms. International Journal of Computer Mathematics, 77(4), 481–506. https://doi.org/10.1080/00207160108805080 [Google Scholar] [CrossRef]

52. Cheng, R., Li, M., Tian, Y., Zhang, X., Jin, Y. et al. (2017). Benchmark functions for the CEC’2017 competition on evolutionary many-objective optimization. http://hdl.handle.net/2086/13857 [Google Scholar]

53. García, S., Fernández, A., Luengo, J., Herrera, F. (2010). Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. Information Sciences, 180(10), 2044–2064. https://doi.org/10.1016/j.ins.2009.12.010 [Google Scholar] [CrossRef]

54. Jena, B., Naik, M. K., Wunnava, A., Panda, R. (2021). A differential squirrel search algorithm. In: Das, S., Mohanty, M. N. (Eds.Advances in intelligent computing and communication, pp. 143–152. Singapore: Springer. [Google Scholar]

55. Naik, M. K., Panda, R., Abraham, A. (2021). An entropy minimization based multilevel colour thresholding technique for analysis of breast thermograms using equilibrium slime mould algorithm. Applied Soft Computing, 113, 107955. https://doi.org/10.1016/j.asoc.2021.107955 [Google Scholar] [CrossRef]

56. Ma, C., Huang, H., Fan, Q., Wei, J., Du, Y. et al. (2022). Grey wolf optimizer based on aquila exploration method. Expert Systems with Applications, 205, 117629. https://doi.org/10.1016/j.eswa.2022.117629 [Google Scholar] [CrossRef]

57. Wu, D., Wang, S., Liu, Q., Abualigah, L., Jia, H. (2022). An improved teaching-learning-based optimization algorithm with reinforcement learning strategy for solving optimization problems. Computational Intelligence and Neuroscience, 2022, 1535957. https://doi.org/10.1155/2022/1535957 [Google Scholar] [PubMed] [CrossRef]

58. Naik, M. K., Panda, R., Wunnava, A., Jena, B., Abraham, A. (2021). A leader Harris hawks optimization for 2-D Masi entropy-based multilevel image thresholding. Multimedia Tools and Applications, 80(28), 35543–35583. https://doi.org/10.1007/s11042-020-10467-7 [Google Scholar] [CrossRef]

59. Hussain, K., Neggaz, N., Zhu, W., Houssein, E. H. (2021). An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection. Expert Systems with Applications, 176, 114778. https://doi.org/10.1016/j.eswa.2021.114778 [Google Scholar] [CrossRef]

60. Coello Coello, C. A. (2002). Theoretical and numerical constraint-handling techniques used with evolutionary algorithms: A survey of the state of the art. Computer Methods in Applied Mechanics and Engineering, 191(11), 1245–1287. https://doi.org/10.1016/S0045-7825(01)00323-1 [Google Scholar] [CrossRef]

61. Shehab, M., Mashal, I., Momani, Z., Shambour, M. K. Y., Al-Badareen, A. et al. (2022). Harris hawks optimization algorithm: Variants and applications. Archives of Computational Methods in Engineering, 29(7), 5579–5603. https://doi.org/10.1007/s11831-022-09780-1 [Google Scholar] [CrossRef]


Cite This Article

APA Style
Cui, H., Guo, Y., Xiao, Y., Wang, Y., Li, J. et al. (2023). Enhanced harris hawks optimization integrated with coot bird optimization for solving continuous numerical optimization problems. Computer Modeling in Engineering & Sciences, 137(2), 1635-1675. https://doi.org/10.32604/cmes.2023.026019
Vancouver Style
Cui H, Guo Y, Xiao Y, Wang Y, Li J, Zhang Y, et al. Enhanced harris hawks optimization integrated with coot bird optimization for solving continuous numerical optimization problems. Comput Model Eng Sci. 2023;137(2):1635-1675 https://doi.org/10.32604/cmes.2023.026019
IEEE Style
H. Cui et al., “Enhanced Harris Hawks Optimization Integrated with Coot Bird Optimization for Solving Continuous Numerical Optimization Problems,” Comput. Model. Eng. Sci., vol. 137, no. 2, pp. 1635-1675, 2023. https://doi.org/10.32604/cmes.2023.026019


cc Copyright © 2023 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 1859

    View

  • 748

    Download

  • 1

    Like

Share Link