Open Access
ARTICLE
An Improved Honey Badger Algorithm through Fusing Multi-Strategies
1 School of Computer Science, Hubei University of Technology, Wuhan, 430068, China
2 Xining Big Data Service Administration, Xining, 810000, China
* Corresponding Author: Chun Liu. Email:
(This article belongs to the Special Issue: Optimization Algorithm in Real-World Applications)
Computers, Materials & Continua 2023, 76(2), 1479-1495. https://doi.org/10.32604/cmc.2023.038787
Received 29 December 2022; Accepted 16 March 2023; Issue published 30 August 2023
Abstract
The Honey Badger Algorithm (HBA) is a novel meta-heuristic algorithm proposed recently inspired by the foraging behavior of honey badgers. The dynamic search behavior of honey badgers with sniffing and wandering is divided into exploration and exploitation in HBA, which has been applied in photovoltaic systems and optimization problems effectively. However, HBA tends to suffer from the local optimum and low convergence. To alleviate these challenges, an improved HBA (IHBA) through fusing multi-strategies is presented in the paper. It introduces Tent chaotic mapping and composite mutation factors to HBA, meanwhile, the random control parameter is improved, moreover, a diversified updating strategy of position is put forward to enhance the advantage between exploration and exploitation. IHBA is compared with 7 meta-heuristic algorithms in 10 benchmark functions and 5 engineering problems. The Wilcoxon Rank-sum Test, Friedman Test and Mann-Whitney U Test are conducted after emulation. The results indicate the competitiveness and merits of the IHBA, which has better solution quality and convergence traits. The source code is currently available from: .Keywords
Meta-heuristic algorithms (MAs) gain wide attention in different science branches because of their higher efficiency and performance in recent years [1–3]. Little core knowledge is demanded to reach encouraging results, and it is effortless to transplant between different disciplines. Also, stochastic optimization techniques of MAs can effectively become retarded from getting stuck to the local optimum [4]. The uncertainty and randomness of MAs bring the possibility of finding better solutions while the algorithm searches in the global search space. Unfortunately, although most algorithms have been verified to be effective, the No-Free-Lunch theory (NFL) states that no one can deal with all real-world optimization problems [5]. This theory stimulates researchers to design a more versatile and efficient algorithms.
Generally, MAs consist of three main branches: evolutionary techniques, physics-based, and swarm-based methods [6]. The evolutionary algorithm (EA) makes use of the strategies of crossing, mutating, recombining, and selecting to obtain the superior offspring [7], such as Genetic Algorithm (GA) [8], Differential Evolution (DE) [9], which selecting high-quality individuals through population evolution and eliminating inferior individuals by the concept of survival of the fittest. The physics-based algorithm is based on natural phenomena and transforms the optimization process into a physical phenomenon. The main examples of this class are simulated annealing (SA) [10] and Equilibrium Optimizer (EO) [11]. The swarm-based algorithm imitates the behaviors of animals preying or searching for foods to update their positions in the population [6], as Particle Swarm Optimization (PSO) [12] and Grey Wolf Optimizer (GWO) [13]. The swarm-based algorithm simulates the mechanism of division of labor and collaboration among animal populations in nature and has good application prospects for realistic and complex problems [14,15]. For instance, Zhao et al. proposed the Manta Ray Foraging Optimization algorithm (MRFO) [16], which was inspired by the behaviors of the manta ray. There were three behaviors in MRFO, including chain foraging, cyclone foraging, and somersault foraging, which updated individual positions according to the global optimal individual position in a different state of behavior to solve optimization problems. The inspiration for the Reptile Search Algorithm (RSA) was based on the two steps of crocodile behavior, such as encircling which was performed by highly walking, and hunting which was performed by aggregating through cooperation [17]. Despite the basic version of the meta-heuristic algorithm being exciting, their variants are still instructive, which are broadly listed into three categories: optimizing the initial position of population or algorithm parameters [18], adding mutation factors [19], and blending other algorithms [20]. Although different MAs have distinguishing strategies for solving problems, they have two equivalent stages in the process of optimization: exploration and exploitation [6]. When the exploration is dominant, the algorithm can search in the solution space more widely, generating a more differentiated solution. When the exploitative is dominant, the algorithm can search more intensively. However, if the ability to explore is improved, it will reduce the ability to exploit, and vice versa [21]. Therefore, it is challenging to attain an acceptable balance between the two main stages.
Honey Badger Algorithm (HBA) is a novel swarm-based algorithm proposed recently, which is characterized by fewer adjustment parameters and better robustness performance [22]. Nassef et al. injected an efficient local search method which was called Dimensional Learning Hunting (DLH) into HBA and operated the photovoltaic systems cells as close as possible to the global value in maximum power point tracking problems [23]. Düzenli̇ et al. added Gauss mapping and oppositional-based Learning (OBL) strategy to control critical value in the exploration and exploitation stages [24]. However, it was still unevenly explored and exploited, which was also common in MAs [25].
With the aspiration of a more versatile and efficient algorithm, this paper proposes an improved HBA with fusing multi-strategies (IHBA) to balance the exploration and exploitation of HBA. Firstly, the Tent chaotic mapping is used to initialize individuals [26]. Secondly, the process of the parameters tuning of HBA is adjusted, which adds several random values to nonlinear parameters. Thirdly, three compound mutation factors are introduced to strengthen the ability of individuals to wander randomly. Finally, a modified position updating rules for individuals is introduced to improve the global optimization ability and avoid algorithm getting stuck in the local optimum. Simulation experiments are conducted on 10 benchmark functions and 5 engineering problems. The results show that the global search capability and convergence accuracy of the IHBA is improved effectively. The major contributions of this paper are given as follows:
• Tent chaotic mapping, control parameters, and mutation factors are introduced to enhance the adaptability of the algorithm to different tasks.
• An improved updating strategy of position to control the population search range has been proposed.
• Compared with EO, MRFO, RSA, HBA, Nonlinear-based chaotic HHO (NCHHO) [26], improved GWO (IGWO) [27], and Differential Squirrel Search Algorithm (DSSA) [28], the Wilcoxon Rank-sum Test, Friedman Test, and Mann-Whitney U Test are performed on the benchmark functions and engineering problems and generate a series of comparative analyses.
The rest of this paper is organized as follows: The original HBA is described in Section 2. Section 3 outlines IHBA using multi-strategies fusion. Section 4 and Section 5 discusses the performance differences on benchmark functions and engineering problems. Finally, Section 6 concludes the work and suggests several future research directions.
2 The Basic Honey Badger Algorithm
Honey Badger Algorithm (HBA) is inspired by the foraging behavior of honey badgers. There are two phases of HBA: the digging phase and the honey phase. In the digging phase, the population of honey badgers is guided by the smell of bees or wanders around the bees. In the honey phase, the population is directly attracted to the bees until they locate the honeycomb [22]. In HBA, the fragrance is distributed by the bees (food), inducing honey badgers to congregate towards them. After sensing the smell, the honey badger moves toward the prey. HBA uses Eq. (1) to imitate the odor intensity (
where
In the digging phase, the honey badger population needs to perceive the location of food in a global environment. The honey badger population is guided by the optimal individual to perform the search task. The digging mode is updated according to Eq. (2).
where
Different from the digging phase, the honey badgers are guided by the honeyguide bird to approach the prey during the honey phase, where the individuals hover around prey until they catch it. It can be simulated as Eq. (3), where
In the basic HBA, the initial position of the individual is generated randomly, which provides an opportunity that the population to search in a wide area. The digging mode and honey mode provide detailed foraging behavior for the population. The honey badger, based on the odor intensity of the prey remaining in the air, identifies the approximate position of the prey and then approaches the prey until it locates. The Honey Badger Algorithm uses global sight to tackle the optimization and to achieve optimal solutions. Chen et al. applied HBA based energy management strategy to solve the problem of performance enhancement in the pressure retarded osmosis and photovoltaic/thermal system by maximizing the multi-variable functions [29]. Ashraf et al. applied HBA to optimally identify the seven ungiven parameters of the proton exchange membrane fuel cells steady-state model [30].
3 The Improved Honey Badger Algorithm
In response to the desire of designing a higher performance algorithm, an improved Honey Badger Algorithm with multi-strategy (IHBA) is proposed, which are detailed are following.
The original HBA obtains the initial solutions randomly, which are difficult to ensure the individuals covering over the entire solution space evenly. Chaotic mapping is often used in optimization algorithms to disperse the population and reduce aggregation. The main two classes of chaotic mappings are Logistic chaotic mapping and Tent chaotic mapping, while the latter creates a more uniform chaotic sequence and has a faster convergence speed [26], the Tent chaos mapping is used as Eq. (4).
where Ndim is the dimension of individuals in the Tent sequence. R is the random number between 0 and 1. The Tent chaotic mapping makes the population widespread and ensures that the population wanders randomly in the global range.
3.2 The Improved Control Parameter
With regard to Eq. (2), the change of control parameters is smooth. Once the population falls into the local optimum, it is difficult to maintain the diversity of the population, resulting in poor quality of solutions. In this paper, the control parameters with random perturbation are introduced. The improved control parameters are shown in Eq. (5).
where t is the current search time,
In Fig. 1, the control parameter α changes irregularly, indicating the ability of the population to escape the current local solution. When the absolute value of α is greater than 1, the individual could prey outside the scope of the population position. On the contrary, individuals search in a more intensive range. Meanwhile, α and A have randomness during the whole search process. When α changes more smoothly, the algorithm will perform a more refined exploration and exploitation.
3.3 The Composite Mutation Factors
Three hybrid mutation factors are utilized for multimodal problems as it is difficult to adapt complex and multivariate functions using simple parameters, which are defined as Eq. (6).
where α and A are control parameters,
The three hybrid mutation factors could escape the local optimum. Since the definite value of these factors is uncontrollable, they have a better ability to adapt to different problems, and the probability of jumping out of the local solution may be higher. Meanwhile,
3.4 An Improved Swarm-Based Strategy
The updating strategy of position is the core of the swarm-based algorithm. A diversified updating strategy of position is proposed to enhance the advantage between convergence and diversity.
3.4.1 The Proposed Exploration Strategy
To improve the diversity of population in exploration, an improved exploration strategy is defined. In detail, three mutation factors are added which can automatically adjust their values with the increasing of the iterations. During the stage of exploration, the IHBA updates the position of individuals according to Eq. (7).
where Eq. (7) is formed by Eq. (2) and Eq. (6), which expands the scope that the individuals preying on. It is known that the search scope is wider, the diversity of the population is better. At latter stage, the population is more aggregated, the distance
3.4.2 The Proposed Exploitation Strategy
The population location information is introduced in the exploitation phase, which enhances the exploitation ability of IHBA and is defined as Eq. (8).
where
IHBA ensures the global search under the dual pressure of fuzzy mapping and mutation factors after combining Tent chaotic mapping, self-adaption time parameters, and three mutation factors. According to the above description, the process of IHBA is as follows:
Step 1: Initialize the position of the population based on the Tent chaotic mapping.
Step 2: Calculate the control parameters α and A via Eq. (5).
Step 3: Calculate the odor intensity
Step 4: Start to perform the exploration task according to Eq. (6) and calculates the fitness value.
Step 5: Update the best individual during the exploration stage. The elite selection is used to obtain the best individual among the original population and the honey badger group generated by step 4 and step 5.
Step 6: Start to perform the exploitation task via Eq. (7). After calculating the fitness value of the population and selecting the elite individual, the best individual
Step 7: Rank the new population according to the fitness value of the individual.
Step 8: Record the global best value. Return to step 2 if the algorithm does not satisfy the termination condition. Otherwise, output the best position and best value.
On the whole, IHBA introduces the Tent chaotic mapping to disrupt the aggregation behavior of the population, which means that individuals can be distributed throughout the search space evenly. By modifying the time parameters and adding three mutation factors, the scope of population searching can be expanded, and the randomness of the algorithm can be greatly improved.
4 Simulation Experiments and Analysis
To verify the performance of the IHBA, simulation experiments are executed by solving 10 benchmark functions. The detail of the benchmark functions is shown in Table 1. IHBA is compared with EO [11], MRFO [16], RSA [17], HBA [22], NCHHO [26], IGWO [27] and DSSA [28]. The former 4 algorithms are basic versions, which are prominent and outstanding in relevant literature. The latter 3 algorithms are improved variants, which are novel and effective. For the fairness of the experiment, the 8 algorithms are run 30 times independently, the max of iterations is 2000, and the size of the population is 50. The specific parameters of algorithms are setting in Table 2. In addition, all experimental results in this paper are carried out on a personal computer with Intel (R) Core (TM) i7-8700 @3.2 GHz CPU, 16.0 GB RAM under Windows 10 system. All algorithms are achieved from MATLAB 2016.
The mean value (Mean) and standard deviation (STD) of the results are the evaluation metrics (Table 3). Meanwhile, to reveal significant differences between all algorithms, Wilcoxon Rank-sum Test, the Friedman Test and Mann-Whitney U Test are adopted. A significance level of the Wilcoxon Rank-sum Test is set to 0.05 and the p is the probability of overturning the hypothesis. When p ≤ 0.05, it indicates a significant difference between the algorithms. Wilcoxon Rank-sum Test, Friedman Test and Mann-Whitney U Test used 300 data to evaluate all the algorithms, which include 10 benchmark functions, each of which had 30 results.
In Table 3, the metrics of Mean reflects the average optimization capability of the algorithm, and the STD reflects the stability of the algorithm. Under the same conditions, IHBA outperforms the others in the 10 benchmark functions. For the unimodal benchmark function
The last three rows of Table 3 show the results based on Wilcoxon Rank-sum Test and Friedman test. As for the symbols “+”, “=”, and “−”, denote that the performance of IHBA is better, the same good, and worse than the compared algorithm respectively. In Table 3, it could be seen that IHBA outperforms the others on the whole, which have comparable performance on 9 benchmark functions commonly. In the worst case, it outperforms the other algorithms on 5 functions with 3 approximate and 2 worse results. In the Friedman Test, the smaller the average ranking value, the better the overall optimization performance of the algorithm. IHBA has smaller sorting results. Furthermore, the p of NCHHO is 0.01 by the Wilcoxon Rank-sum Test, indicating that there is a significant difference between NCHHO with IHBA. The values of p in HBA, MRFO, RSA, and EO are 0.1, which is greater than 0.05, indicating that there was no significant difference between this compared algorithm and IHBA. The result of the Mann-Whitney U Test shows that there are significant differences between IHBA and the comparison algorithm, except for DSSA. All in all, the overall performance of IHBA is better than the compared algorithms.
Fig. 2 shows the convergence curves of different algorithms. IHBA and DSSA get better solutions than the others significantly in
5 Simulation in Engineering Optimization Problems
To compare the performance of algorithms involved in the paper on engineering problems, simulation experiments are performed on 5 engineering problems [31]. Specifically, F1 is the tension compression spring design problem, F2 is the three-bar truss design problem, F3 is the gear train design problem, F4 is the I-beam design problem, and F5 is the tubular column example. Parameters and indicators are the same as those set in Section 4.
In Table 4, it is clear that all algorithms can obtain acceptable solutions in 5 tasks. Specifically, in the F2 and F4 tasks, the mean values of 8 algorithms are 2.64E + 02 and 1.31E − 02. Moreover, the STD of IHBA is subtle, which indicates that there is better stability. For the F1 and F3 tasks, although IHBA does not gain the global optimal solution, there is a lower accuracy. For the F5 task, IHBA can obtain the global optimal solution in 30 runs as the STD is 0. In addition, there is no significant difference among RSA, NCHHO, and IHBA in the benchmark functions, but there is a gap in the constrained engineering problems when significance parameter is 0.05. It verifies the NFL again. Combining the Wilcoxon Rank-sum Test, Friedman Test and Mann-Whitney U Test, IHBA has a wider adaptability between the benchmark problem and the engineering problem compared.
Fig. 3 shows the box plots of the solutions of different algorithms on 5 engineering problems. The blue box area is the scope of solutions. The smaller the box height indicates the better stability of the algorithm, and the lower the box signifies the superior solution accuracy of the algorithm. The solutions of IHBA are more concentrated in all five tasks, which denotes that IHBA has the advantage in the stability of the algorithm. Specifically, compared with the other algorithms, the box positions of MRFO and IHBA are lower in the F1. The box positions of MRFO, EO, HBA, IGWO, and IHBA are lower in F2, F4, and F5. The solutions of MRFO and EO are more attractive in F3 than others. In the mass, the performance of IHBA is not inferior to other novel algorithms in the above engineering problems. On the whole, IHBA uses a variety of strategies to enhance global search. Compared to other algorithms, IHBA performed better in terms of stability and optimization ability for the benchmark functions and engineering problems.
HBA is a new MAs with good expansibility and high efficiency suffering from the local optimum and unbalanced exploration and exploitation. In this paper, we propose a hybrid approach using multi-strategies to improve the performance of HBA. It introduces Tent chaotic mapping and mutation factors to HBA, meanwhile, the random control parameter is improved, moreover, a diversified updating strategy of position is put forward to enhance the advantage between exploration and exploitation. IHBA is compared with EO, MRFO, RSA, HBA, NCHHO, IGWO and DSSA on 10 benchmark functions and 5 engineering problems. On 10 benchmark functions, the numerical results reflect that IHBA gets effective solutions for most functions. The Friedman test reveals that IHBA has a desirable average ranking on entire optimization problems. The Wilcoxon signed-rank test and Mann-Whitney U Test demonstrate that the performance of IHBA is dramatically different from other algorithms. The box plots show that IHBA achieves higher stability than other algorithms in engineering optimization problems. However, IHBA is attracted to local solutions and has difficulty finding global solutions for tasks with a large number of local minima.
In the future, we will continue to improve the exploration and exploitation capabilities of IHBA and apply IHBA to a wider range of issues, including multi-objective optimization, feature selection, and hyperparameter optimization in machine learning.
Funding Statement: The work is supported by National Science Foundation of China (Grant No. 52075152), and Xining Big Data Service Administration.
Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.
References
1. A. Mahdavi-Meymand and M. Zounemat-Kermani, “Homonuclear molecules optimization (HMO) meta-heuristic algorithm,” Knowledge-Based Systems, vol. 258, pp. 110032, 2022. [Google Scholar]
2. P. Lu, L. Ye, Y. Zhao, B. Dai, M. Pei et al., “Review of meta-heuristic algorithms for wind power prediction: Methodologies, applications and challenges,” Applied Energy, vol. 301, pp. 117446, 2021. [Google Scholar]
3. C. Renkavieski and S. R. Parpinelli, “Meta-heuristic algorithms to truss optimization: Literature mapping and application,” Expert Systems with Applications, vol. 182, pp. 115197, 2021. [Google Scholar]
4. W. Gao, S. Liu and L. Huang, “Particle swarm optimization with chaotic opposition-based population initialization and stochastic search technique,” Communications in Nonlinear Science and Numerical Simulation, vol. 17, no. 11, pp. 4316–4327, 2012. [Google Scholar]
5. H. D. Wolpert and G. W. Macready, “No free lunch theorems for optimization,” IEEE Transactions on Evolutionary Computation, vol. 1, no. 1, pp. 67–82, 1997. [Google Scholar]
6. S. Li, H. Chen, M. Wang, A. A. Heidari and S. Mirjalili, “Slime mould algorithm: A new method for stochastic optimization,” Future Generation Computer Systems, vol. 111, pp. 300–323, 2020. [Google Scholar]
7. S. Mirjalili, “SCA: A sine cosine algorithm for solving optimization problems,” Knowledge-Based Systems, vol. 96, pp. 120–133, 2016. [Google Scholar]
8. P. Zhang, J. Wang, Z. Tian, S. Sun, J. Li et al., “A genetic algorithm with jumping gene and heuristic operators for traveling salesman problem,” Applied Soft Computing, vol. 127, pp. 109339, 2022. [Google Scholar]
9. Z. Zeng, M. Zhang, H. Zhang and Z. Hong, “Improved differential evolution algorithm based on the sawtooth-linear population size adaptive method,” Information Sciences, vol. 608, pp. 1045–1071, 2022. [Google Scholar]
10. C. Shang, L. Ma, Y. Liu and S. Sun, “The sorted-waste capacitated location routing problem with queuing time: A cross-entropy and simulated-annealing-based hyper-heuristic algorithm,” Expert Systems with Applications, vol. 201, pp. 117077, 2022. [Google Scholar]
11. A. Faramarzi, M. Heidarinejad, B. Stephens and S. Mirjalili, “Equilibrium optimizer: A novel optimization algorithm,” Knowledge-Based Systems, vol. 191, pp. 105190, 2020. [Google Scholar]
12. Q. Liu, J. Li, H. Ren and W. Pang, “All particles driving particle swarm optimization: Superior particles pulling plus inferior particles pushing,” Knowledge-Based Systems, vol. 249, pp. 108849, 2022. [Google Scholar]
13. S. Mirjalili, M. S. Mirjalili and A. Lewis, “Grey wolf optimizer,” Advances in Engineering Software, vol. 69, pp. 46–61, 2014. [Google Scholar]
14. D. Zhang, G. Ma, Z. Deng, Q. Wang, G. Zhang et al., “A self-adaptive gradient-based particle swarm optimization algorithm with dynamic population topology,” Applied Soft Computing, vol. 130, pp. 109660, 2022. [Google Scholar]
15. T. Si, B. P. Miranda and D. Bhattacharya, “Novel enhanced salp swarm algorithms using opposition-based learning schemes for global optimization problems,” Expert Systems with Applications, vol. 207, no. 8, pp. 117961, 2022. [Google Scholar]
16. W. Zhao, Z. Zhang and L. Wang, “Manta ray foraging optimization: An effective bio-inspired optimizer for engineering applications,” Engineering Applications of Artificial Intelligence, vol. 87, pp. 103300, 2020. [Google Scholar]
17. L. Abualigah, M. E. Abd, P. Sumari, Z. W. Geem and A. H. Gandomi, “Reptile search algorithm (RSAA nature-inspired meta-heuristic optimizer,” Expert Systems with Applications, vol. 191, pp. 116158, 2022. [Google Scholar]
18. Y. Luo, W. Dai and W. Y. Ti, “Improved sine algorithm for global optimization,” Expert Systems with Applications, vol. 213, pp. 118831, 2023. [Google Scholar]
19. H. Dong, Y. Xu, D. Cao, W. Zhang, Z. Yang et al., “An improved teaching-learning-based optimization algorithm with a modified learner phase and a new mutation-restarting phase,” Knowledge-Based Systems, vol. 258, pp. 109989, 2022. [Google Scholar]
20. H. Nenavath, K. R. Jatoth and S. Das, “A synergy of the sine-cosine algorithm and particle swarm optimizer for improved global optimization and object tracking,” Swarm and Evolutionary Computation, vol. 43, pp. 1–30, 2018. [Google Scholar]
21. K. J. Xue and B. Shen, “A novel swarm intelligence optimization approach: Sparrow search algorithm,” Systems Science & Control Engineering, vol. 8, no. 1, pp. 22–34, 2020. [Google Scholar]
22. A. F. Hashim, H. E. Houssein, K. Hussain, S. M. Mabrouk and W. Al-Atabany, “Honey Badger Algorithm: New metaheuristic algorithm for solving optimization problems,” Mathematics and Computers in Simulation, vol. 192, no. 2, pp. 84–110, 2022. [Google Scholar]
23. M. A. Nassef, H. E. Houssein, E. B. Helmy and H. Rezk, “Modified Honey Badger Algorithm based global MPPT for triple-junction solar photovoltaic system under partial shading condition and global optimization,” Energy, vol. 254, no. 3, pp. 124363, 2022. [Google Scholar]
24. T. Düzenli̇, K. F. Onay and B. S. Aydemi̇r, “Improved Honey Badger Algorithms for parameter extraction in photovoltaic models,” Optik, vol. 268, pp. 169731, 2022. [Google Scholar]
25. Z. Zhang and C. W. Hong, “Application of variational mode decomposition and chaotic grey wolf optimizer with support vector regression for forecasting electric loads,” Knowledge-Based Systems, vol. 228, pp. 107297, 2021. [Google Scholar]
26. A. A. Dehkordi, S. A. Sadiq, S. Mirjalili and K. Z. Ghafoor, “Nonlinear-based chaotic harris hawks optimizer: Algorithm and internet of vehicles application,” Applied Soft Computing, vol. 109, pp. 107574, 2021. [Google Scholar]
27. H. M. Nadimi-Shahraki, S. Taghian and S. Mirjalili, “An improved grey wolf optimizer for solving engineering problems,” Expert Systems with Applications, vol. 166, pp. 113917, 2021. [Google Scholar]
28. B. Jena, K. M. Naik, A. Wunnava and R. Panda, “A differential squirrel search algorithm,” Advances in Intelligent Computing and Communication, vol. 202, pp. 143–152, 2021. [Google Scholar]
29. Y. Chen, G. Feng, H. Chen, L. Gou, W. He et al., “A multi-objective honey badger approach for energy efficiency enhancement of the hybrid pressure retarded osmosis and photovoltaic thermal system,” Journal of Energy Storage, vol. 59, pp. 106468, 2023. [Google Scholar]
30. H. Ashraf, S. O. Abdellatif, M. M. Elkholy and A. A. El-Fergany, “Honey badger optimizer for extracting the ungiven parameters of PEMFC model: Steady-state assessment,” Energy Conversion and Management, vol. 258, pp. 115521, 2022. [Google Scholar]
31. H. Bayzidi, S. Talatahari, M. Saraee and C. Lamarche, “Social network search for solving engineering optimization problems,” Computational Intelligence and Neuroscience, vol. 2021, pp. 32, 2021. [Google Scholar]
Cite This Article
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.