Abstract

The marine predators algorithm (MPA) is a novel population-based optimization method that has been widely used in real-world optimization applications. However, MPA can easily fall into a local optimum because of the lack of population diversity in the late stage of optimization. To overcome this shortcoming, this paper proposes an MPA variant with a hybrid estimation distribution algorithm (EDA) and a Gaussian random walk strategy, namely, HEGMPA. The initial population is constructed using cubic mapping to enhance the diversity of individuals in the population. Then, EDA is adapted into MPA to modify the evolutionary direction using the population distribution information, thus improving the convergence performance of the algorithm. In addition, a Gaussian random walk strategy with medium solution is used to help the algorithm get rid of stagnation. The proposed algorithm is verified by simulation using the CEC2014 test suite. Simulation results show that the performance of HEGMPA is more competitive than other comparative algorithms, with significant improvements in terms of convergence accuracy and convergence speed.

1. Introduction

Solving optimization problems in engineering and scientific research is a common problem. An optimization problem is the process of finding the best value of a decision variable that satisfies the maximum or minimum objective value without violating the constraints. Traditional gradient-based deterministic algorithms show difficulty in solving practical problems [1]. With the development of science and technology today, the optimization problems we encounter are becoming more and more complex. These real-world optimization problems often involve many decision variables, complex nonlinear constraints and nonconvexity, dynamic objective functions, and expensive computational costs [2, 3]. Although these algorithms achieve faster processing speeds, they can easily fall into local optima. In addition, the performance of the algorithms depends heavily on the characteristics of the problem and the initial parameter values. However, metaheuristic algorithms, which do not depend on the characteristics of the problem, are simple in structure, flexible, and do not rely on gradient information and have therefore received widespread attention and flourished among scholars. As a result, it is widely used to solve various optimization problems, such as task planning [4, 5], feature selection [6, 7], parameter optimization [8, 9], and image segmentation [10, 11].

Over the past decades, many metaheuristic algorithms have been proposed. These algorithms can be divided into three categories: evolution-based algorithms, physics-based algorithms, and swarm-based algorithms. Evolutionary-based algorithms are a class of algorithms that simulate the laws of evolution in nature. Genetic algorithm (GA) [12] is a widely used evolution-based algorithm proposed by Holland. It updates populations by simulating the natural law of superiority and inferiority. With the popularity of GA and GA variants, more and more evolutionary-based algorithms have been proposed, including differential evolution (DE) [13], genetic programming (GP) [14], and evolutionary strategy (ES) [15]. In addition to these evolutionary algorithms, new evolutionary-based algorithms have recently been proposed, such as the artificial algae algorithm (AAA) [16] and monkey king evolutionary (MKE) [17]. Physics-based algorithms simulate the laws of physics in nature or in the universe. Inspired by the phenomenon of annealing in metallurgy, simulated annealing (SA) [18] is one of the best-known physics-based algorithms. Other physics-based algorithms have been proposed, including the gravitational search algorithm (GSA) [19], nuclear reaction optimization (NRO) [20], water cycle algorithm (WCA) [21], and sine cosine algorithm (SCA) [22]. Population-based algorithms simulate the social behaviour of species such as self-organisation and division of labour. Particle swarm optimization (PSO) [23] and ant colony optimization (ACO) [24] are two classical swarm-based algorithms. Inspired by these two algorithms, an increasing number of scholars have studied this topic and proposed different swarm-based algorithms such as grey wolf optimizer (GWO) [25], whale optimization algorithm (WOA) [26], sparrow search algorithm (SSA) [27], firefly algorithm (FA) [28], artificial bee colony algorithm (ABC) [29], and tuna swarm optimization (TSO) [30].

The marine predators algorithm (MPA) is a novel population-based natural heuristic optimization algorithm proposed by Faramarzi et al. [31], which is mainly inspired by the different foraging strategies of marine predators and the optimal encounter rate strategies between predators and prey. Simulation results in the literature [31] show that MPA has better performance compared to GA, PSO, GSA, CS, SSA, and CMA-ES and thus has been widely used to tackle many practical engineering problems such as photovoltaics [32, 33], power systems [34], image classification [35], and task scheduling [36].

Although MPA has been applied in several fields, there is less analysis and improvement on the shortcomings of MPA. MPA mainly searches near the optimal individuals when performing population position update, without using the effective information of more individuals. The insufficient diversity leads the MPA into a local optimum. The FADs process is designed to enhance the diversity of the population, but it does not determine the superiority of the offspring and the parent, which is not conducive to the optimization and convergence of the algorithm to a certain extent.

Currently, there are three main approaches to improve the performance of natural heuristic optimization algorithms. The first is parameter tuning. Tang et al. [37] used chaos mapping to optimise key parameters of sparrow search algorithm. Ewees and Elaziz [38] investigated the effect of different chaotic mapping tools on the parameter settings of the algorithm. The second approach is to design different search operators. Zhang et al. [39] used a triangular variational strategy and a logarithmic spiral strategy to improve the development and exploration performance of the algorithm. Nor et al. [40] proposed an adaptive switching particle swarm algorithm based on a hybrid update sequence. The third approach is to introduce other techniques. The fractional order is an effective tool that has been used in other areas [41, 42]. Deep neural networks can also be combined [43, 44]. In addition, traditional gradient-based methods can be combined with metaheuristic algorithms. Inspired by these ideas, this paper proposes a hybrid MPA combined with the estimation of distribution algorithm for improving the basic MPA performance. And, we use Tent mapping and Gaussian random walk to further improve performance. The performance of HEGMPA was evaluated on the CEC2014 test suite and compared with five advanced algorithms. The superiority of the proposed algorithm is verified by numerical analysis, convergence analysis, stability analysis, and statistical analysis.

The left part of this paper is organized as follows: A review of the MPA is presented in Section 2. Section 3 shows a detailed description of the proposed algorithm. In Section 4, the effectiveness of the proposed improvement strategy is verified using CEC2014 test suite. Finally, we summarize this work in Section 4 and present directions for future research.

2. Marine Predators Algorithm (MPA)

In this section, the procedure of the basic MPA is presented. MPA is a novel swarm-based metaheuristic algorithm mainly inspired by the different foraging strategies of marine predators and the optimal encounter rate strategy between the predator and prey.

Similar to most metaheuristics, the initial solution of MPA is distributed as evenly as possible in the search space. The initialization formula is as follows:where and denote the upper and lower bounds of the population variables, respectively. is a random vector obeying a uniform distribution.

The MPA search process is divided into three phases based on different speed ratios: (1) a high-speed phase, where the prey speed is faster than the predator speed; (2) a unit speed ratio phase, where the prey speed and the predator speed are similar; and (3) a low-speed phase, where the prey speed is slower than the predator speed. In each stage, the movement of the predator and prey in nature is imitated separately.

Phase 1. The high-speed phase occurs at the beginning of the iteration, when the prey obeys Brownian motion and the predator performs mainly exploratory behaviour. The mathematical model for this phase is shown as follows:where is a random vector that follows a normal distribution based on Brownian motion. is a constant taking the value 5. is a uniformly distributed random vector. denotes the number of current iterations. denotes the maximum number of iterations.

Phase 2. In the second stage, both exploitation and exploration are required, so the stock is divided into two parts. One part is used for exploitation, and the other for exploration. The prey is used for the exploitation process and the predator for the exploration process. The mathematical model of this stage is described as follows:The first part of the population carries out exploitation behaviour.where is a random vector obeying a Lévy distribution.
The second part of the population performs exploratory behaviour.where is an adaptive parameter that controls the predator step size.

Phase 3. As the last phase, the exploitation process is mainly carried out and the mathematical model of this phase is described as follows:In addition, environmental issues can cause changes in predator behaviour, and fish aggregating devices (FADs) are a factor that affects predator behaviour and are seen as a local optimum in this search area, assuming that the local optimum can be jumped out of by longer steps. The effect of FADs can be expressed mathematically as follows:where denotes the probability that affect the optimization process. is a binary vector including 0 or 1. When a random vector from 0 to 1 is generated and is less than 0.2, the array is changed to 0, and vice versa. is a uniformly distributed random number. and are two randomly selected individuals. The pseudo-code for the MPA is shown in Algorithm 1 and Figure 1.

(1)Initialize search agents (Prey) populations
(2)While termination criteria are not met
(3)Calculate the fitness, construct the Elite matrix, and accomplish memory saving
(4)if iter < itermax/3
(5)Update prey based on equations (2) and (3)
(6)else if itermax/3 < iter < 2 × itermax/3
(7)For the first half of the populations (i = 1, …, NP/2)
(8)Update prey based on equation (6)
(9)For the other half of the populations
(10)Update prey based on equation (8)
(11)else if iter > 2 × itermax/3
(12)Update prey based on equation (10)
(13)end if
(14)Accomplish memory saving and Elite update
(15)Applying FADs effect and update based on equation (11)
(16)end while

3. The Proposed MPA Variant

The basic MPA uses only the best individuals for iterative search, not making full use of valid information from the remaining individuals, resulting in reduced population diversity. The FADs process is performed at each iteration, which increases the computational cost. We use three strategies to improve the performance of the algorithm. Firstly, we take advantage of the good traversal and randomness of chaotic mapping to generate the initial solution of the population and increase the population diversity. Secondly, we use EDA to sample the dominant population information and correct the evolutionary direction. A Gaussian wandering strategy is used to enhance the population diversity when the algorithm stalls, helping the algorithm to jump out of the local optimum. Finally, a greedy strategy is used to ensure that the algorithm converges efficiently.

3.1. Population Initialization Based on Cubic Mapping

The initial population of most current intelligent optimization algorithms is randomly generated in the search space, and the quality of the initialized population has a great impact on the efficiency of the optimization algorithm. A uniformly distributed population is conducive to expanding the search range and thus improving the convergence speed and accuracy of the algorithm.

MPA, like other intelligent algorithms, suffers from a reduction in population diversity late in the iteration when solving complex problems, which can easily fall into local optima leading to premature maturity resulting in poor convergence accuracy. To improve its global search capability and avoid the reduction of population diversity in the postiteration period, the chaos operator is used to initialize the population, considering that it has the characteristics of randomness and regularity and can traverse all states within a certain range without repetition, so the cubic mapping chaos operator is used. Figure 2 shows the effect of cubic mapping and logistic mapping.

The cubic mapping formula is shown as follows:

The cubic mapping is used to initialize the prey population by generating a random vector of −1 to 1 in each dimension as the first individual, then iterating over each dimension of the first individual to obtain the remaining M − 1 individuals using equation (12), and finally mapping the values of the variables generated by the cubic mapping onto the prey individuals using

3.2. Estimation of Distribution Algorithm

The estimation of distribution algorithm is an algorithm that uses probability models to represent relationships between individuals. EDA has been employed for hybridization with other algorithms and has achieved better results [45, 46]. It uses the current dominant population to calculate a probability distribution model and generates a new population of children based on the probability distribution model sampling, thus iterating continuously to eventually obtain the optimal solution. In this paper, the distribution model is estimated using a weighted maximum likelihood estimation method and the top one-half population with better performance is taken as the dominant population. The mathematical model of the algorithm is described as follows:where denotes the weighted mean of the dominant population. is the population size. denotes the weight coefficients in the dominant population in the descending order of fitness values. is the weighted covariance matrix of the dominant population.

3.3. Medium Solution Gaussian Random Walk

A random walk strategy is used to help the algorithm jump out of stagnation and enhance its exploration capabilities when it falls into a local optimum late in the iteration. A stalled algorithm is considered to have stalled if the average fitness of the top half of the dominant population does not change in two consecutive iterations, and the random walk strategy is then used to update the population. The random walk strategy is a probabilistic model that simulates the random movement of organisms in nature and is widely used in the design and improvement of various optimization algorithms. In this paper, we propose a Gaussian random walk strategy for constructing new offspring using a medium population.

As the vector of differences between dominant and intermediate populations can improve the diversity of populations, information on intermediate populations is considered in this paper. The sampling points are related to the relative positions of the medium and dominant populations, extending the search range and providing a strong exploration capability. The mathematical model of the strategy is described as follows:where and represent randomly selected individuals from the dominant and intermediate populations, respectively.

At the end of each iteration, HEGMPA used a greedy strategy to retain the best NP individuals in the parent and offspring, thus forming a new population, which facilitated the global convergence ability of HEGMPA. In summary, the flow chart of the improved algorithm proposed in this paper is shown in the following. The pseudo-code for the HEGMPA is shown in Algorithm 2.

(1)Initialize search agents (Prey) populations
(2)While termination criteria are not met
(3)Calculate the fitness, construct the Elite matrix and accomplish memory saving
(4)if stagnation
(5)Update prey based on equation (18)
(4)else if rand < 0.5
(6) if iter < itermax/3
(7)  Update prey based on equations (2) and (3)
(8) else if itermax/3 < iter < 2 × itermax/3
(9)  For the first half of the populations (i = 1, …, NP/2)
(10)   Update prey based on equation (6)
(11)  For the other half of the populations
(12)   Update prey based on equation (8)
(13) else if iter > 2 × itermax/3
(14)  Update prey based on equation (10)
(15) end if
(16)else
(17) Update prey based on equation (17)
(18)Accomplish memory saving and Elite update
(19)end while

4. Simulation Experiments and Analysis of Results

To comprehensively validate the performance of the improved algorithms, we first verify the effectiveness of different improvement strategies and then verify the superiority and competitiveness of the improved algorithms by comparing them with recently proposed ones.

The CEC2014 test suite contains 30 test functions, which can be divided into four categories of test functions according to different characteristics: F1–F3 for single-peaked functions, F4–F16 for multipeaked functions, F17–F22 for mixed functions, and F23–F30 for combined functions. The definitions and optimal values of the functions are shown in Table 1. In the CEC2014 test, the maximum number of iterations is 600 and the population size is 500. All algorithms were run independently 51 times to record statistical values. The program was run on a MATLAB 2016b platform.

4.1. Comparison of HEGMPA Improvement Strategies

HEGMPA was compared with MPA-1 using cubic mapping to initialize the population, MPA-2 with fused EDA, MPA-3 using a moderately solved Gaussian random wandering strategy, and basic MPA to verify the effectiveness of the different improvement strategies. Table 2 records the mean error, standard deviation, and ranking of the different algorithms for solving the test functions. The last column shows the average ranking of each algorithm.

From Table 2, HEGMPA with a full search strategy has the best search performance, while the basic MPA ranks last. Specifically, for unimodal test functions F1–F3, MPA-2, which only incorporates the EDA algorithm, performs similarly to HEGMPA in unimodal functions and far outperforms the other compared algorithms, indicating that incorporating the EDA strategy can effectively improve the development capability of the algorithm. For the multimodal functions F4–F16, HEGMPA and EDA-2 also ranked in the top two positions, while MPA was the least effective, suggesting that using dominant population information to generate offspring is beneficial in enhancing the diversity of individuals in the population. For the combinatorial functions F17–F30, HEGMPA only underperformed MPA-2 on F21, F24, and F26, indicating that the Tent chaotic mapping and moderately solved Gaussian random walk strategies can improve the algorithm’s performance in solving complex combinatorial functions and effectively help the algorithm to jump out of the local optimum when the algorithm stalls. In summary, the improvement strategy proposed in this paper can effectively improve the MPA optimality finding performance.

4.2. An Analysis of HEGMPA Compared with Other Algorithms

To further illustrate the superiority of the improved algorithms, five algorithms, TLMPA, VCS [47], MMPA [48], CPIJADE [49], and HFPSO [50], are selected for comparison with HEGMPA. CPIJADE is an improved JADE algorithm using a new framework. HFPSO is an improved particle swarm algorithm mixed with the firefly algorithm. To ensure fairness, the parameters of each algorithm are referred to the original literature, as shown in Table 3. In the experiments, NP = 500, dim = 30, and the maximum number of evaluations is 300,000. Table 4 records the average error and ranking of each algorithm in each test function for 51 independent runs.

The analysis in Table 4 shows that for single-peaked test functions F1–F3, HEGMPA outperforms all the comparative algorithms and can consistently find the optimal values of these three test functions, demonstrating the advantage of HEGMPA in solving pathological functions and once again verifying that the improved strategy can effectively improve the development capability of the algorithm. For the multipeaked test functions F4–F16, HEGMPA also outperforms all the comparative algorithms, indicating that the improved algorithms can maintain good enough population diversity to avoid falling into local optima. For complex combinatorial functions, each algorithm has its own advantages and disadvantages. HEGMPA obtains optimal solutions on F17–F22, F26, and F29–F30. VCS achieves better solutions on F23–F25. MMPA outperforms the other comparative algorithms on F23, F25, and F27–F28. CPIJADE performs best on F22. F17–F30 are more complex combinatorial test functions, and HEGMPA achieves the best results on eight of them, providing better evidence of HEGMPA’s potential to solve complex optimization problems in the real world.

To further illustrate the convergence performance of the algorithms, Figure 3 shows the average error convergence curves of the six algorithms for solving the CEC2014 test set. HEGMPA has better convergence accuracy and faster convergence on F1–F4, F6–F7, F13, and F16–F21. In solving F5, F8–F10, and F12, HEGMPA converges faster and with better convergence accuracy in the later part of the iteration, although the convergence speed is slower in the early part of the iteration. In summary, HEGMPA outperforms the comparison algorithms in terms of convergence accuracy and convergence speed.

To analyse the distributional properties of the solutions solved by the improved algorithms, box plots were drawn based on the results of 51 independent solutions for each algorithm, as shown in Figure 4. For each algorithm, the centre marker of each box indicates the median of the results of 51 solved functions, the bottom and top edges of the box indicate first- and third-degree points, and the symbol “+” indicates bad values that are not inside the box. As can be seen from Figure 4, HEGMPA has no outliers when solving 17 of the test functions (F1–F3, F7–F9, F11, F13–F15, F18, F21, F23–F24, F26–F27, and F29), indicating that the distribution solved by HEGMPA is very concentrated; meanwhile, for the other test functions where bad values exist, the HEGMPA has a small median, indicating that the quality of HEGMPA’s solutions is relatively better. Therefore, the improved algorithm proposed in this paper has a strong robustness.

To avoid chance in testing, this paper uses the Wilcoxon signed-rank test to verify whether the improved algorithms are statistically significantly different from the comparison algorithms. Table 5 presents the results of the Wilcoxon signed-rank sum test for each algorithm and HEGMPA. In the table, “+” indicates that HEGMPA outperforms the comparison algorithm in terms of optimization results, “−” indicates poorer results, “=” indicates similar results, and the symbol “R+” is a positive rank value indicating the extent to which HEGMPA is better than the comparison algorithm and “R−” indicates the opposite result. As can be seen in Table 4, HEGMPA outperformed the basic MPA in at least 23 of the 30 tested functions due to all comparison algorithms and in all tested functions, which statistically validates the excellent performance of the improved algorithm.

The computational efficiency of the algorithm is also another important aspect in evaluating the performance of the algorithm. Table 6 lists the average time taken by each algorithm in solving the test function, and the last column lists the average ranking of each algorithm. We can learn that HEGMPA takes more time to compute, ranking only fourth, while MPA ranks third, due to the increased computational cost caused by the introduction of EDA, and the computation of the covariance matrix based on the Gaussian distribution model increases the computational time taken by the improved algorithm. Although the introduction of the improved strategy leads to an increase in the computational time consumed by the basic MPA, the performance improvement it brings is significant and therefore the computational time consumed by the HEGMPA proposed in this paper is acceptable.

5. Conclusions

In this paper, we propose a variant of MPA, called HEGMPA. The performance of the algorithm is improved using Tent mapping, distribution estimation strategy, and Gaussian random walk. To evaluate the effectiveness of the improved strategy and the superiority of HEGMPA, it was validated using the CEC2014 test suite. It was compared with five state-of-the-art algorithms through numerical analysis, convergence analysis, stability analysis, and statistical tests. The simulation results show that the HEGMPA algorithm balances development and exploration and is competitive with other algorithms. On the other hand, there is still room for improvement in HEGMPA. The effect of initialization of small populations needs to be investigated. The calculation of the covariance matrix increases the computational cost. Therefore, how to reduce the computational cost while maintaining performance is an issue that needs to be further investigated.

In future work, we plan to further apply HEGMPA to medical image recognition detection. In addition, we plan to develop a multiobjective version of HEGMPA to address optimization problems in other domains.

Data Availability

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors acknowledge funding received from the Key Scientific and Technological Research and Development Projects of Jilin (20200401093GX).