Abstract

Differential evolution has made great achievements in various fields such as computational sciences, engineering optimization, and operations management in the past decades. It is well known that the control parameter setting plays a very important role in terms of the performance improvement of differential evolution. In this paper, a differential evolution without the scale factor and the crossover probability is presented, which eliminates almost all control parameters except for the population size. The proposed algorithm looks upon each individual as a charged particle to decide on the shift of the individual in the direction of the difference based on the attraction-repulsion mechanism in Coulomb’s Law. Moreover, Taguchi’s parameter design method with the two-level orthogonal array is merged into the crossover operation in order to obtain better individuals in the next generation by means of better combination of factor levels. What is more, a new ratio of the signal-to-noise is proposed for the purpose of fair comparison of the numerical experiment for the tested functions which have an optimal value with 0. Numerical experiments show that the proposed algorithm outperforms the other 5 compared algorithms for the 10 benchmark functions.

1. Introduction

With its efficiency and effectiveness, differential evolution (for short, DE) proposed by Storn and Price has been successfully applied in many different engineering fields [1, 2]. In order to keep improving the performance of DE, various efforts have been devoted over the past decades.

The researchers proposed three discrete DEs for the scheduling problems in the permutation flow shop environment [3]. These approaches focus on converting vectors of the continuous domain into permutation vectors of the discrete domain and self-adjusting the control parameters of these algorithms based on JADE [4] and SADE [5]. The results show that these proposed approaches are promising for scheduling problems.

For the parameter identification of solar cells, the original FSDE in reference [6] was improved, which is the hybridization between free search and DE with opposition-based learning by using a simple greedy strategy instead of a Gaussian noise update in the process of the potential solution generation for the proposed best solution update strategy [7]. Reference [8] also employed a DE with opposition-based learning for estimating optimum hourly energy generation scheduling of a hydro-thermal system.

The authors emphasized the population initialization on increasing the accuracy and convergence speed of DE and designed a new DE variant with a modified initialization scheme by combining the strengths of both chaotic maps and oppositional-based learning strategy in order to generate the initial population with a good quality of mean fitness and diversity of the solutions. Extensive simulation studies on benchmark functions show that the proposed algorithm outperforms its peers [9].

A cultural DE algorithm using a measure of population diversity was proposed as an alternative method for solving the economic load dispatch problems of thermal generators [10]. Based on the cultural algorithm technique using normative and situational knowledge sources, the proposed algorithm is able to balance well the trade-off between the exploration and the exploitation of the search space.

The scale factor and the crossover probability are two vital parameters in DE, which usually greatly improve the performance. Various strategies for parameter setting have been researched.

The values of and were suggested by Storn and Price [1]. The was set to the normal distribution rand number with expectation 0 and standard deviation 1 for multiobjective optimization in reference [11].

Qin and Suganthan considered and as the random numbers following normal distribution and according to the learning experience, where the parameter is set at 0.5 and updated once every 25 generations [5].

Kim et al. proposed that the scale factor is calculated by the formula , where and [12].

Ali and trn empirically obtained an optimal value  = 0.5 and calculated automatically the scale factor using the maximum and the minimum for focusing on the exploration at early generation and the exploitation at latter generation [13].

The parameters and were given, respectively, following -stable distribution and , where and denote the successfully evolved individuals’ and based on some feedbacks from the optimization process [14].

The scale factor was set using the Tsallis distribution in economic view for the optimization model in shell-and-tube heat exchangers [15]. is fist initialized with uniform random values between 0.8 and 1.1, and then is determined by at each generation, where obeys a -Gaussian distribution or Tsallis distribution with the means and the variance , the parameter is linked to the type of distribution that assumes values from 1 to 3.

A self-adaptive scaling factor was utilized in reference [16] for maximizing the profit of the distribution company with the several constraints based on the basic idea of the penalty function approach for solving optimal planning of energy storage systems in order to improve the rate of convergence of DE, where , , and are an acceleration factor, a linear decreasing factor, and a deceleration factor, respectively.

Based on the different setups created by a simple orthogonal experimental design method, the paper [17] revealed that with and is promising to optimize the vector Jiles–Atherton vector hysteresis model from a workbench containing a rotational single sheet tester. Similarly, the self-adapting parameter strategy was used in reference [18].

Some researchers designed the novel selection operator or employed the classical derivative-free methods in DE or analyzed the search behavior in theory for improving the performance of DE [1922].

These versions of DE do improve the algorithm performance. However, each of them only is superior to the other in some special aspects. The best setting for the control parameters can be different for different problems. Even though the self-adapting parameter strategies seem to be able to overcome the problem of parameter setting, some new control parameters are used. Several references reported that choosing the proper control parameters for DE is more difficult than expected. How to set reasonably these parameters is a nuisance [2, 23, 24].

A differential evolution without the scale factor and the crossover probability is presented in the paper. The algorithm calculates dynamically the scale factor F using the attraction-repulsion mechanism in Coulomb’s Law and executes the crossover operation using Taguchi’s parameter design method based on the orthogonal array. The proposed algorithm avoids the parameter settings. Numerical experiments show that the performance of the proposed algorithm is superior to that of the other compared algorithms.

The paper is the extended version which has been further researched based on “almost-parameter-free differential evolution” proposed by Zhang and Liu [24]. There are four different points between them. Firstly, this paper describes in detail the idea and particulars of the proposed algorithm. Secondly, we regard the scale factor in the mutant equations (13) and (14) in Section 4 as the two different charges for the purpose of a better interpretation of the algorithms’ idea and a better numerical experiment results. Thirdly, the vital shortcoming of the original definition of the ratio of the signal to noise (SNR) is analyzed in Section 4 and reveals the fact that it has thought of the optimal value of the tested problem before being solved as 0, then presents a modified definition of SNR for the sake of fairness. Finally, a brief convergence analysis is given under two assumptions.

The main contributions of this paper, which distinguish from the related literatures, are summarized as follows:(i)Use the electromagnetism-like mechanism to decide on the step length in the direction of the difference for the mutation operation;(ii)Employ Taguchi’s parameter design with a two-level orthogonal array based on a new ratio of the signal to noise that is proposed for the crossover operation;(iii)Eliminate almost all the control parameters of DE except for the population size.

The remainder of the paper is organized as follows. In Section 2, differential evolution algorithm is briefly introduced. Taguchi’s parameter design method is described in the next section. In Section 4, a DE without the parameters is proposed and the convergence in probability is analyzed. In Section 5, the results of numerical experiments are given. Finally, we conclude this paper and consider the further research issues.

2. Differential Evolution

Like other evolutionary algorithms (EAs), DE starts with an initial population individual, followed by the successive operations of mutation, crossover, and selection. However, there are two main differences between them. (i) Mutation is caused not by the small changes of the genes in EAs, but by adding the weighted difference of two randomly selected individuals to a third randomly selected one in DE. The direction information from the current population is used to guide the search process. (ii) New individual is generated by adopting a greedy selection scheme in DE, which is only accepted if it improves on the fitness of the parent individual.

Storn and Price proposed several different mutation strategies [1]:DE/Rand/1: DE/Rand/2: DE/Best/2:

In the above, , and they are the random numbers distributing uniformly in , where is denoted by the population size. For the strategy DE//, represents the individual being perturbed and is the number of difference vectors used to disturb . Take DE/rand/1 as an example, it means that the target individual is randomly selected, and only one difference vector is used.

Although there are several variants of DE, a common variant, which is known as DE/rand/1, or “classic DE,” is the most widely used in practice. Hence, this DE is described as follows:(i)Initialization: like other EAs, classic DE initializes an initial population that distributes uniformly in the feasible domain.(ii)Mutation: for each parent vector , a mutant vector is generated according to (1) where the random indexes , , and are mutually distinct integers following uniform distribution in and also are different from the current index . The scale factor is used to control the amplification of the differential variation.(iii)Crossover: the trial individual is generated using the parent and mutant individuals as follows:In the above formula, is denoted by the -th component of the individual, represents a random number with uniform distribution in [0, 1] for each , the crossover probability is set to a given number in , and the integer is randomly chosen in , where denotes the dimension of the tested problem. The trial individual is a stochastic combination of the parent and mutant individuals. When is equal to 0, at least one of the components of the trial individual will differ from the parent because of the condition .(iv)Selection: DE implements a very simple selection procedure. The offspring is generated only if the fitness of the offspring is better than that of the parent. Due to the greedy selection scheme, all the individuals of the next generation are as good as or better than their counterparts in the current generation.

The above process ii–iv repeats until the number of function evaluations or the number of the iterations reaches a given constant, namely, the termination criteria are satisfied. Further detailed descriptions about DE can be found in references [1, 23].

3. Taguchi’s Parameter Design

Taguchi method [25] is a parameter design approach in the production and process conditions optimization. It can make high-quality products using less development and manufacturing costs. Two major tools used in the Taguchi method are the orthogonal array [26] and the signal-noise-ratio , which are briefly described as follows.

The orthogonal array is a fractional factorial matrix, which assures a balanced comparison among the factors or its levels. A two-level orthogonal array is a matrix consisting of 1 or 2 arranged in rows and columns. Each row represents the combination of factor levels in each experiment, and each column represents the special level of each factor. Let the element 2 in the orthogonal array be −1, then all column vectors are orthogonal to each other, namely, the dot product is zero. Generally, a two-level orthogonal array is denoted by , where , which is equal to , represents the number of experiments; is a positive integer; the number 2 shows that each factor has two levels: 1 and 2; is the number of the factors or columns. The two-level orthogonal arrays are commonly used in practice: , , , and . For more clearness, the following table (see Table 1) shows the orthogonal array with the canonical form.

There are 8 factors in the array . For each factors, it can choose either 1 or 2. In order to obtain the better or best the combination of factor levels, only 8 experiments are under considered in the two-level orthogonal array instead of all combinations of the factors which can reach up to experiments. The notation represents the -th experiment or row, and the -th column vector or factor. For simplicity, the sign denotes the level of the -th factor in the -th experiment. For instance, , , and . If each 2 in array is thought of as , for all and from 1 to 7.

The conception of the is originally introduced in communication and electronic engineering, which is defined as the ratio of the signal to noise and is used to evaluate the quality of communication. In 1957, Taguchi applied the conception to the design of engineering experiments, hence, Taguchi parameter design method was proposed. This method utilizes the to evaluate quality and applies the orthogonal array to arrange experiments. According to the type of characteristic, the can be classified into smaller-the-better, larger-the-better and nominal-the-best. Given a set of characteristics , then in the case of smaller-the-better characteristic the is as follows:

4. Differential Evolution without and

After the brief description about classical DE and Taguchi’s parameter design, the ideals and the advantage of eliminating the parameters in DE are described, respectively. Finally, the differential evolution without the scale factor and the crossover probability are proposed.

Besides the parameters and , classic DE has a control parameter which are closely related to the problem under consideration. The population size, , is typically larger than a threshold value in order to obtain a global optimum and improve the success rate of convergence. However, too large may increase the number of function evaluations. Generally, separable and unimodal functions require the smallest population sizes, while parameter-dependent multimodal functions require the largest populations. For simplicity, the parameter is set as a constant according to the dimension of the problem under consideration.

The parameter determines the amplification of the difference. A high (low) value of makes DE more exploratory (less exploratory). The parameter controls the distribution of coordinate points in the trial individual. A high (low) value of means that the coordinates of the mutant individual dominate the trial individual. Between the two parameters and , is much more sensitive to the problem’s properties and complexity such as the multimodality, while is more related to the convergence speed.

Finding the optimal values for these parameters is a difficult task as these values are problem specific, especially when one wants to strike a balance between reliability and efficiency. Thus, the performance of DE depends on how these control parameters are selected. However, how to set well these parameters is generally based on trial and error. An optimal parameter setting can be found via the boring preliminary numerical experiments for a special problem, whereas it is not probably optimal for the other problems.

In order to overcome these contradictions, we eliminate the scale factor and the crossover probability with exception of the population size by using the modified attraction-repulsion mechanism and Taguchi method. In the following subsections, how to eliminate these parameters is described in detail.

4.1. Eliminating the Scale Factor

According to the attraction-repulsion mechanism in Coulomb’s Law, electromagnetism-like (EM) algorithm [27, 28] first calculates the charge of each individual in terms of its objective function value and then determines the resultant force exerted on each particle by all other particles in the population. The charge of each particle determines its power of attraction or repulsion. The particles with better objective function values attract others while those with inferior function values repel.

Like the method of calculating the force, the electromagnetic force exerted on the particle by other particles is obtained by the vector addition following the parallelogram law. For example, the charge of is less than that of while is greater than that of in Figure 1. Thus is a repulsive force and is an attractive force acting on by and , respectively. The resultant force exerted on is . In a similar way, the resultant forces exerted on , and on can also be calculated.

The charge of each is determined by the objective function value of itself relative to that of the current best particle :where is the dimension of the problem. The force vector exerted on by is then determined by

From (6), the particles with the relatively good objective function values will attract the other particles in the population while the particles with the worse objective function values repel the others. The resultant force vector exerted on a particle by other particles in the population is calculated as follows:

However, each particle has only one particle exerting force on it in a version of EM proposed by Debels et al. [29]. In this approach, the charge of is calculated based on the relative difference in the objective function values and :where and denote, respectively, the worst and the best solutions, is chosen randomly from the population. By the new definition of , obviously, a better(worst) particle gives the higher(lower) value. Moreover, if , then is positive, otherwise, is negative. After calculating the charge of , the particle moves to the new particle , where

It is obvious that when is positive (negative), attracts(repels) . This modified EM remains the basic ideal of EM, moreover, it is more simple and easier to utilize. Hence, for DE/Rand/1, the mutant individual can be transformed towhere . If we regard the scale factor and in equation (10) as the two different charges as shown in equation (8), viz.then the equation (10) can be interpreted as the motion of the particle in the direction of the resultant force . The magnitude of the motion is determined by the scale factors and . Hence, the mutant individual is modified in our algorithm as follows:

Similarly, we also haveor

As described , equations (13) and (14) are easy to understand. The idea implied in equation (13) comes from : the individual moves in the direction of . The magnitude of the motion is not controlled artificially in DE/Rand/1, but is determined self-adaptively according to its charge obtained by the particle . The similar interpretation is also done for equation (14).

Besides the self-adaptation of F and the simplicity of calculation, preliminary numerical experiments show that the modified equations (12)–(14) can generally improve the performance of DE, and equation (12) might avoid DE(DE/Ra nd/1) searching wrongly in the direction of “up hill.” The detailed description is as follows.

For six hump camel back function (see in Appendix), it is well known that the optimal value is . Given and , then two cases are given.CASE 1 Let , , and . Thus, can be obtained by equation (12) (see Figure 2).CASE 2 Let , , and . Then, , see Figure 3.

Figures 2 and 3 show the contour of SHCB on with the corresponding function value marked. The stars denote the optimal solutions; the circle denotes the individual ; two outer squares 10 represent and , respectively; Two outer real line denote the shift of in direction of the force and , respectively. The mutant individual obtained by equation (12) is denoted by the diamond. The inner real line represents the shift of in direction of the resultant force . Two bunches of squares locating in outer dashed line denote the motions of the individual in directions of and with the different scale factor , respectively. The scale factor is chosen orderly from the set , the corresponding results are shown in Figures 2 and 3 by the squares with the different number marked. A bunch of squares between outer squares gives the different mutant individual (see equation (15)). All squares can be matched by the numbers locating in them.

It is worth noting that equation (15) is different from . Five different mutually random individuals are selected in DE/Rand/2 while three individuals in equation (15). However, If and are thought of as two new individual, then equation (15) is the same as DE/rand/1 in essence. Thus a comparison is done between equation (12) and equation (15). The two formulas have the similar structure and is easier to distinguish in the figures if some dissimilarities appear in.

From Figure 2, only if  = {0.2, 0.3, 0.4}, the mutant individual obtained by equation (15) is better, whereas that obtained by equation (12) is closer to the global optimal solution. In Figure 3, it is very clear that equation (12) is superior to equation (15). Though the function value of the individual obtained by equation (15) for is almost same as that of obtained by equation (12), it moves uphill wrong.

4.2. Avoiding the Crossover Probability Cr

Taguchi method can obtain the better combination of the factor level with less cost. In the paper, a two-level orthogonal array is used. Since the number of factors (or variables) is , where is an integer greater than 1, the number of experiments is dependent on the dimension of the problem. In our paper, is given as follows:

For instance, if , then ; if , then . In equation (16), the minimal value subjecting to is chosen for avoiding the possible repeating experiments.

In what follows, a simple algorithm generating the two-level orthogonal array is described. The algorithm forms the array by using Hadamard matrix .

Definition 1. if any two columns in a matrix consisting of 1 or −1 are orthogonal, then the matrix is called Hadamard matrix [30].
In the above definition, denotes the order of the Hadamard matrix . There are several operations on Hadamard matrices which preserve the Hadamard property:(i)Permuting rows (columns)(ii)Changing the sign of some rows (columns)(iii)The Kronecker productIf and are known, then can be obtained by their Kronecker product, namely, by replacing all 1s in by and all -1s by .

Example 1. If , then where denotes the Kronecker product. Hadamard matrix of high order can be similarly generated from that of lower order: , , , , etc.
After a Hadamard matrix is obtained, a two-level orthogonal array can be given by discarding the all-one column and changing −1 s to 2 s in . However, this obtained array is not generally canonical form. Therefore, the simple exchange of rows can fix it for consistency (see Table 1 and the gray part in ).
Recall the notations about the orthogonal array in Section 3: is denoted by the -th experiment, by the -th factor and by the level of the -th factor in the -th experiment. The effects of the factors can be defined as follows:For simplicity, the is calculated as follows:where , , and . This conception is used here to evaluate the level of the factor. If , the optimal level of the factor is 1, otherwise, the optimal level is 2. When each is determined, a new individual (an optimal or near-optimal combination) is generated.

Example 2. An example is shown to illustrate this process of Taguchi parameter design method acting on two individuals, where . Without loss of generality, let and . This problem has 7 variables (factors), thus according to equation (16): , the orthogonal array is chosen(See Table 1). If is equal to 1 in Table 1, then the corresponding in Table 2 is the -th component of the mutant individual , otherwise, the corresponding is , see bold in Table 2.
Next, calculate the function value and the SNR of each combination of the factor level in Table 2, respectively. All results appear in the two most right hand columns. Then the effect of each factor is determined in terms of equation (18) (take and as an example).Finally, we obtain the new individual or the trial vector . The optimal level of the factor is decided by its effect. Since , 2 is the optimal level of the factor ; , therefore the optimal level of the factor is 2. The optimal levels of the other factors can be determined in a similar way. The component of the new individual consists of either or for all , which is dependent on the optimal level of the factor . If the optimal level is 1, then the corresponding component of the new individual is that of the individual , otherwise, it is equal to that of the individual .
Obviously, Taguchi parameter design method executes only 8 experiments instead of all combinations of factor levels for obtaining a new individual with the lower function value 1 (see the last row in Table 2). It is necessary to mention that only the first columns is used in orthogonal array while the other columns are ignored if .
In reference [31], the hybrid Taguchi-genetic algorithm (HTGA) is proposed for global numerical optimization with the continuous variables, which uses the systematic reasoning ability of Taguchi parameter design to gain the better genes in the crossover operation. The comparison results between HTGA and OGA/Q [32] show that HTGA can find the optimal or the near-optimal solutions with less function evaluations and better average values. However, this superiority is not very obvious for the tested function with nonzeros optimal values. Let we recall the original definition of in the case of smaller-the-better characteristic, which is described in equation (4), and change it toIn Taguchi method, the item represents the average loss of quality, where denotes the ideal signal in the case of smaller-the-better characteristic. Therefore, equation (21) shows that HTGA has thought of the optimal value of the tested problem as 0 before this problem is solved. This is unfair and unreasonable. As described above, we found that the superiority of HTGA is not very obvious for those function with nonzero optimal value from Tables IV and V on page 273 and 275 in the reference [31], Hence, is modified as follows:where is defined as the current optimal value after the -th iteration.
In what follows, the differential evolutions without the scale factor and the crossover probability (for short, DE∖FCr) are proposed. For the sake of clarity, the flowcharts of DE and DE∖FCr(take DE∖FCr2 as an example) are also given in Figure 4, where cross() represents the crossover operation in equation (2) and Taguchi() denotes Taguchi parameter design method in equation (13) (see Algorithm 1).
In Step 2, a termination criterion is given since is the denominator in Eq.(8). When this difference approximates to zero, the numerical stability of the algorithm will lose.
Let be the population at the -th generation. Through Step 4-5 in , transforms into the next population . Since the limitation of numerical calculation accuracy and relies only on the state of , the population sequence generated by DE\FCr can be described as finite-state Markov stochastic process.
Suppose (i) the objection function has a unique global optimal solution. Let be the state space of the stochastic process , be the state space of the global optimal solution, and be the global optimal value.
Because of the limitation of state space or search space, the probability that the algorithm can find the optimal solution at the next generation is greater than 0 if it cannot find at the -th generation, hence, suppose (ii) for and , where .
Now, we consider the probability that the proposed algorithm can not find the global optimum at the generation.It is very obvious that the current known optimal solution still can be retained in the next generation from Step 5. Once finds the optimal solution, the will hold the current state . Hence, in the equation (23),Summarizing the result of equation (23), we haveBecause the sequence is strictly monotonic decreasing as t, soTherefore,Equation (26) shows that the population sequence generated by can convergence in probability to the global optimum.

Step 1: Initialization: population , population size , maximal generation , current generation and , .
Step 2: If or , then output the current optimal value .
Step 3: Mutation. For each in the population, calculate the mutant individual according to equations (12) or (13) or equation (14). The corresponding algorithm is denoted by DE\FCr1, DE\FCr2 and DE\FCr3, respectively.
Step 4: Crossover. Execute Taguchi parameter design method with the denoted as equation (22) for the individual and the mutant individual , so the trail individual is generated.
Step 5: Selection. If , then and .
Step 6: If , goto Step 3; otherwise, , goto Step 2.

5. Numerical Experiments

The proposed Algorithms are executed in Matlab R2017 for the known numerical benchmark functions listed in Appendix with the default parameters and . Based on this parameter setting, each needs function evaluations at each iteration for 30 dimensional tested functions. with the four strategies below are compared with our algorithms, respectively.Strategy 1(DE): DE/Rand/1. , . This is a recommend parameters setting for DE/Rand/1 in most of the references [113];Strategy 2(DEG): , [11];Strategy 3(DE0.4): , [12]Strategy 4(DEM): and is calculated by the following formula [13]:

For Strategy 1–4, the population size and the maximal generation are set as 100 and 2000, respectively. All algorithms are performed with 10 independent runs for each tested function with 30 variables. According to these settings, our algorithm has the almost same function evaluations as with Strategy 1–4, that is, for and for . Hence, the results listed in Tables 3 and 4 are obtained under the assumption of the not same but different function evaluations. Obviously, the proposed algorithm evaluates 90 function values less than . The average values of the obtained results are given in Tables 3 and 4. The number of evaluations(#EVALU.), the best function value(BEST), the worst function value(WORST), the mean of function values(MEAN) and the standard deviation of the best function values(STD.) are used for the comparisons among these algorithms.

Table 3 summarizes the results obtained by the 7 algorithms for -. For , obtains the best mean of function values and the smallest standard deviation; , and find the better results; The means and standard deviations given by both and are worst among all algorithms; After 20 independent runs finds a better function value, however, it obtains the worst standard deviation because it encounters twice the worst function value 15.37. For , and find the best function values with the almost same precision E − 15; The precision given by is E − 14; and give the precision of E − 12; However, the lower precisions provided by and are E − 9 and E − 7 respectively. All of algorithms obtain the best function values with the almost same precision for tested function . For , the results given by are best, and those provided by are a little bit less promising; find the less promising optimum with the precision E − 14, while the precisions provided by and reach only E − 2; fails in finding the optima of among 20 independent runs, and traps into the local optima. Both , , and solve efficiently with the precision E − 24, E − 22, E − 16 and E − 13 respectively; The results given by and are worse; Similarly, fails to solve .

In Table 4, Both and find the optima of , however the fewer number of function evaluations 148533 is used by ; The results obtained by and are sightly worse than those obtained by and ; However, , and have the lower precisions of about E − 3. finds the optimum of tested function with highest precision and smallest STD and the fewer number of function evaluations than the other algorithms; and reach the a little bit worse precision, and the anther algorithms fail to find the optimum of with 200100 function evaluations. For 30 dimensional Rosenbrock tested function , none of all algorithms is obviously superior to the other one, namely, all algorithms can not find a satisfactory optima. For , produces a best results, while and provides the sightly worse results; The precisions of the mean function values given by and reach E − 10 and E − 6, respectively; can not find the reasonable result, and only the mean value 60.796 is presented. and find the best means of function values of with the precision of E − 4; The precisions given by and are E − 2 and E − 1 respectively; , and give the almost same results and fail to find a satisfactory optima of .

In order to show further the efficiency of , the means of the current optimal values obtained by 7 algorithms with the almost same number of function evaluations for each tested functions are respectively given in Figures 5 and 6. As mentioned in previous section, is used for the numerical stability, hence each stops probably before the maximal generation is reached. For convenience to draw the following figures, the current optimal value is recorded repeatedly in succeeding generations if the algorithm stops in advance since we think that the algorithm cannot be improved greatly in succeeding running.

Since the proposed algorithm is not same as the compared algorithms in the number of function evaluations at each iteration, it is inconvenient to draw the evolution curves describing the variations of with the number of function evaluations in a figure window for the reasonable comparison among all algorithms. Therefore, the current optimal values with the number of function evaluations which is denoted by for respectively are drawn in Figures 5 and 6 without considering the number of function evaluations costed by initialization. In fact, the sequence is an arithmetic sequence with the initial term 9900 and the common difference 9900, where denotes the least common multiple. 990 and 100 represent the number of function evaluations of and those of at each iteration, respectively. It needs to be emphasized that each proposed algorithm evaluates 70(=100−30) less than the compared algorithm at each given iteration as above. Consequently, the current optimal value obtained by each compared algorithm under the given number of function evaluations is just recorded at certain generation which is for each and for . Hence, only according to the recorded current optimal value at each generation can the figures below be given easily.

From each figure(see Figures 5 and 6), outperforms the other algorithms for , and , whereas surpasses the other ones for and . For , , and , is on the top three of 7 algorithms in terms of the performance. However, For and , drops out of the top three almost into the last three. It is worth noting that both and don’t be considered because all algorithms, especially and , obtain the almost same results. We also find that can obtain the optimum with the higher precision at earlier generation than the other algorithms and enters easily into the local optimum at latter generation for , and . finds the satisfactory results of most of tested functions except and also has not the tendency toward the local optimum with the default parameters.

In a summary, does rather well in terms of the performance, and are a little bit less promising, and are even less promising, and and are worst.

Furthermore, the numerical comparison experiments are done between and (see Table 5). is proposed in reference [19], which employs three other candidate individuals to design a new selection operator for improving the ability to escape the local optimum. In Table 5, find the optima of the tested functions with higher precision. For , Both algorithms have the almost same precision, however, reduces 90 function evaluations. Obviously, give the global minimal value 0 with the fewer for .

6. Conclusion

For avoiding the settings of the parameters, the differential evolutions without and are presented. The proposed algorithms use the attraction-repulsion mechanism in Coulomb’s Law and Taguchi parameter design method for the purpose of eliminating the scale factor and the crossover probability, respectively. Numerical experiments show that the proposed algorithm , which can balance well between exploration and exploitation, is superior to the compared algorithms with other strategies and can find quickly the optima or the near-optima of the problems. Although a smaller population size 30 is given in the proposed algorithms for all 30 dimensional tested functions, this small population maybe lead to the prematurity of algorithm such as . However a larger population will expend too many function evaluations because of using the two-level orthogonal array which is related with the dimension of the problems. Obviously, In our algorithms the number of function evaluations of each proposed algorithm at each generation is . Therefore, as for future work, the following problems are going to be investigated: (i) decrease the function evaluations at each generation and increase the population size without the loss of the algorithmic performance; (ii) analyze the accelerated convergence behavior of the current optimal value after the -th iteration in equation (22).

Appendix

Data Availability

All the data, MATLAB codes, and plots used in the paper can be found at https://github.com/zhang-xiaowei/Code.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the projects of UESTC under grant nos. JPKC20203-18 and JPKC20192-40.