157
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Analysis of dependence of grey wolf optimizer to shift-transformations and its shift-invariant improved methods adaptively controlling the search areas

&
Pages 66-79 | Received 25 Sep 2023, Accepted 22 Jan 2024, Published online: 28 Feb 2024

ABSTRACT

As a metaheuristic method for the continuous optimization problem, the grey wolf optimizer (GWO) has attracted much attention from researchers because the method is reported to be superior to other methods. However, some works show that the GWO is too specialized only to problems having the zero-optimal solution, which can lead to a significant deterioration of the efficiency for other problems. In this paper, we, first, theoretically prove the shift-dependence of the GWO, which is the underlying cause of the over-specialization of the GWO, and we experimentally analyze the property by using a larger number of problems. Secondly, we propose am shift-invariant GWO, GWO-SR, and, modify the GWO-SR by adding two methods: an adjustment technique the size of the search area and a mutation process to enhance the diversity of the search (GWO-AS) Finally, we show advantages of two proposed GWOs by comparing them with other metaheuristic methods.

1. Introduction

The grey wolf optimizer (GWO) is one of the nature-inspired metaheuristic methods which exploits hunting techniques and social hierarchy of the grey wolves in order to solve global optimization problems [Citation1, Citation2]. The GWO has attracted significant attention from researchers because the method has been reported to have high performance in seeking high-quality solutions regardless its simple algorithm according to many studies [Citation3–8], and thus, it has been applied to a wide range of fields [Citation9–12].

However, a previous work [Citation13] shows that the GWO is not invariant to shift-transformations in the orthogonal coordinate representation of the optimization problem. In addition, the some works [Citation14, Citation15] show that the performance of the GWO is often considerably low when the optimal solution is far from the origin. In addition, a work [Citation16] showed that the search behavior of the GWO is too specialized for problems having the optimal solution at the origin through numerical experiments, and that the updating systems of the GWO widely depend on the distance of the origin to the best solution found so far. The property can lead to a significant deterioration of the efficiency of the search process for other problems whose optimal solution is not the origin. Thus, many improved GWOs, which have been investigated [Citation3–8], can have the same drawbacks if its updating systems are not shift-invariant similar to the original GWO. Furthermore, in the work, authors proposed a shift-invariant method GWO with reference points (GWO-R) to overcome the drawback, in which grey wolves are updated by shift-invariant systems exploiting the reference points instead of the origin.

In this paper, we first analyze the underlying causes of too specialized properties of the GWO in more detail. We theoretically show that the GWO is not invariant to shift-transformations, and experimentally clarify ineffective search of the GWO due to the property by using more shifted problems than in work [Citation16]. These results show that the performance evaluations of the GWO in many studies are not fair for a general-purpose metaheuristic method for global optimization problems.

Secondly, we propose a modified GWO-R, GWO-SR, to improve the insufficient intensive search of GWO-R in the final stages, in which the size of search areas are adjusted more reasonably with respect to the distance between the reference point and the best solutions. Then, we experimentally show that the search is effective if the reference point is chosen appropriately. Moreover, we improve the GWO-SR by adding an adjusting method of search areas and a mutation process in order to strengthen the diversity of the search, which is called the GWO with adjustment of the search area (GWO-AS). Finally, we show the good performance of the proposed GWOs, GWO-SR and GWO-AS, by comparing them with other metaheuristic methods such as particle swarm optimization (PSO) [Citation17, Citation18] and firefly algorithm (FA) [Citation19, Citation20].

This paper is organized as follows. In Section 2, we introduce the GWO, and in Section 3, we analyze the property of its search process, and point out the shift-dependence and the over-specialization of its search process. In Sections 4 and 5, we propose new modified GWOs exploiting a reference point or an adjustment of search area and a mutation process, which overcome the above drawbacks. In Section 6, through numerical experiments, we verify the effectiveness of the proposed methods. Finally, we conclude this paper in Section 7.

2. Grey wolf optimizer

The GWO is one of metaheuristic methods for continuous global optimization, in which the hunting technique and the social hierarchy of grey wolves are mathematically exploited in order to solve global optimization problems [Citation1, Citation2]. Regardless of its simple algorithm, the GWO is reported to have a high performance of finding desirable solutions of the optimization. In this section, we introduce the standard GWO.

Throughout this paper, we mainly consider the following continuous optimization problems having many local minima and the rectangular constraint. (P1)minx f(x) s.t. xX:=j=1n [xjl,xju].Here, f:n denotes the objective function to be minimized, and X n does the feasible region. In order to solve the problem, the GWO uses a number of search agents called grey wolves. The position of each grey wolf i, iW denotes a candidate solution of the problem, which are represented as x(i)(t) n at iteration t, where W denotes an index set of all grey wolves. These positions are updated in a way that models the social hierarchy and tracking, encircling, and attacking a prey of the grey wolves as follows: (1) d(p,i)(t):=|2r(p,1)x(p)(t)x(i)(t)|,(1) (2) Δx(p,i)(t):=a(t)(2r(p,2)1)d(p.i)(t),p{α,β,δ}(2) (3) x(i)(t+1):=x(i)(t)+13p{α,β,δ}Δx(p,i)(t),(3) where x(α)(t), x(β)(t) and x(δ)(t) denote the first, second and third best solutions obtained by grey wolves until iteration t, respectively, which are called the alpha, beta and delta wolves. In addition, 1 denotes the vector whose components are all one, and |u| for a vector u n denotes a vector whose components are |ui|, iN:={1,,n}, and ⊗ denotes the element-wise product of two vectors. The vector a(t) n is a coefficient vector whose components are linearly decreased from 2 to 0 over the course of iterations, and components of r(p,1), r(p,2) are random numbers which are independently and uniformly chosen from (0,1) for each p and i. In Matlab code [Citation21], the vector a(t) is given by a(t)=2(1t/tmax)1, where tmax denotes the maximal number of iterations. We also used the above linear function a(t) in numerical experiments at 3.3.

In numerical experiments in this paper, in order to keep the feasibility of x(i)(t), iW, x(i)(t+1) obtained in (Equation3) are projected onto the feasible region X of (P1) for the original GWO and the proposed GWOs. In the GWO, the size of search area of each wolf i depends on d(p,i)(t) and Δx(p,i)(t) by updating systems (Equation1)–(Equation3). In (Equation2), Δx(p,i) is uniformly reduced by using a(t) for all i W, p{α,β,δ}, which works to enhance the diversification of the search at the early stages and the intensification at the last stages according to only iteration t.

On the other hand, d(p,i) is determined by (Equation1) to adjust the search area for each grey wolf i according to its position or alpha, beta and delta wolves: When random variables rj(p,1), jN are all close to 0.5, then dj(p)(t)=|xj(p)(t)xj(i)(t)| and thus, the wolf i can search for solutions around x(p)(t), p{α,β,δ}. Conversely, when rj(p,1), jN are all close to 1 or 0, then d(p,i)(t)=|2x(p)(t)x(i)(t)| or d(p,i)(t)=|x(i)(t)|, which means that wolf i can explore a wider area. In many studies, the GWO is reported to have excellent intensification ability, while it is relatively weak in the extensive search, and thus, various improvement methods have been proposed such as improving updating mechanisms, introducing new operators, encoding scheme of the solutions and social hierarchy of grey wolves [Citation3–8].

However, the GWO is not invariant to the shift-transformations and is too specialized only for problems having the zero-optimal solution, which can cause the ineffectiveness of the search for other problems. Therefore, in the next section, we point out its drawbacks and analyze the properties.

3. Analysis of the search behavior of GWO

3.1. Shift-dependence of original GWO

In this subsection, we show that the GWO is not invariant to shift-transformations in the orthogonal coordinate representation of the optimization problem, and compare it with other metaheuristic methods. Although the previous work [Citation13] shows that the GWO is not invariant to affine transformations, in this subsection, we show more precisely that the GWO is not invariant to shift-transformations as well as affine transformations.

First, let's consider the orthogonal coordinate system which is shifted by a constant vector s0n from the original orthogonal coordinate system of (P1). In the shifted coordinate system, the position x in the original coordinate system is represented as x¯=x+s. Then, the problem (P1) can be expressed in the shifted-coordinate system as follows: (P2)minx¯ fs(x¯)=f(x¯s) s.t.x¯i=1n [xil+si,xiu+si],where note that problems (P1) and (P2) are essentially the same except the coordinate system.

Now, we define the invariance of a metaheuristic method for shift-transformations in the coordinate representation of optimization problems. If the updating system is invariant to any shift-transformation, the metaheuristic method is called to be invariant, where for the sake of simplicity, we assume that all random variables used in the updating system are same in the original and shifted systems.

Definition 3.1

Suppose that for a constant vector s0 and all search agents iI of a metaheuristic method, the positions x(i) and x¯(i) of search agent i are feasible solutions for (P1) and (P2), respectively, such that (4) x¯(i)=x(i)+s,iI,(4) where I denotes the index set of all search agents. In addition, xnew(i) and x¯new(i) are the next points which are obtained by updating x(i) and x¯(i), respectively. If xnew(i) and x¯new(i) satisfy the relation (5) x¯new(i)=xnew(i)+s,iI,(5) then the metaheuristic method is said to be shift-invariant.

The shift-invariance guarantees that the method can search for solutions in exactly the same way and independently to the shift-transformations. Conversely, if a metaheuristic method is not invariant to shift transformations, the search performance can vary depending on the choice of the coordinate representation of even the same problem, which can cause difficulties in executing an efficient search for each problem. Therefore, the property is significant with respect to the general purpose metaheuristic method. Moreover, many metaheuristic methods such as variations of PSO, FA, DE and others have the property because, in those methods, x(i)(t+1)x(i)(t) is determined by only difference vectors between two points, which are independent of the vector s. On the other hand, the GWO is not shift-invariant as follows:

Theorem 3.2

The GWO is not shift-invariant in the coordinate representation of the optimization problem.

Proof.

Now, we focus on a grey wolf i and assume that x(i), x¯(i), x(p) and x¯(p), p{α,β,δ} satisfy (Equation4). In addition, we have that r(p,1) and r(p,2) p{α,β,δ} are the same constant vectors in the different coordinate systems. Then, from (Equation1), we can derive that d(p,i)=|2r(p,1)x(p)x(i)|,d¯(p,i)=|2r(p,1)x¯(p)x¯(i)|=|(2r(p,1)x(p)x(i))+(2r(p,1)1)s)|,p{α,β,δ}.We can easily show that d(p,i)(t)d¯(p,i)(t) unless random variables r(p,1) are all 0.5 or rj(p,1)=(2xj(i)+sj)/(4xj(p)+2sj), jN. In addition, from (Equation2) and (Equation3), we have that (6) xnew(i)x¯new(i)=x(i)x¯(i)+13p{α,β,δ}(Δx(p,i)Δx¯(p,i)),=s+13p{α,β,δ}a(t)(2r(p,2)1)(d(p,i)d¯(p,i)).p{α,β,δ}.(6) The second term of (Equation6) is not zero unless random variables r(p,2) are all 0.5 or d(p,i)=d¯(p,i), p{α,β,δ}. As a result, we can see that (Equation5) does not necessarily hold. Thus, the GWO is not invariant.

3.2. Search area control of GWO

Next, let us consider that the updating system of GWO is also too specialized only for optimization problems having the zero-optimal solution. In the previous work [Citation16] inefficient searches of the GWO were analyzed by the relations between the size of search area and the degree of degradation of search.

Figure 1. Relations of d(p,i), x(p) and x(i).

Figure 1. Relations of d(p,i), x(p) and x(i).

Now, we focus on the size of the search area of grey wolf i, iW at iteration t. Since the size is mainly determined by d(p,i)(t)=|2r(p,1)x(p)(t)xj(i)(t)|, p{α,β,δ}, let us consider the upper bound dˆj(p,i) of jth component of d(p,i)(t) at iteration t, which is shown in Figure . Here, we define u(p,i)(t):=x(p)(t)x(i)(t). We can see that the jth component of 2r(p,1)x(p) is varied in (0,2xj(p)(t)) or (2xj(p)(t),0), jN, which are represented as light blue areas in Figure . Therefore, the upper bound dˆj(p,i)(t) is given by (7) dˆj(p,i)(t)=max{|xj(i)(t)|,|2xj(p)(t)xj(i)(t)|}=max{|xj(p)(t)uj(p,i)(t)|,+|xj(p)(t)+uj(p,i)(t)|}(7) (8) ={xj(p)(t)+|uj(p,i)(t)|if xj(p)(t)0,xj(p)(t)+|uj(p,i)(t)|if xj(p)(t)<0,=|xj(p)(t)|+|uj(p,i)(t)|,j{1,,n}.(8) From the results we can see that d(p,i)(t) increases as x(p)(t) or u(p)(t) increases. It is reasonable to choose a large d(p,i)(t) for a large u(p,i)(t) because the solution x(i)(t) which is far from the three best solutions x(p)(t), p{α,β,δ} should be changed drastically. However, it is not well-founded to choose a large d(p,i)(t) for a large x(p)(t) because x(p)(t) just represents the distance from the origin, which is simply a reference point in the coordinate representation, and thus, x(p)(t) does not contain any useful information for a search. This property can cause an inefficient search, which is discussed in Ref. [Citation16].

From these results, we can see that when the problem has the global optimal solution at the origin, an intensive search around the origin, and an extensive search far from the origin can effectively work in the GWO. On the other hand, when the optimal solutions are far from the origin, the intensive and extensive searches cannot be expected to work so effectively. In the next subsection, we experimentally clarify the ineffective search of the GWO due to the property by using a large number of shifted problems in more detail than in work [Citation16].

3.3. Numerical experiments

In this subsection, we investigate how search efficiency of GWO for (P2) changes as the shift vector s is varied through numerical experiments, which analyzes the quantitative relevance such as the relationship between the size of the shift vector and the mean size of the search areas, the means of d(α,i)(t,l) and Δx(α,i)(t,l), (9) dm(α)(t)=l=150iWd(α,i)(t,l)50 |W|,Δxm(α)(t)=l=150iWΔx(α,i)(t,l)50 |W|,(9) on 50 trials at iteration t, where d(α,i)(t,l) and Δx(α,i)(t,l) denote vectors obtained at iteration t on trial l, respectively. In addition, we observed the mean of the best function values obtained on 50 trials at iteration t, (10) fm(α)(t)=l=150fs(x(α)(t,l))50,(10) where x(α)(t,l) denotes the best solution obtained at iteration t on trial l. The maximal number of iteration tmax is 5000, the number of wolves |W| was 80, and we used a(t)=2(1t/tmax)1 as mentioned at Section 2.

We used 18 basic functions of CEC'17 benchmark problems [Citation22] (C1: BentCigar, C2: Sum of Different Power, C3: Zakharov, C4: Rosenbrock, C5: Rastrigin, C6: Expanded Schaffer F6, C7: Levy, C8: Modified Schwefel, C9: High Conditioned Elliptic, C10: Discus, C11: Ackley, C12: Weierstrass, C13: Griewank, C14: Katsuura, C15: HappyCat, C16: HGBat, C17: Expanded Griewank plus Rosenbrock, C18: Schaffer F7 ). In the experiments, relatively simple functions were used to compare the behavior of the original GWO for shift transformations. All functions have a hypercube constraint whose sides are 100, and functions C4, C15–C17 have the optimal solution at 1 or 1 n, while others have the zero-optimal solution. In addition, we used other two problems: O1:2n-minima Func. and O2:Schwefel Func., which have the optimal solution far from the origin as shown in Table . All problems have a hypercube constraint whose side is the same constant cs, namely, xil=xiu=cs, iN. As a shift vector s, four vectors ρcsu, ρ{0.0,0.1,0.5,0.9} were used, where u is a randomly selected unit vector, and thus, ρ determines the relative size of s toward cs. Note that (P2) with ρ=0 is equivalent to (P1).

Table 1. Benchmark problems.

Figure  depicts the results of seven problems for which typical transitions of each of (Equation9) and (Equation10) were observed. They are selected from all results obtained by the GWO to 20 benchmark problems with four shift vectors, respectively.

Figure 2. Means of Δx(α,i)(t), d(α,i)(t), and f(x(α)(t)) for four shift vectors.

Figure 2. Means of ‖Δx(α,i)(t)‖, ‖d(α,i)(t)‖, and f(x(α)(t)) for four shift vectors.

Furthermore, Table  indicates the mean function values and their standard deviations (SD) obtained by GWO for 20 benchmark problems with four shift vectors on 50 trials, where the first and second best values of the means or SDs obtained for the same problem with four shift vectors are indicated by bold and italic numbers, respectively.

Table 2. Comparison of results obtained by GWO for 20 benchmark problems (P1) and (P2) with three kinds of shift vectors (ρ=0.1,0.5,0.9).

First, we focus on the results of problems C1–C18 which have the optimal solution at the origin or close to the origin. From the figure and table, we can see in most of problems with a small ρ, the GWO can find exactly the global optimal solutions or considerably high-quality solutions, and that Δx(α,i)(t) and d(α,i)(t) are large at early stages of the search, gradually decrease and finally approach zero, which shows that the adjustment of the size of search area is appropriate. On the other hand, GWO finds considerably low-quality solutions for the problems with a large ρ. In this case, although Δx(α,i)(t) and d(α,i)(t) are decreasing, the rate of decrease becomes slower as ρ becomes larger. In particular, d(α,i)(t) is relatively large until the final stage, which means that the search is not sufficiently intensified. Moreover, the function value does not decrease as smoothly as the problem with smaller ρ, and the function values obtained finally are considerably worse for the problem with a larger ρ. These results show that adjustment of the search area is not appropriate, which are consistent with the previous discussion.

Next, let us consider the two problems, O1 and O2, in which the optimal solution is far from the origin. In these cases, the mean of obtained function values is smallest for (P2) with σ=0.9, and the obtained solutions are not overwhelmingly high quality, which are considerably different from the case for problems having the zero-optimal solution. In these cases, we find that the search of GWO is not so efficient regardless of the shift transformations.

From the above observations, the following conclusions can be derived. Since, in general, many global optimization problems can be expected to have many local solutions other than the origin, the technique of adjusting the search area in GWO as mentioned above can be regarded as a fatal drawback as a metaheuristic method, and it is difficult to find high-quality solutions by the GWO to many real-world problems. Therefore, in the subsequent sections, we propose modified GWOs to overcome the drawbacks which were pointed out in this section.

4. Shift-invariant GWO

4.1. Modification of GWO-R

In this subsection, we first introduce the GWO-R, which was proposed in Ref. [Citation16], and show that it is shift-invariant to overcome the drawback of the original GWO. Secondly, we theoretically and experimentally point out that the intensification of the GWO-R is insufficient in the final stages and propose its modified method such that its search areas are more reasonably controlled than GWO-R. Then, we show theoretically the shift-invariance of the GWO-R and the modified GWO-R.

In GWO-R, the point μ(t) called a reference point is used to adjust adaptively the search area of each grey wolf i W on the basis of the distance of μ(t) to x(p)(t), p{α,β,δ} or x(i)(t): (11) d(p,i)(t):=|2cmb(t)r(p,1)(x(p)μ(t))(x(i)(t)μ(t))|,(11) (12) ξ(p,i)(t):=x(i)(t)+2b(t)(2r(p,2)1)d(p.i)(t),p{α,β,δ}(12) (13) x(i)(t+1):=13p{α,β,δ}ξ(p,i)(t),(13) where (Equation11) was derived by replacing 2r(p,1) with cmb(t)r(p,1) in (Equation1), b(t) is given by b(t)=1t/tmax, and cm is a positive value for adjusting the search area. The GWO using (Equation11)–(Equation13) is called GWO with reference points (GWO-R).

The reference point μ(t) was selected by using information obtained until iteration t through the search process. As the reference point, three candidates were proposed as follows:

  1. Centroid of all pbests:μ(1)(t)=1|W|iWp(i)(t),

  2. Centroid of all grey wolves:μ(2)(t)=1|W|iWx(i)(t),

  3. Centroid of three best solutions:μ(3)(t)=13(x(α)(t)+x(β)(t)+x(δ)(t)),

where p(i)(t), iW denotes the personal bests (pbests), which is the best solution found by grey wolf i until iteration t as follows: p(i)(t)=argmin{f(x)| x{x(i)(τ)}τ{1,,t}}.The pbests have been used in PSO, which exploits information obtained by each search agent. In addition, as reference points, authors introduced internally dividing point of two reference points, μ(i)(t) and μ(j)(t), selected from the three kinds of candidates, in which its weights vary linearly as follows: μ(i,j)=1tmax((tmaxt)μ(i)(t)+tμ(j)(t)),ij{1,2,3}.By using these kinds of reference points, updating systems (Equation11)–(Equation13) of the GWO-R are shift-invariant, as shown at the end of this subsection,

Moreover, in Ref. [Citation16], the relations between the upper bound of d(p,i)(t) and the difference vector between the best solution and the reference point, namely v(p):=x(p)(t)μ(t), were derived in the similar way in 3.2 as follows: (14) dˆj(p,i)(t)=max{|vj(p)(t)uj(p,i)(t)|,|(2cmb(t)1)vj(p)(t)+uj(p,i)(t)|},(14) The relations indicate that the search process can be classified into three stages: in the first stage the upper bound dˆj(p,i)(t) increases with a large proportional constant to |vj(p)|, which can be considered to be an extensive search, and in the second and third stages, it increases with a small proportional constant, which means an intensive search. These results show that independently of the distance between the optimal solution and the origin, GWO-R gradually shifts from diverse to intensive search as the search progresses. At the same time, the effective search by GWO-R was reported through numerical experiments in Ref. [Citation16].

Furthermore, we can observe that in the first and second stages, under the assumption that u(p,i)(t)=x(p)(t)x(i)(t) is fixed, dˆj(p,i)(t) is minimal when xj(p)(t)=μj(t). The observation means that the search area is minimal when the best solution is equal to the reference point. Such behavior is reasonable because if additionally u(p)(t)=0 holds, namely, x(i)(t)=x(p)(t)=μ(t), the size of the search area is zero.

On the other hand, under the same assumption, only in the third stage, dˆj(p,i)(t) is not minimal even if xj(p)(t)=μj(t), which is not reasonable. In such cases, the detailed search may be insufficient.

Therefore, in this paper, we modify the updating system (Equation11) such that dˆj(p,i)(t) is minimal when x(p)(t)=μ(t) as follows: (15) d(p,i)(t):=|(cmb(t)r(p,1)+1)(x(p)(t)μ(t))(x(i)(t)μ(t))|.(15) The GWO with updating systems (Equation12), (Equation13) and (Equation15) is called simplified GWO-R (GWO-SR), in which the search process can be classified into two stages, and under the same assumptions as above, the upper bound dˆj(p,i)(t) is minimal in all stages when x(i)(t), x(p)(t) and μ(t) are equal. The properties can be shown as follows.

Now, we analyze the search area of each grey wolf. First, let us consider the upper bound dˆj(p,i)(t) of d(i)(t) in (Equation15). In the same way to derivate (Equation14), we have that (16) dˆj(p,i)(t)=max{|xj(i)(t)μj(t)|,×|(cmb(t)+1)(xj(p)(t)μj(t))(xj(i)(t)μj(t))|}=max{|vj(p)(t)uj(p,i)(t)|,×|cmb(t)vj(p)(t)+uj(p,i)(t)|}.(16) Thus, dˆj(p,i)(t) can be evaluated by dividing into the following cases: (17) Ifcmb(t)=1, then dˆj(p,i)(t)=|vj(p)(t)|+|uj(p,i)(t)|,(17) (18) Ifδj(p,i)(t)vj(p)(t)[2|uj(p,i)|1cmb(t),0], cmb(t)>1,orδj(p,i)(t)vj(p)(t)[0,2|uj(p,i)|1cmb(t)], cmb(t)<1,thendˆj(p,i)(t)=|vj(p)(t)uj(p,i)(t)||vj(p)(t)|+|uj(p,i)(t)|,(18) (19) Ifδj(p,i)(t)vj(p)(t)[2|uj(p,i)|1cmb(t),0], cmb(t)>1,orδj(p,i)(t)vj(p)(t)[0,2|uj(p,i)|1cmb(t)], cmb(t)<1,thendˆj(p,i)(t)=|cmb(t)vj(p)(t)+uj(p,i)||cmb(t)vj(p)(t)|+|uj(p,i)|,(19) where we define δj(p,i)(t):=uj(p,i)(t)/|uj(p,i)(t)|.

Note that cmb(t)<1 and cmb(t)>1 mean the first and second halves of the search such that t<ts and t>ts, respectively, and ts:=(11cm)tmax is the iteration at which the two kinds of searches are switched, as mentioned below:

From (Equation17), (Equation18) and (Equation19), we can see that in the first half, dˆj(p,i)(t) increases with proportional constant cmb(t) (>1) for large |vj(p)(t)|, and does with 1 for small |vj(p)(t)|, respectively, which widens the search area at a relatively high rate according to v(p)(t), and that in the second half of the search, dˆj(p,i)(t) increases with the proportional constant 1 for large |vj(p)(t)|, and does with cmb(t)(<1) for small |vj(p)(t)|, respectively, which means that the effect of v(p)(t) on the search area is relatively small. These relations are depicted in Figure , in which cm was chosen to be 3.0, and the switching occurs at ts=(11cm)tmax=23tmax. The same cm was used in the numerical experiment at the next subsection. The results coincide with the above results. Furthermore, note that the results show that the upper bound dˆj(p,i)(t) is always minimal at xj(p)(t)=μj(t) under the assumption that u(p,i)(t) is fixed, which means that the intensive search of GWO-SR is more reasonable than that of GWO-R.

Figure 3. Relation between dˆ(p,i)(t) and v(p)(t) when cm=3 and u(p,i)=1.5.

Figure 3. Relation between dˆ(p,i)(t) and v(p)(t) when cm=3 and u(p,i)=1.5.

As discussed above, the search of the GWO-SR aims to suitably adjust the search area in a feasible way for more general cases. Note that cm determines not only the proportional constant cmb(t) of the search but also the switching iteration ts. Thus, an appropriate selection of cm can tune the balance between diversification and intensification of the search.

Finally, we theoretically show the invariance of GWO-R and GWO-SR. Since in the updating systems (Equation11), (Equation12), (Equation13) of GWO-R, or in (Equation12), (Equation13), (Equation15) of GWO-SR, the position of grey wolf i is updated by using only difference vectors between μ(t) and x(i) or x(p), we can show shift-invariance of them as follows:

Theorem 4.1

GWO-R and GWO-SR are shift-invariant in the coordinate representation of the optimization problem.

Proof.

In the same way to Theorem 3.2, we assume that x(i), x¯(i), x(p) and x¯(p), p{α,β,δ} satisfy (Equation4). In addition, μ(t) and μ¯(t) denote the reference points in the original and shifted coordinate systems, namely for (P1) and (P2), respectively. Then, since from the definition of the reference point, μ(t) is a weighted sum of points satisfying the relation (Equation4), we have that μ¯(t)=μ(t)+s, which means that x(p)(t)μ(t)=x¯(p)(t)μ¯(t), p {α,β,δ}, and x(i)(t)μ(t)=x¯(i)(t)μ¯(t), iW. Therefore, the relation d(p,i)(t)=d¯(p,i)(t) always holds at t + 1 if all random variables are the same in both systems. As a results, the position of any grey wolf satisfies the relation (Equation5).

4.2. Numerical experiments

In this subsection, we applied GWO-R and GWO-SR to the same problems (P) with no shift-transformation as 3.3, namely, twenty 50-dimensional problems. Note that even if they are applied to (P2) with different shift vectors, the same results are obtained because of shift-invariance of them. We compared GWO-SRs with different reference points and GWO-R with μ(1,2). In Ref. [Citation16], GWO-R with μ(1,2) was shown to have better performance than others. We compared the mean function values of the best solutions obtained by GWO-R and GWO-SR on 50 trials. The maximal number of iteration tmax was 5000, the number of wolves |W| was 80, the constant cm of GWO-R was 1.5, and cm of GWO-SR was 3, which was selected in the preliminary experiments.

Table 3. Comparison of results obtained by GWO-SRs with seven different reference points and GWO-R with μ(1,2) for 20 benchmark problems (P1).

Table  shows the result obtained by GWO-R using μ(1,2) and GWO-SRs using μ(i), i = 1, 2, 3 and four internally dividing points, μ(1,2), μ(2,1), μ(1,3) and μ(3,1), which are the first to fourth best results in ones obtained by GWO-SRs using internally dividing points of all combinations of two points, μ(i,j), ij{1,2,3}. We can see that GWO-SRs with μ(1), μ(1,2) and μ(1,3) obtained better solutions than others, which demonstrates that it is advantageous to use the centroid of pbests for the diversification in the first half of the search, and the centroid of x(i), iW or x(p), i{α,β,δ} for the intensification in its second half. At the same time, the results indicate that the selection of the reference point is considerably important because the reference point is a significant criterion to adjust the size of the search area.

Next, comparing GWO-R with GWO-SR using the same μ(1,2), we can observe that GWO-SR is superior to GWO-R, which indicates that the method of adjusting reasonably the search area works effectively in GWO-SR due to (Equation15). In addition, since most of the results of the GWO-SR are better than those of the original GWO for (P2) with different shift vectors (ρ0) shown in Table , we can conclude that the proposed GWO-SR is more useful than the original GWO. These results and the reasonable control of the search areas in GWO-SR show that if the reference point is a temporal prey, GWO-SR can be considered to search for solutions reasonably around the prey, which means that GWO-SR more naturally and effectively simulates the hunting mechanism of grey wolves than the original GWO or GWO-R.

On the other hand, GWO-SR has a search ability that is almost comparable to other metaheuristic methods, as shown in Section 6, while it does not have significant advantages over others. The limitation of its search ability can be mainly caused by the fact that the search areas are controlled only on the basis of the distance from the reference points, and in addition, there is room for improvement in the method. Therefore, in the next section, we add two techniques to the proposed GWO, which can control the search areas adaptively as the search progresses.

5. GWO with adjustment of search area

In this section, first, we propose a method which adaptively selects the size of the search area for each grey wolf by evaluating objective function values at candidates of its next point additionally. Its updating system is given as follows: (20) d(k,i)(t)=|(ckb(t)r(k,1)+1)(x(α)(t)μ(t))(x(i)(t)μ(t))|,(20) (21) ξ(k,i)(t)=x(i)(t)2b(t)(2r(k,2)1)d(k.i)(t),k{1,2,3},(21) (22) x(i)(t+1)=argmin{f(x)|x{ξ(1,i)(t),ξ(2,i)(t),ξ(3,i)(t)}}.(22) In this method, at the beginning, d(k,i)(t), k{1,2,3} are calculated by using the best solution x(α)(t) with different ck, k{1,2,3}, such that c1<c2<c3 by (Equation20) for grey wolf i. Then, three candidate points ξ(k,i)(t), k{1,2,3} for i are selected from the search areas determined by d(k,i)(t), k{1,2,3}, respectively, by (Equation21). Next, among three candidate points, the point with the smallest objective function value is selected as the next point of x(i)(t) by (Equation22). Namely, the next point of each grey wolf is the best point of three candidates selected from search areas with different sizes.

Here, note that since function values of three candidate points are evaluated for each grey wolf per iteration, with regard to the number of evaluations of function values the method requires three times greater than GWO-SR. In addition, although selection from more candidates based on all of x(p)(t), p{α,β,δ} might improve current solutions more significantly, it requires a larger amount of computational resources. Thus, this method uses only three candidates based on x(α)(t).

Next, let us remind that cm in (Equation11), which corresponds to ck in (Equation20), determines the switching iteration ts=(11/cm)tmax between the extensive and intensive searches, as discussed in 4.1. Since tsk:=(11/ck)tmax is different for different ck, the balance between extensive and intensive searches is also different for each ck. Hence, we can expect the various searches by selecting different ck. In the numerical experiments in the next section, we selected (c1, c2, c3)=(3.0,4.5,6.0), in which we have ts1=23tmax, ts2=79tmax and ts3=56tmax.

Secondly, we introduce a mutation procedure for a more extensive search because the diversity of the search may be relatively weakened due to selecting the next point from three candidates based on the same xα. In the procedure, at each iteration t, a mutation procedure is executed with a probability of (1t/tmax) at t as follows: A grey wolf i W and an index k {1,2,3} are selected uniformly at random, respectively, and the jth component of ξ(i,k)(t) is also selected uniformly at random among {1,,n}. Then, after updating ξ(i,k)(t) by (Equation21), a mutational modification of Δxmrm is added as follows: (23) ξj(k,i)(t)=ξj(k,i)(t)+Δxmrm,(23) where rm is a uniform random variable from (1,1), and Δxm is the upper bound of modification amount.

This modified GWO-SR is called the GWO with adjustment of the search area (GWO-AS). Here, note that we can easily show that the modified GWO-AS is also invariant to shift-transformations in the same way as Theorem 4.1 because the updating systems (Equation20)–(Equation22) and (Equation23) include only difference vectors between μ(t) and x(i) or x(p) similarly to GWO-SR.

6. Numerical experiments

6.1. Experimental conditions

In this section, we apply the proposed GWOs, GWO-SR and GWO-AS, and other metaheuristic methods, FA and PSO, to 18 basic functions of CEC'17 benchmark problems and the other two problems, which were used in 3.3 and 4.2. Note that in order to evaluate the characteristics of the proposed methods, the search behavior was investigated by comparing them with the original metaheuristic methods rather than the various modifications of them. Since the four methods differ in the number of function evaluation per iteration, we selected the maximum number of iteration for each method such that the total number of function evaluations fmax were equal in four methods. We set fmax=1200000 and 4800000 for 50 and 200-dimensional problems, respectively, in which 50 trials were performed for 50-dimensional and 30 trials were 200-dimensional problems. All methods used 80 search agents such as particles, fireflies and grey wolves. In addition, the parameter values of PSO were selected as the recommended ones [Citation23]. Those of FA and the proposed methods were done by preparatory experiments. For GWO-SR, cm=3.0 and the reference point was μ(1,2), while for GWO-AS, Δxm=0.6cs and (c1,c2,c3)=(3.0,4.5,6.0) the reference point was μ(2,1).

First, we compared the mean and standard deviations of function values obtained on 50 trials of four methods for twenty 50-dimensional benchmark problems, and compared those on 30 trials for nineteen 200-dimensional benchmark problems, where all four methods did not obtain any computable objective function value for 200-dimensional problem C2, which are shown in Tables  and , respectively.

Table 4. Comparison of PSO, FA, GWO-SR and GWO-AS for 50-dimensional problems (fmax=1200000).

Table 5. Comparison of PSO, FA, GWO-SR and GWO-AS for 200-dimensional problems ( fmax=4800000 ).

Next, we evaluated the transitions of the best function values obtained by four methods with fmax=1200000 for all 50-dimensional problems, and selected six typical results in them, which are shown in Figure . In the figures, its horizontal axis denotes the function evaluation count per agent, and its vertical axis does the mean of the best function values in the search process of each method on 50 trials. Finally, we focus on the number of times ξ(k,i)(t+1), k{1,2,3} is selected in (Equation22) of GWO-AS at iteration t + 1 as follows: W(k)(t+1):={iW|x(i)(t+1)=ξ(k,i)(t)}, k{1,2,3}, and compare the mean of |W(k)(t+1)| on 50 trials at each iteration in GWO-AS for four typical results, which are shown in Figure .

Figure 4. Means of the best function values obtained by four methods, PSO, FA, GWO-SR and GWO-AS, until function evaluation count per agent for 50-dimensional six benchmark problems.

Figure 4. Means of the best function values obtained by four methods, PSO, FA, GWO-SR and GWO-AS, until function evaluation count per agent for 50-dimensional six benchmark problems.

Figure 5. Means of |W(k)(t+1)|, k{1,2,3} for three candidates on 50 trials at each count when GWO-AS is applied to 50-dimensional four benchmark problems.

Figure 5. Means of |W(k)(t+1)|, k∈{1,2,3} for three candidates on 50 trials at each count when GWO-AS is applied to 50-dimensional four benchmark problems.

6.2. Discussion of results

First of all, we discuss obtained function values by four methods, as shown in Tables  and . Now, comparing the results of four methods and the original GWO with ρ0, which are shown in Table , we can see that four methods obtained better solutions than the original GWO with ρ0 for many problems, especially, GWO-SR and GWO-AS did. These results well demonstrate the disadvantages of the original GWO and the advantages of GWO-SR and GWO-AS. Next, concerning the comparison of three methods, PSO, FA and GWO-SR, the number of the problems for which the least objective functions was obtained by three methods are almost the same, which indicates that GWO-SR is not only invariant to shift-transformations by just introducing the reference point, but also has a similar search ability to other two methods. In addition, let us compare four methods, PSO, FA, GWO-SR and GWO-AS. From the above two tables, we can see that GWO-AS has better search ability than other three methods: GWO-AS obtained the first least function value for ten or eleven problems, and the second least function value for five ones in four methods. From these results, we can conclude that even though the GWO has serious drawbacks, the drawbacks can be overcome by applying the proposed methods, which implies that the basic concept for search in GWO is still promising, comparing other metaheuristic methods.

Secondly, we focus on the two proposed methods, GWO-SR and GWO-AS. From preparatory experiments for the methods with different reference points, which are shown in Tables  for GWO-SR, we observed that the appropriate reference points for GWO-SR were μ(1) and μ(1,2), while those for GWO-AS were μ(2) and μ(2,1), respectively. It means that the pbests are effective as the reference point for the search in GWO-SR, while the three best solutions are effective in GWO-AS. The difference can be attributed to the different search characteristics of the two methods. Although GWO-AS using the three best solutions may enhance the intensive search too strongly than GWO-SR, the balance between diversification and intensification is considered to be kept due to the proposed adaptive selection and mutation. Moreover, GWO-AS has better performance than GWO-SR in Tables  and , which indicates that the two proposed methods,the adaptive selection of search areas and the mutation procedure, are effective.

Thirdly, we discuss the relationship between computational complexity and search ability. In Figure , we can observe that all four methods decrease the objective function values sequentially as the search progresses. On the other hand, in the early stage of the search, the GWO-AS decreases more slowly than others, where the slow decline can be explained by the fact GWO-AS requires three times more function evaluations per iteration than the others, as mentioned previous section, while in the middle to end of the search, the GWO-AS decreases the function value significantly faster for many problems. In addition, as discussed above, the final function value obtained by GWO-AS was the least for many problems. These results suggest that the performance of search ability of GWO in terms of computational complexity is higher than others.

Finally, we analyze the mean of selection times of ξ(k,i)(t) per each count in GWO-AS, as shown in Figure , in which we can see that three kinds of candidates ξ(k,i)(t), k{1,2,3} are variously selected as the search progresses: The candidate ξ(k,i)(t) which is most often selected is different in the first and second halves of the search, and its switching count is between 5000 and 8000. In the first half of the search, ξ(1,i)(t) with c1=3 was most often selected, which is relatively small, while in its second half, ξ(3,i)(t) with c3=6 was most often selected. From these results, we can see that three kinds of search areas are adaptively selected in GWO-AS for each problem as the search progresses.

7. Conclusion

In this paper, we have theoretically shown that the original GWO relies on a shift transformation of the problem, and have experimentally shown that its search is too specialized for problems with an optimal solution at the origin, which can significantly reduce its effectiveness of the search for other problems.

Next, we have proposed a modified method of GWO-R (GWO-SR) that is not only invariant to shift-transformations, and that can adjust reasonably its search areas. In addition, we have experimentally shown that its search area can be appropriately varied as the search progresses, in which GWO-SR is more efficient than GWO-R, and as efficient as PSO and FA. Moreover, we have proposed the GWO-AS by adding two methods, an adaptive selection method of the size of search area and a mutation procedure, to GWO-SR. Through numerical experiments, we have verified that the GWO-AS outperforms other methods. From these results, we can conclude that the basic concept of the search inspired by the hunting mechanism of grey wolves is promising despite of its serious drawbacks of the orignal GWO because the drawbacks can be easily overcome by the proposed methods.

Furthermore, since various methods have been investigated for the original GWO [Citation3–8], the proposed GWO can be also modified by some of them. As a future issue, the performance evaluation of the proposed GWOs with such modifications should be verified. In addition, note that the whale optimization algorithm (WOA) [Citation24], which is also one of the popular metaheuristic methods, uses basically the same update systems as the GWO. Thus, the method has the same drawbacks, as mentioned in Ref. [Citation15]. It is possible to add the shift-invariance to WOA and modify its search ability as a general-purpose metaheuristic method by applying the methods used in GWO-SR and GWO-AS. Therefore, we would like to verify how useful those methods are when they are applied to the WOA, or what variations are particularly effective.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Notes on contributors

Keiji Tatsumi

Keiji Tatsumi, received the Ph. D. degree from Kyoto University, Japan, in 2006. In 1998, he joined Osaka University, where he is currently an Associate Professor of the Graduate School of Engineering. His research interests comprise metaheuristics for global optimization and machine learning.

Nao Kinoshita

Nao Kinoshita, received his M.S. degrees from Osaka University, Japan, in 2022. He is currently working for Nippon Steel Corp. His research interests are in metaheuristics for global optimization.

References

  • Mirjalili S, Mirjalili SM, Lewis A. Grey wolf optimizer. Adv Eng Softw. 2014;69:46–61. doi: 10.1016/j.advengsoft.2013.12.007
  • Muro C, Escobedo R, Spector L, et al. Wolf-pack (Canis lupus) hunting strategies emerge from simple rules in computational simulations. Behav Processes. 2011;88:192–197. doi: 10.1016/j.beproc.2011.09.006
  • Mosav SK, Jalalian E, Gharahchopog FS. A comprehensive survey of grey wolf optimizer algorithm and its application. Int J Adv Robot & Expert Syst (JARES). 2018;1(6):23–45.
  • Faris H, Mirjalili S. Grey wolf optimizer: a review of recent variants and applications. Neural Comput Appl. 2018;30:413–435. doi: 10.1007/s00521-017-3272-5
  • Panda M, Dan B. Grey wolf optimizer and its applications: a survey. In: Proceedings of the Third International Conference on Microelectronics, Computing and Communication Systems. 2019.p. 179–194.
  • Ali S, Sharma A, Jadon S. A survey on grey wolf optimizer. J Emerg Technol Innov Res (JETIR). 2020;7(11):789–790.
  • Almufti SM, Ahmad HB, Marqas RB, et al. Grey wolf optimizer: overview, modifications and applications. Int Res J Sci Technol Educ Manage. 2021;1(1):44–56.
  • Wu G, Mallipeddi R, Suganthan P. Ensemble strategies for population-based optimization algorithms – A survey. Swarm Evol Comput. 2019;44:695–711. doi: 10.1016/j.swevo.2018.08.015
  • Mirjalili S. How effective is the grey wolf optimizer in training multi-layer perceptrons. Appl Intell. 2015;43:150–161. doi: 10.1007/s10489-014-0645-7
  • Muangkote N, Sunat K, Chiewchanwattana S. An improved grey wolf optimizer for training q-Gaussian Radial Basis Functional-link nets, 2014 International Computer Science and Engineering Conference (ICSEC); 2014.
  • Saremi S, Mirjalili SZ, Mirjalili SM. Evolutionary population dynamics and grey wolf optimizer. Neural Comput Appl. 2015;26:1257–1263. doi: 10.1007/s00521-014-1806-7
  • Shankar K, Eswaran P. A secure visual secret share (VSS) creation scheme in visual cryptography using elliptic curve cryptography with optimization technique. Aust J Basic Appl Sci. 2015;9(36):150–163.
  • Jian Z, Zhu G. Affine invariance of meta-heuristic algorithms. Inf Sci (Ny). 2021;576:37–53. doi: 10.1016/j.ins.2021.06.062
  • Niu P, Niu S, Liu N, et al. The defect of the grey wolf optimization algorithm and its verification method. Knowl Based Syst. 2019;171:37–43. doi: 10.1016/j.knosys.2019.01.018
  • Askari Q, Younas I, Saeed M. Emphasizing the importance of shift invariance in metaheuristics by using whale optimization algorithm as a test bed. Soft Comput. 2021;25:14209–14225. doi: 10.1007/s00500-021-06101-9
  • Tatsumi K, Kinoshita N. Shift-invariant grey wolf optimizer exploiting reference points and random selection of step-sizes. Proc SICE Ann Conf. 2022;2022:1201–1206.
  • Kennedy J, Eberhart RC. Particle swarm optimization. Proc IEEE Int Jt Conf Neural Netw. 1995;4:1942–1948.
  • Poli R, Kennedy J, Blackwell T. Particle swarm optimization – an overview. Swarm Intell. 2007;1:33–57. doi: 10.1007/s11721-007-0002-0
  • Yang XS. Nature-Inspired metaheuristic algorithms. Frome: Luniver Press; 2008.
  • Fister I, Fister J.I, Yang XS, et al. A comprehensive review of firefly algorithms. Swarm Evol Comput. 2013;13:34–46. doi: 10.1016/j.swevo.2013.06.001
  • Grey Wolf Optimizer (GWO) version 1.6 (1.85 MB) by Mirjalili S. GWO is a novel meta-heuristic algorithm for global optimization. https://www.mathworks.com/matlabcentral/fileexchange/44974-grey-wolf-optimizer-gwo.
  • Awad NH, Ali MZ, Suganthan PN, et al. Problem definitions and evaluation criteria for the CEC 2017 special session and competition on single objective real-parameter numerical optimization, Technical Report of Nanyang Technological Univ., Jordan Univ. and Zhengzhou Univ., 2016.
  • Clerc M. Particle swarm optimization. London: ISTE Publishing; 2006.
  • Mirjalili S, Lewis A. The whale optimization algorithm. Adv Eng Softw. 2016;95:51–67. doi: 10.1016/j.advengsoft.2016.01.008