SUPERNOVA IMPROVEMENT FOR PLANAR LANDSCAPE AND NEEDLE MINIMUMS Eddy Mesa, Juan David Velásquez. Used nomenclature Variable 𝑁 𝐱0 𝐱𝑘 𝐩𝑘 𝐬𝑘 𝑣𝑘 𝐝𝑘 𝐧𝑘𝑡 𝑟𝑖𝑗 𝐮𝑖𝑘 𝐡𝑘 𝐱∗ 𝜆 𝛼 𝑇 𝜂 𝑅 𝑡, 𝑎, 𝑟 Type Integer Vector of 𝑁 × 1 Vector of 𝑁 × 1 Vector of 𝑁 × 1 Vector of 𝑁 × 1 Scalar, 𝑣𝑘 ≥ 0 Unitary vector of 𝑁 × 1 Vector of 𝑁 × 1 Scalar value Unitary vector of 𝑁 × 1 Vector of 𝑁 × 1 Vector of 𝑁 × 1 Scalar Scalar Integer Scalar (0, 1) Integer Integers Meaning Number of dimensions Initial point Current position of particle 𝑘 Previous position of particle 𝑘 Next position of the particle 𝑘 Velocity of the particle 𝑘 Current direction of the particle 𝑘 Best point visited by the particle 𝑘 until the iteration 𝑡 Distance between the particles 𝑖 and 𝑗 Direction between particles 𝑖 and 𝑘 Particle interaction for gravity Local optimum for iteration 𝑡 Radius of the hyper-sphere containing the particles in the initial explosion Factor for reducing the radius of explosion when the algorithm restarts Maximum number of iterations Reduction factor for restarts Number of restarts Algoritm’s counters 1 INTRODUCTION The difficultly of real-world optimization problems has increased over time (there are shorter times to find solutions, more variables, greater accuracy is required, etc.). Classic methods of optimization (based on derivatives, indirect methods) are not enough to solve all optimization problems. So, direct strategies were developed, despite the fact that they can’t guarantee an optimal solution. The first direct methods implemented were systematic searches over solutions spaces. Step by step, they evolved to become more and more effective methods until to arrive to artificial intelligence. Those new methods were called heuristics and metaheuristics, respectively (Himmelblau, 1972). Metaheuristics is iterative and direct methods developed for solving certain types of optimization problems using simple rules for search [3]. It is inspired in physical, natural or biological phenomena [6, 9]. Recently, gravity force had inspired some metaheuristics; Its mean, for a given set of particles with mass, they will be attracted among themselves in proportion to their mass and distance. If we assume the value of the objective function (or its inverse) is the mass, we would have a search process capable to solve optimization problems in a similar way to hill-climbing methods. In literature, we found the following algorithms inspired on gravity: gravitational local search (GLSA) [10], big–bang big–crunch (BBBC) [4], gravitational search (GSA) [14] and central force optimization (CFO) [5]. These four methods differs each other. The first one is a local method while others are global. BBBC use integer variables and calculate inertial mass point,. CFO is deterministic and GSA is random. Following this idea, We proposed a novel metaherustic called Supernova (mesa, 2010). It is a new approach based on the gravity analogy and inspired in the supernova phenomenon [8]. Our methodology shows those advantages: The method is able to find a better optimum in fewer calls to objective function for convex,quasi-convex and even for a smoothness slopes and high rugosity functions. It is robust for high dimensional problems, especially for functions described above. It has a small set of easily tuned parameters. By the way, supernova is a new metaheuristics that is still been developing. In this stage, the methodology is immature and have weakness as: handle of constrains, unsuccessful performance in planar regions and needle minima, control and tune of parameters, high computational costs, among others . We focus in one of them: The performance over planar search regions and needle minima. Although there are many ways to improve the metaheuristics, Our focus are how parameters selection and control could increase the effectiveness over planar regions and initialization and scaled functions as tools to have more and better exploration with metaheuristic. So, The main goal of this work is to find a method for Supernova to have a successful optimization over planar regions with hard minimums (needles). This paper is organized as follows. In Section 2, we present the developed strategy. In Section 3, we show the results for a set of well-known benchmark functions and compare them with other three metaheuristics. In Section 4, we discuss about Supernova results, and the weakness presented. In Section 5 some approach to improve other similar metaheuristics are present. And finally, we conclude and propose future working in Section 6. 2 SUPERNOVA METHODOLOGY This section is focused in how we develop a novel analogy for optimizing nonlinear functions inspired by Supernova phenomenon. When a star exploits, many particles are ejected out from the star. Particles move until they reach a steady state with the environment. This phenomena is very complex and it does not make any sense to try of simulate the entire process with the aim of solving optimization problems [7]. In our approach, we simplify and idealize the real phenomena recreating, only, a simply dynamics to government the particles’ interaction (expel and repel). The particles’ movement is planted as a strategy for exploring the search space. Thus, only two forces are used in our heuristic: the attraction force generated by the gravity; and the impulse force caused by the explosion of the star [12, 15, 17]. In order to exploit the best regions found, the strategy is restarted with a smaller impulse; thus, the algorithm archives a combination of exploration and exploitation necessary to global optimization [12]. In the supernovae phenomenon, the magnitudes are huger compared with typical optimization applications, so, the constants for gravity and impulse are parameters chosen for the problem. In our algorithm, we have 𝐾 particles. In Fig. 1, we analyze the displacement of the particle 𝑘 assuming that other particles remain static. Black points in Fig. 1 represent the trajectory of the analyzed particle, while white points represent the other particles. Fig 1. Forces actuating over a single particle In Fig. 1a, we show the displacements caused by the impulse force and attraction forces with particles 1 and 2. The displacement 𝐢𝑘 of the particle 𝑘 caused by the impulse force is: 𝐢𝑘 = 𝐹 ∙ 𝑚𝑘 ∙ 𝑣𝑘 ∙ 𝐝𝑘 (1) where 𝐹 is a constant parameter for the impulse force; 𝐝𝑘 is the current direction; 𝑚𝑘 is the mass and 𝑣𝑘 is the velocity. The displacement caused by the attraction force between particles 𝑘 and 𝑖, 𝐚𝑘𝑖 , is calculated as (see Fig. 1a): 2 −1 ) ∙ 𝐮𝑘𝑖 𝐚𝑘𝑖 = 𝐺 ∙ 𝑚𝑘 ∙ 𝑚𝑖 ∙ (𝑟𝑘𝑖 (2) where 𝐺 denotes the gravitational constant, 𝑟𝑘𝑖 denotes the distance between particles 𝑘 and 𝑖, and 𝐮𝑘𝑖 denotes a unitary vector in the same direction of particles 𝑘 and 𝑖. Thus, the net displacement 𝚫𝑘 is calculated taking into count the contribution of the displacements caused by the attraction forces among all particles and the inertia: 𝐾 𝐾 𝚫𝑘 = 𝐢𝑘 + ∑ 𝐚𝑘𝑖 = 𝐹 ∙ 𝑚𝑘 ∙ 𝑣𝑘 ∙ 𝐝𝑘 + ∑ 𝑖=1 𝑖≠𝑘 𝑖=1 𝑖≠𝑘 𝐺 ∙ 𝑚𝑘 ∙ 𝑚𝑖 ∙ 𝐮𝑘𝑖 2 𝑟𝑘𝑖 (3) Thus, the next position of the particle is calculated as the current position plus the new displacement, as show in Fig. 1b. 2.1 Algorithm description The developed methodology is presented in the Algorithm 1. The nonlinear optimization problem is stated as 𝐱 𝐺 = arg min 𝑓(𝐱), where 𝐱 is a vector of 𝑁 × 1 coordinates or components, and 𝑓(∙) is a bounded real-valued, continuous and multimodal function, such that 𝑓: ℛ 𝑁 → ℛ; and 𝐱 𝐺 is a vector representing the global optimum, such that 𝑓(𝐱 𝐺 ) ≤ 𝑓(𝐱) in a compact domain. We have 𝐾 particles. The particle 𝑘 in the iteration 𝑡 has the following properties: a current position, 𝐱 𝑘 ; a previous position, 𝐩𝑘 ; a next position, 𝐬𝑘 ; a mass,𝑚𝑘 ; a unitary direction, 𝐝𝑘 ; and a velocity, 𝑣𝑘 . In addition, 𝐧𝑘𝑡 denotes the best point visited by the particle 𝑘 until the iteration 𝑡. Supernova iterates over 𝑡 = 1, … , 𝑇. Initialization. Initial velocity, 𝑣𝑘 , and direction, 𝐝𝑘 , of each particle 𝑘 are random. For this, we use the function runif(𝑛) (lines 3 and 4) which returns is a random number between zero and one for velocity and a random a vector of 𝑛 components for 𝐝𝑘 .. For the initial iteration (𝑡 = 1), an initial point 𝐱 0 is required for generating the set of expelled particles; the position of each particle 𝐬𝑘 is obtained in line 9; 𝑤ℎ𝑒𝑟𝑒 𝜆 is a scalar value representing the radius of an hyper-sphere that contained the set of expelled particles. Thus, the star is located in 𝐱 0 when it explodes.a Function evaluation. In our methodology, the value of objective function evaluated in 𝐱 𝑘 is proportional to the mass 𝑚𝑘 of the particle in eq. (1)–(3) with 𝑚𝑘 > 0. Thus, for function minimization, it is necessary to use 𝑚𝑘−1 as the mass of particle 𝑘, where: 𝑚𝑘 = 𝑓(𝐱 𝑘 ) + 1 − min 𝑓(𝐱 𝑘 ) (4) 1≤𝑘≤𝐾 Previous transformation guarantee positive values of 𝑚𝑘 . In line 18, we use 𝑓 ∗ to denote the minimum of 𝑓(∙) which is calculated in the line 12. Calculation of displacement. In this step, we calculate the total displacement for each particle using eq. (3). In line 20, 𝐝𝑘 is the unitary vector in direction from the previous position to the current position of the particle 𝑘. In line 21, the unitary vector in the direction of points 𝑖 and 𝑘, 𝐮𝑖𝑘 , is calculated. In line 22, we calculate the net displacement caused for all attraction forces (see eq. (2)); however, the accurated use of eq. (2) is not possible due to the magnitude of the values used for measuring masses and distances in the real phenomenon are not comparable with 2 −1 2 −1 ) (𝑟𝑘𝑖 ) in eq. (2) for a new expression 𝐚𝑘𝑖 in eq. (5) to guarantee a optimization problems. Thus, we replace (𝑟𝑖𝑘 bigger influence for near particles and lower influence for high distances between particles as in the gravity fields, without the problem of division by zero for close particles: 𝐾 −1 𝐚𝑘𝑖 = 𝐺 ∙ 𝑚𝑘 ∙ 𝑚𝑖 ∙ [1 − 𝑟𝑖𝑘 ( ∑ 𝑟𝑖𝑗 ) ] ∙ 𝐮𝑘𝑖 𝑗=1|𝑗≠𝑖 (5) Algorithm 1. Supernova. 01: 02: 03: 𝑡 ← 1; 𝑟 ← 0; 𝜆 ← 𝜆0 loop while 𝑟 ≤ 𝑅 if 𝑡 ≠ 1 then, 𝐱 0 ← arg min 𝑓(𝐧𝑘 ), end if 1≤𝑘≤𝐾 04: 05: 06: 07: 08: 09: 10: 11: 12: for 𝑘 =1 to 𝐾 do 𝑣𝑘 ← runif(1) 𝐝𝑘 ← 2 ∙ runif(𝑁) − 1 𝐝𝑘 ← 𝐝𝑘 /‖𝐝𝑘 ‖ 𝐱𝑘 ← 𝐱0 𝐬𝑘 ← 𝐱 𝑘 + 𝜆 ∙ 𝑣𝑘 ∙ 𝐝𝑘 if 𝑡 = 1 then, 𝐧𝑘 ← arg min{𝑓(𝐱 0 ), 𝑓(𝐬𝑘 )}; end if end for 𝑓 ∗ ← min 𝑓(𝐬𝑘 ) 13: 14: 15: 16: 17: 18: 19: 20: 21: 22: 𝑎←0 loop while 𝑡 ≤ 𝑇 for 𝑘 = 1 to 𝐾 do: 𝐩𝑘 ← 𝐱 𝑘 ; 𝐱 𝑘 ← 𝐬𝑘 ; end for for 𝑖 = 1 to 𝐾 do: for 𝑗 = 1 to 𝐾 do: 𝑟𝑖𝑗 ← ‖𝐱 𝑖 − 𝐱𝑗 ‖; end for; end for for 𝑘 = 1 to 𝐾 do: 𝑚𝑘 ← 𝑓(𝐱 𝑘 ) + 1 − 𝑓 ∗ 𝐧𝑘 ← arg min {𝑓(𝐧𝑘 ), 𝑓(𝐩𝑘 ), 𝑓(𝐱 𝑘 )} 𝐝𝑘 ← [𝐩𝑘 − 𝐱 𝑘 ]⁄‖𝐩𝑘 − 𝐱 𝑘 ‖ for 𝑖 = 1 to 𝐾 do: 𝐮𝑖𝑘 ← [𝐱 𝑖 − 𝐱 𝑘 ]⁄‖𝐱 𝑖 − 𝐱 𝑘 ‖; end for −1 −1 𝐾 −1 𝐡 𝑘 ← ∑𝐾 𝑖=1|𝑖≠𝑘 𝐺 ∙ 𝑚𝑘 ∙ 𝑚𝑖 ∙ [1 − 𝑟𝑖𝑘 ∙ (∑𝑗=1|𝑗≠𝑖 𝑟𝑗𝑘 ) ] ∙ 𝐮𝑖𝑘 23: 24: 25: 26: 27: 28: 29: 30: 31: 32: 33: 34: 35: 36: 37: 38: 39: 40: 1≤𝑘≤𝐾 𝐬𝑘 ← 𝐱 𝑘 + m𝑘−1 𝑣𝑘 ∙ 𝐹 ∙ 𝐝𝑘 + 𝐡𝑘 end for 𝐱 ∗ ← arg min 𝑓(𝐬𝑘 ); 𝑓 ∗ ← 𝑓(𝐱 ∗ ) 1≤𝑘≤𝐾 if 𝑡 = 1 or 𝑓 ∗ ≤ 𝑓(𝐱 𝐺 ) then 𝐱𝐺 ← 𝐱∗ 𝑎←0 else 𝑎 ← 𝑎+1 end if if 𝑎 > 𝜂 ∙ 𝑇 then 𝑡 ←𝑇+1 end if 𝑡 ←𝑡+1 end loop 𝑡 ←1,𝑟 ←𝑟+1 𝜆 ← 𝜆 ⁄𝛼 end loop end algorithm Algorithm restart. In heuristic methods, it is well-known that the initial exploration usually is insufficient for certain multimodal regions because there are no enough focus to find the best point. When algorithm restarts, a search in a smaller region around the best point found begins, improving the accuracy. When the algorithm reaches 𝜂 ∙ 𝑇 iterations without improving the current optimum (see lines 28 and 30), the algorithm is restarted (line 33), and a new set of particles is created. For this, the radius of the hyper-sphere containing the initial particles is reduced (line 38); following, the best point found until the current iteration (lines10 and 19) is used as the initial point (line 03), and a new explosion occurs (lines 4 to 10). Stop criterion. The process finishes when the maximum number of iterations 𝑇 or restarts 𝑅 are reached (lines 2 and 14) [11]. 3 TESTING After implementation, we started to test supernova. Test had three parts. Initially, parameters were evaluated. In the second part, some functions were rotated and move to obtain different problem’s configuration to check the robustness. Finally, Supernova was validated against another metaheuristics. This section is divided in four parts: Recommended values, performance for specific functions configurations, description of benchmark functions and results. 3.1 Recommended values for the parameters The first step of testing was to search adequate general values for the parameters of the proposed algorithm. For this, we optimize the set of benchmark functions proposed by De Jong [2] varying each parameter. The obtained results were: Number of iterations (𝑇): 100. Number of restarts (R): 50. Number of particles (𝐾): 60–100 Momentum parameter (𝐹): 2 Attraction parameter (G): 2 Initial radius of the hyper-sphere for generating the particles (𝜆0 ): 200 Factor for reducing the radius of the hyper-sphere in each restart (𝛼): 1.5 Complete results are reported in [8]. Those parameters are used for the complete test of Supernova, except 𝜆0 for small search spaces, less 30 units, or huge spaces, over 200 units. For small regions, we use 𝜆0 = 30; for big regions, we use 𝜆0 = 600. 3.2 Supernova performances over special regions 3.3 Experimental setup Supernova has been compared with Differential Evolution (DE) [1], Adaptive Differential Evolution (ADE) [1] and Fast Evolutionary Programming (FEP) [16] using 23 benchmark functions taken from [1] and [16]. In Table 1, the equation of each function, the range of the search space, the number of dimensions (𝑁) and the optimal value for each function are presented. Table 2. Benchmark functions 𝑓 Equation 𝑓(𝐱 𝑜𝑝𝑡 ) 2 𝑓1 = ∑𝑁 𝑖=1 𝑥𝑖 0.0 𝑁 𝑓2 = ∑𝑁 𝑖=1|𝑥𝑖 | + ∏𝑖=1|𝑥𝑖 | 0.0 2 𝑖 𝑓3 = ∑𝑁 𝑖=1(∑𝑗=1 𝑥𝑗 ) 0.0 𝑓4 = max{|𝑥𝑖 |, 1 ≤ 𝑖 ≤ 𝑁} 0.0 2 2 2 𝑓5 = ∑𝑁−1 𝑖=1 [100(𝑥𝑖+1 − 𝑥𝑖 ) + (𝑥𝑖 − 1) ] 0.0 2 𝑓6 = ∑𝑁 𝑖=1(⌊𝑥𝑖 + 0.5⌋) 0.0 4 𝑓7 = ∑𝑁 𝑖=1 𝑖𝑥𝑖 + 𝑟𝑎𝑛𝑑𝑜𝑚[0,1) 0.0 𝑓8 = − ∑𝑁 𝑖=1(𝑥𝑖 sin(√|𝑥𝑖 |)) −12569.5 2 𝑓9 = ∑𝑁−1 𝑖=1 [𝑥𝑖 − 10 cos(2𝜋𝑥𝑖 + 10)] 0.0 𝑓10 1 1 𝑁 𝑁 𝑁 2 1 = −20 exp (−0.2√ ∑𝑁 𝑖=30 𝑥𝑖 ) − exp ( ∑𝑖=1 cos(2𝜋𝑥1 )) + 20 + e Range 𝑁 [−500, 500]𝑁 30 [−10, 10]𝑁 30 [−100, 100]𝑁 30 [−100, 100]𝑁 30 [−30, 30]𝑁 30 [−100, 100]𝑁 30 [−1.28, 1.28]𝑁 30 [−500, 500]𝑁 30 [−5.12,5.12]𝑁 30 [−32,32]𝑁 30 [−600, 600]𝑁 30 0.0 𝑓11 𝑓12 𝑓13 = ∑𝑁 𝑖=1 𝑥𝑖2 4000 𝑥 𝑖 − ∏𝑁 𝑖=1 (cos ) + 1 √𝑖 0.0 π 2 [−50, 50]𝑁 = {10 sin2 (𝜋𝑦1 ) + ∑𝑁−1 𝑖=1 (𝑦𝑖 − 1) ∙ 𝑁 0.0 [1 + 10 sin2 (𝜋𝑦𝑖+1 + 1 )] + (𝑦𝑛 − 1)2 } + ∑30 𝑖=1 𝑢(𝑥𝑖 , 10,100,4) 2 2 [−50, 50]𝑁 = 0.1{sin2 (𝜋3𝑥1 ) + ∑𝑁−1 𝑖=1 (𝑥𝑖 − 1) [1 + sin (3𝜋𝑥𝑖+1 )] 0.0 +(𝑥𝑛 − 1)2 [1 + sin2 2𝜋𝑥𝑁 ]} + ∑𝑁 𝑖=1 𝑢(𝑥𝑖 , 5,100,4) 𝑘(𝑥𝑖 − 𝑎)𝑚 , 𝑥𝑖 > 𝑎 1 0, −𝑎 ≤ 𝑥𝑖 ≤ 𝑎 and 𝑦𝑖 = 1 + (𝑥𝑖 + 1) 𝑢(𝑥𝑖 , 𝑎, 𝑘, 𝑚) = { 4 𝑚 𝑘(−𝑥𝑖 − 𝑎) , 𝑥𝑖 < −𝑎 30 30 −1 𝑓14 =[ 1 500 + ∑25 𝑗=1 1 6 𝑗+∑2 𝑖=1(𝑥𝑖 −𝑎𝑖𝑗 ) ] [−65.536, 65.536]𝑁 2 [−5, 5]𝑁 4 [−5, 5]𝑁 2 [−5,10][0,15] 2 0.0 𝑥1 (𝑏𝑖2 +𝑏𝑖 𝑥2 ) 2 𝑓15 = ∑11 𝑖=1 [a i − 𝑓16 3.075 × 104 1 = 4x12 − 2.1𝑥14 + 𝑥16 + 𝑥1 𝑥2 − 4𝑥22 + 4𝑥24 3 −1.031628 𝑓17 = (x2 − 0.398 𝑏𝑖2 +𝑏𝑖 𝑥3 +𝑥4 5.1 2 𝑥 4𝜋2 1 5 ] 2 + 𝑥1 − 6) + 10 (1 − 𝜋 1 8𝜋 ) cos 𝑥1 + 10 𝑓19 [−2, 2]𝑁 = [1 + (𝑥1 + 𝑥2 + 1)2 (19 − 14𝑥1 + 3𝑥12 − 14𝑥2 +6𝑥1 𝑥2 + 3𝑥22 )] 3 × [30 + (2𝑥1 − 3𝑥2 )2 × (18 − 32𝑥1 + 12𝑥12 + 48𝑥2 − 36𝑥1 𝑥2 + 27𝑥22 )] 2 [0, 1]𝑁 = − ∑4𝑖=1 𝑐𝑖 exp [− ∑𝑁 𝑗=1 𝑎𝑖𝑗 (𝑥𝑗 − 𝑝𝑖𝑗 ) ] 𝑓20 −3.86 2 = − ∑4𝑖=1 𝑐𝑖 exp [− ∑𝑁 𝑗=1 𝑎𝑖𝑗 (𝑥𝑗 − 𝑝𝑖𝑗 ) ] 𝑓18 𝑓21 𝑓22 𝑓23 −3.32 = ∑5𝑖=1[(x − a i )(𝑥 − 𝑎𝑖 )𝑇 + 𝑐𝑖 ]−1 0.0 = ∑7𝑖=1[(x − a i )(𝑥 − 𝑎𝑖 )𝑇 + 𝑐𝑖 ]−1 0.0 𝑇 −1 = ∑10 𝑖=1[(x − a i )(𝑥 − 𝑎𝑖 ) + 𝑐𝑖 ] 0.0 2 3 [0, 1]𝑁 6 [0, 10]𝑁 4 [0, 10]𝑁 4 [0, 10]𝑁 4 The characteristics of test set are important to know the metaheuristic. How it works agree to dimensionality, convexity, modality and if the function are separable or not. In table 2 are describe the characteristics o this functions. It is important to note in the Table, all characteristics are defined for the range of the function and we declare quasi-convex functions as convex because the global behavior. Table 2. Functions characteristics Función Número de dimensiones 𝑓1 30 𝑓2 30 𝑓3 30 𝑓4 30 𝑓5 30 𝑓6 30 𝑓7 30 𝑓8 30 𝑓9 30 𝑓10 30 𝑓11 30 𝑓12 30 𝑓13 30 𝑓14 2 𝑓15 4 𝑓16 2 𝑓17 2 𝑓18 2 𝑓19 3 𝑓20 6 𝑓21 4 𝑓22 4 Convexty diferenciable modalidad Separabilidad convex convex diferenfiable No diferenciable convexa No diferenciable diferenciable No diferenciable No convexa No diferenciable unimodal unimodal unimodal unimodal multimodal unimodal unimodal Multimodal multimodal multimodal multimodal multimodal multimodal Multimodal multimodal multimodal multimodal multimodal multimodal multimodal multimodal multimodal separable Separable No separable separable No separable separable Separable Separable separable No separable No separable No separable No separable convexa convexa convexa diferenciable 𝑓23 4 multimodal ̅̅̅∗ ), the standard deviation (𝜎) For each benchmark function, we run Supernova 50 times and calculate the average (𝑓 of the best function value found in each run and the number of function calls (FO). The obtained results are reported in Table 3. [coello,06, DE],[a note on the extended rosenbrock function][CEC’05] 3.3.1 Results For the 23 benchmark functions, the algorithm improve the answers (see Table 1) and have less calls to objective function in 26% of the functions. Presented less calls to objective function in 22% of the functions for similar minimums; Supernova presents a similar behavior in 30% of the test functions. The best behavior of Supernova is observed for functions with a kind of descend geometry. For plane regions, Supernova shows the worst performance. In [10], the GSA method is validated with the same test of functions but with different parameter values, such the number of calls to the function objective is not similar and uses a smaller population to search too. Therefore, the results presented in [10] are not comparable because the test conditions are not similar. Table 3. Results and comparison for benchmark functions. f Supernova ____________________________ f∗ σ 6,3 x 10 —29 1,0 x 10 —28 1,1 x 10 —26 F.O f1 972 f2 1812 1,1 x 10 —26 f3 852 1,0 x 10 —13 8,8 x 10 —14 f4 1212 6,1 x 10 —35 1,7 x 10 —34 2,5 x 10 —2 f5 510 28.89 f6 677 0 0 —3 —3 f7 f8 552 612 3,2 x 10 —7,4 x 10 15 f9 612 f10 972 f11 612 f12 100 f13 250 1.09 f14 100 1,9 x 10 4,28 x 10 0 4,7 x 10 —15 0 2,3 x 10 —15 0 0 1,1 x 10 1,0 x 10 5,7 x 10 1,3 x 10 2000 1,0 x 10 —24 9,7 x 10 —24 5000 3,1 x 10 —14 5,9 x 10 —14 1500 8,2 x 10 5,9 x 10 —9 9,9 x 10 —10 8,1 x 10—3 7,7 x 10 —4 —11 7,4 x 10 —11 1,6 x 10 —2 1,4 x 10 —2 1,5 x 10 6,8 x 10 5000 0 0 0 0 0.3 0.5 20000 0 0 0 0 5.06 5.87 1500 0 0 0 0 0 0 —3 —4 —3 —3 0.1415 0.3522 -12554.5 52.6 3000 9000 3,15 x 10 -12569.5 5000 1500 7,5 x 10 7,0 x 10 —12 0 7,7 x 10 —15 2000 0 1,4 x 10 —15 0 4,63 x 10 -11080.1 69.2 9,7 x 10 0 —8 1,2 x 10 574.7 38.8 4,2 x 10 0 —8 0 4,6 x 10 —2 1,2 x 10 —2 1,8 x 10 —2 2,1 x 10 —3 1,6 x 10 —2 2,2 x 10 —2 —15 8,0 x 10 —15 9,2 x 10 —6 3,6 x 10 —6 0.55 1500 5,0 x 10 —29 3,9 x 10 —29 5,1 x 10 —14 4,8 x 10 —14 1,6 x 10 —4 7,3 x 10 —5 1.39 1.2 100 2,6 x 10 —16 3,3 x 10 —16 —4 4,9 x 10 —5 4,4 x 10 -1.0315 1,36 x 10 —4 f17 204 0.39884 1,2 x 10 —3 f18 204 3.0024 2,9 x 10 —3 f19 50 -3.862 7,4 x 10 —5 f20 60 -3.26 3,68 x 10 —2 f21 510 2,17 x 10 —1 2,5 x 10 —1 3,5 x 10 —1 4 σ —4 7,9 x 10 204 510 f* —4 —30 f16 f23 σ —14 7,9 x 10 404 510 f∗ —14 —30 f15 f22 σ —28 6,6 x 10 2,5 x 10 —2 f∗ —28 F.O 1500 2,9 x 10 —2 15 ADE DE FEP _____________________________ ______________________ ___________________ -8,73 x 10 —1 —8,8 x 10 —1 -1.08 4000 0.998004 4,4 x 10 —4 2,7 x 10 9,7 x 10 —4 —12 0.998004 4,5 x 10 —4 3,1 x 10 —13 1.22 0.56 3,2 x 10 —4 -1.03 4,9 x 10 —7 0.398 1,5 x 10 —7 5,0 x 10 —4 100 -1.03163 100 0.397887 100 3 100 N. A. N. A. N. A. N. A. -3.86 1,4 x 10 —5 200 N. A. N. A. N. A. N. A. -3.27 5,9 x 10 —2 100 -10.1532 100 100 -10.4029 -10.5364 2,3 x 10 1,7 x 10 —8 -1.03163 3,3 x 10 —4 0.397887 —15 2,2 x 10 —6 4,9 x 10 —7 5,8 x 10 —6 3 9,9 x 10 2,0 x 10 —9 —15 2,5 x 10 —6 -10.4029 3,9 x 10 —7 -10.5364 1,9 x 10—7 -10.1532 DISCUSSION Agree with the descriptions given, for the functions, on section 4.3. We can conclude that: Planar regions and penalties (f5,f12,f13,f16,f17,18) Needle minimums (f5,f14,f21,f23,f24) Are problematic topologies for Supenova. 3.02 0.11 -5.52 1.59 -5.52 2.12 -6.57 3.14 5. HOW HAD BEEN OTHER METAHEURISTICS IMPROVED? Generally, there is not a specific landscape associated to specific improve over performance. Regularly, an improved is present numerically for a set of functions. Follow the functions with low performance, we found a improvement of the performance for rosenbrock function (f5) with particle swarm optimization (PSO). And have a second idea to improve the performance having better control of the parameters. Both ideas will be exposed below in this section. 5.1. PSO with LDS 5.2. Parameters Stochastic methods have not the same pathway each run of algorithms. Also, the intelligence immerse in metaheuristics change dynamically from the characteristics found in the search. Parameters give to metaheuristics robustness and flexibility (Talbi, 2009). To choose parameters is not a simple task because a wrong set of parameters would make that metaheuristics have premature convergence or even no convergence at all (Bäck, 1996; Talbi, 2009). Metaheuristics need to have enough parameters to be flexible, but each parameter increase the complexity to the tune and control of the method and the necessary changes to optimize a new problem. Each metaheuristic is a complete different world, has different parameters, and they influence the metaheuristic in different ways. There is not a unique right to choice parameters in metaheuristics. In literature two points of view are related, but they are only for Evolutionary Computing (EC). Traditionally, EC includes genetic algorithms (GA), evolution strategies (ES), and evolutionary programming (EP). By the way, other metaheuristics met the conditions propose by Angeline (1995) to be an EC metaheuristic, method that iterated until specific stopping criteria. So, we would extend this classification to another metaheuristics and heuristics in general, but they do not use the same operators that EC ( Fogel, D. B., et al, 1991). Specifically, most part of another heuristics has not operators formalized as operators, they use a complete rules with parameters inside, so maybe it is not a direct analogy (Zapfel, et al., 2010). In the next two sections the formal parameter classification proposed in literature was discussed and follow an example of other heuristic adaptation are presented. 5.2.1. PARAMETERS CLASSIFICATION Each metaheuristic has a parameter’s set to control the search. Although there is no consensus about a unique classification to metaheuristic parameters, one approximation had been developed for EC (Angeline, 1995; Hinterding,et al., 1997; Kramer, 2008). According with the definition provided by Angeline (1995) this classification could be extended to others paradigms. The most complete classification is proposed by Kramer (2008) who links Angeline (1995) point of view and Hinterding et al. (1997) 5.2.1.1 Exogenous. These parameters are those whose are affected by metaheuristic performance, but they are external to it, for example: constrain changes, problems with parts functions. 5.2.1.2 Endogenous These are internal to the method and could be change by the user or method itself. Even though, Endogenous parameters are our focus, we cannot forget the exogenous ones because they will affect their selection Population level: In this level the parameters are global for the optimization. Operator of EC that use this type or parameters control the next generation. For example: population size, stopping criteria, etc. Individual level .This kind of parameters only affects each individual, for example: It could be the pass for each individual. Component level. In this level, the parameters affect part of the individual like a gene of a chromosome in Generic Algorithms (GA). Also, it is important to notice that, authors propose this classification, but they do not talk about a right or unique manner to adapt and automatize each level of parameters. The techniques are addressed by next classification. 5.2.2. PARAMETERS SETTINGS In last section, parameters were changed according to their level of sort in the metaheuristic. In this case, the parameters are chosen in two stages the first is before to use metaheuristic, and it is called “to tune” and the second one is called “to control” and both have different ways to be select. 5.2.2.1. Tuning. Parameters are tuned before to use metaheuristic. Those initial parameters could be chosen by three different levels accord Eiben, et al (1999). Manually: The initial parameters could be chosen by and expert ad hoc it is a right manner but lately are not the most recommended (Kramer, 2008; Talbi, 2009) Design of experiments (DOE): It implies design test to show the behavior, and use or define a metric analyze the result and take a decision about a value or a range of values for each parameter. (Kramer, 2008; Talbi, 2009). Metaevolutionary: It means that parameters are chosen by other metaheuristics, in literature a good example is given by bacterial chemotaxis method proposed in (Müller, et al., 2000). Additionally, Kramer (2008) extend this procedure to control like in Hyper-metaheuristics (Hamadi, et al., 2011) 5.2.2.2. Control Parameters can change while metaheuristic is running. Control process is this change in “real time”. Accord to Hinterding et al. (1997) and Eiden et al (1999), Kramer (2008) describes and discusses about three methods to control parameters. Deterministic: It could be static or dynamics. Statics means there is no change at all and dynamic it change with a specific rule like a dynamic penalty than change with the distance to feasible zone (Kramer, 2008; Mesa, 2010) Adaptive: In this case, parameter changes agree a rule like if it happens, then do this. A classic example of adaptive parameters is the 1/5th rule for mutation use in ES (Bäck, 1996; Kramer, 2008). Self-adaptive: In this type of control, parameters evolves agree to the problem for example the self-adaptation of the mutation step-size with the direction to mutation (Kramer, 2008; Schwefel, 1995) REFERENCES [1] J. Brest, S. Greiner, B. Boskovic, M. Mernik, V. Zumer, Adapting Control Parameters in Differential Evolution: A comparative study on numerical benchmark problems, IEEE T. Evolut. Comput. 10 (2006) 646– 657. [2] K.D. De Jong, An analysis of the behavior of a class of genetic adaptive systems, Ph.D. Thesis, University of Michigan, Michigan, 1975. [3] M. Dorigo, M. Birattari, T. Stützle, Ant colony optimization: Artificial ants as a computational intelligence technique, IEEE Comput. Intell. M. 1 (2006) 28–39. [4] O.K. Erol, I. Eksin, A new optimization method: Big Bang -Big Crunch, Adv. Eng. Soft. 37 (2006) 106–111. [5] R.A. Formato, Central force optimization: a new metaheuristic with applications in applied electromagnetics, Prog. Electromagn. Res. 77 (2007) 425–491. [6] M. Gendreau, J.-Y. Potvin, Handbook of Metaheuristics, Int. Ser. Oper. Res. Man. 146, second ed., Springer, New York, 2010. [7] A. Lieddle, Introduction to modern cosmology, Wiley, West Sussex, 2003. [8] E. Mesa. Supernova: un algoritmo novedoso de optimización global. Master Thesis. Universidad Nacional de Colombia–Sede Medellín. 2010. (available at http://www.bdigital.unal.edu.co/2035/) [9] S. Olariu, A.Y. Zomaya. Handbook of bioinspired algorithms and applications. Computer & Information Science Series.Boca Ratón, Chapman & Hall/CRC, 2006. [10] E. Rashied, H. Nezamabadi-pour, S. Saryazdi. GSA: A Gravitational Search Algorithm. Inform. Sci. 179 (2009) 2232–2248. [11] E. Talbi, Metaheuristics: From design to implementation, first ed., Wiley, Hoboken, 2009. [12] A. Törn, A. Zilinskas. Global optimization, Lect. Notes Comp. Sci. 350, Springer, New York, 1989. [13] J. Vesterstrom, R. Thomsen. A comparative study of differential evolution, particle swarm optimization, and evolutionary algorithms on numerical benchmark problems. Proceedings of congress of evolutionary computation, Portland, (2004) 1980–1987. [14] B. Webster, P. Bernhard. A local search optimization algorithm based on natural principles of gravitation. Technical Report, Florida Institute of technology, Melbourne, 2003. [15] D.H. Wolpert, W.G. Macready. No free lunch theorems for optimization. IEEE T. Evolut. Comput. 1 (1996) 67–82. [16] X. Yao, Y. Liu, G. Lin. Evolutionary programming made faster. IEEE T. Evolut. Comput. 3 (1996) 82–102. [17] S.H. ZanaLa descripción de supernova REFERENCES Ali, M., and Törn, A. (2004). Population set-based global optimization algorithms: Some modifications and numerical studies. Comput. Operational Research, 31(10), pp. 1703–1725. Angeline, P. (1995). Adaptative and Self-Adaptative Evolutionary Computations. IEEE, Computational Intelligence (pp. 152-163). IEEE. Brest, J., Greiner, S., Boskovic, B., Mernik, M., and Zumer, V. (2006). Self-Adapting Control Parameters in Differential Evolution: A Comparative Study on Numerical Benchmark Problems. IEEE Transactions on Evolutionary Computation, 10(6), 646-657. Bäck, T. (1996). Evolutionary Algorithms in Theory and Practice. Oxford University press. Eiben, E., Hinterding, R., and Michalewicz, Z. (1999). Parameter control in evolutionary algorithms. IEEE Transactions on Evolutionary Computation, 3(2), 124-141. Fogel, D. B., Fogel, L. J., and Atmar, J. W. (1991). Meta-evolutionary programming. Signals, Systems and Computers, 1991. 1991 Conference Record of the Twenty-Fifth Asilomar Conference on pp. 540–545. IEEE. Hamadi, Y., Monfroy, E., and Saubion, F. (2011). What Is Autonomous Search? In: Hybrid optimization.Ed M. Milano and P. Van Hentenryck . pp. 357-392. Springer. Himmelblau, D. (1972). Applied Nonlinear Programming (p. 498). McGraw hill. Hinterding, R., Michalewicz, Z., and Eiben, A. E. (1997). Adaptation in evolutionary computation: a survey. Proceedings of 1997 IEEE International Conference on Evolutionary Computation (ICEC ’97) (pp. 65-69). IEEE. Kramer, O. (2008). Self-adaptative Heuristics for Evolutionary Computation. Design (First Ed., p. 175). Berlin: springer. Mesa, E. (2010). Supernova : un algoritmo novedoso de optimización global. National University. Retrieved from http://www.bdigital.unal.edu.co/2035/. Mezura-Montes, E., Velázquez-Reyes, J., and Coello Coello, C. A. (2006). A comparative study of differential evolution variants for global optimization. Proceedings of the 8th annual conference on Genetic and evolutionary computation (p. 485–492). ACM. Müller, S., Airaghi, S., Marchetto, J., and Koumoutsakos, P. (2000). Optimization algorithms based on a model of bacterial chemotaxis. in Proc. 6th Int. Conf. Simulation of Adaptive Behavior: From Animals to Animats,, SAB 2000 Proc. Suppl (pp. 375-384). Citeseer. Price, K. V. (1996). Differential evolution: a fast and simple numerical optimizer. Proceedings of North American Fuzzy Information Processing (pp. 524-527). Ieee. doi: 10.1109/NAFIPS.1996.534790. Schwefel, H. P. (1995). Evolution and Optimun Seeking (p. 435). N.Y. Wiley. Talbi, E. (2009). Metaheruistics. Search (p. 618). Wiley. Yang, X.-S. (2010). Engineering Optimization An Introduction with Metaheuristics Applications. Engineering Optimization. Hoboken: Wiley. Zanakis, S. H., and Evans, J. R. (1981). Heuristic “Optimization”: Why, When, and How to Use It. Interfaces, 11(5), 84-91. Zapfel, G., Braune, R., and Bogl, M. (2010). Metaheuristic Search Concepts. Search (1st ed., p. 316). Berlin: Springer. Zilinskas, A., and Törn, Aimo. (1989). Global Optimization. In G. Goos and J. Hartmanis (Eds.), Lecture Notes in Computer Science vol 350 (p. 255). Berlin: Springer.