• Nem Talált Eredményt

Validating Quality of Multi-objective Search and Optimisation

In document Óbuda University PhD Dissertation (Pldal 31-40)

2.1 Literature Synopsis

2.1.6 Validating Quality of Multi-objective Search and Optimisation

Various stochastic search and optimisation methods based on stochastic gradient descent variations like simulated annealing, taboo search; swarm intelligence methods like ant colonies, bacterial search; and evolutionary algorithms like genetic algorithms have been successfully used for fuzzy system optimisation [32], [16], [2], [9]. Genetic algorithms are efficient, well studied simple stochastic optimisation methods of proven convergence, capable of global multi-objective search and optimisation of versatile complex systems. Capabilities and performance characteristics of properly formed GAs are defined in [27] so the effect on the GA performance becomes obvious if any supplementary operator – like a new vector comparison or a new niching operator is included into a simple GA.

This paper will prove the validity of a new general vector comparison operator, suitable for applications in any mathematical or engineering field, including stochastic search, and evolutionary algorithms. For a relation to be a proper strict inequality relation it must have the properties of irreflexivity, antisymmetry and transitivity. I shall

mathematically prove these traits by analysing basic set cardinality properties of my proposed vector comparison operators. These new vector comparison methods provide more information when comparing two vectors than the classic Pareto-based comparison, thus the GA is faster, more efficient in its search.

Many benchmark problems were designed to test the effectiveness of a multi-objective GA. Of the most interest to us, thus what I will use is the analytically designed GA hard multi-objective function set presented in [79]; I have also used their generalisation to 4, 8 and 16 objectives. To have a clear, familiar simple baseline I have also implemented a single objective GA, which minimises the sum of 42, 84, 168 and 336 parameters – since the 2, 4, 8 and 16 multi-objective problems use this many parameters in total.

2.1.6.1 Simple Two Objective Optimisation Problem

According to [79] a simple two objective optimisation problem can be defined as:

Minimize f1(x) = f1(x1, x2,…, xm),

Minimize f2(x) = g(xm+1,…, xN).h(f1(x1,…, xm),g(xm+1,…, xN)). (15)

Figure 4.

Four hyperbolic lines (f1f2 = c) with c1 < c2 < c3 < c4 are shown.

The most simple case is where f1 = x1 and g(x2) (>0) is a function of x2 only, and h = g/f1. The first objective function f1 is a function of x1 only and the function f2 is a function of both x1 and x2. In the function space (that is, a space with (f1; f2) values), the above two functions obey the following relationship: f1(x1; x2) . f2(x1; x2) = g(x2).

For a fixed value of g(x2) = c, a f1 – f2 plot becomes a hyperbola (f1f2 = c) as shown for a

objective f1, along a Pareto-front like f1fi = ci, for i=2,3,4,5. Results of the optimisation that correspond to the smallest achievable c are the global, Pareto-optimal solutions.

For the general case of (15) the Global Pareto-optimal solutions are: 0 x1 1 and xi = 0 for i = 2, 3,…,N. In [79] the proposal is N=20.

2.1.6.2 Deceptive Multi­objective Optimisation Problem

A deceptive g function is defined over binary alphabets, thereby making the search space discontinuous. Let us say that the following multi­objective function is defined over l bits, which is a concatenation of N substrings of variable size li such that sum(li)

= l:

Minimise f1 = 1 + u(l1),

Minimise f2 = sum(g(u(li)) / (1+u(l1)), (16) where u(l1) is the unitation of the first substring of length l1.

The first function f1 is a simple one­min problem, where the absolute minimum solution is to have all 0s in the first sub-string. A one is added to make all function values strictly positive.

Figure 5.

10 000 randomly generated individuals and the true deceptive two objectives Pareto-front.

The function g(l1) is defined as g(u(l1)) = 2 + u(l1); if u(l1) < l1, or g(u(l1)) = 1; if u(l1) = l1.

This makes the true attractor (with all values of one in the substring) to have worst neighbours and with a function value g(l1) = 1 and the deceptive attractor (with all 0s in

the substring) to have good neighbours and with a function value g(0) = 2. Since, most of the substrings lead towards the deceptive attractor, GAs may find difficulty to converge to the true attractor (all 1s) as in Figure 5. Since each g function has two minima (one true and another deceptive), there are a total of 2N-1 local minima, of which only one is global. Global Pareto-optimal solutions are: 0 u(x1 li and u(xi) = li

for i = 2, 3,…,N. In [79] the proposal is N=20.

2.1.6.3 Multi­modal Multi­objective Problem

When the function g(x2) is multi­modal with local x2 and global x2 minimum solutions, the corresponding two objective problem also has local and global Pareto­optimal solutions corresponding to solutions (x1; x2) and (x1; x2), respectively. The Pareto­optimal solutions vary in x1 values.

Figure 6.

10 000 randomly generated individuals and the true multi-modal 2 objectives Pareto-front.

We create a modal, two objective optimisation problem by choosing a multi-modal g(x2) function:

Minimize f1(x) = f1(x1, x2,…,xm),

Minimize f2(x) = g(xm+1,…,xN).h(f1(x1,…,xm),g(xm+1,…,xN)), h = g/f1, where

g(x2,…,xN ) = 1+ 10(N-1) + sum(xi2 – 10cos(2xi)). (17) Global Pareto-optimal solutions are: 0 x  1 and x = 0 for i = 2, 3,…,N as in Figure 6.

To highlight the specifics, where the difficulty lies with optimisations along a multi-modal Pareto front, please observe the magnified part of the search space of the solution distribution in Figure 7.

Figure 7.

A magnified portion of the true multi-modal Pareto-front.

2.1.6.4 Convex and Non­convex Pareto­optimal Fronts We choose the following function for h:

Minimize f1(x) = f1(x1, x2,…,xm),

Minimize f2(x) = g(xm+1,…,xN).h(f1(x1,…,xm),g(xm+1,…,xN)), where

h(f1, g) = 1 – (f1/g), if f1g; and 0 otherwise. (18) With this function, we may allow 0f1, g may be any function 0<g. The global Pareto­optimal set corresponds to the global minimum of g function. The parameter  is a normalisation factor to adjust the range of values of functions f1 and g. To have a significant Pareto­optimal region,  may be chosen as f1,max/gmin, where f1,max and gmin are the maximum value of the function f1 and the minimum (or global optimal) value of the function g, respectively.

The above function can be used to create multi­objective problems having convex Pareto­optimal set by setting   1.

Global Pareto-optimal solutions are: 0 x1 1 and xi = 0 for i = 2, 3,…,N as in Figure 8.

In [79] the proposal is to use N=20 in equation (18), which results in a suffitienty versatile search space.

Figure 8.

10 000 randomly generated individuals and the true convex two objectives Pareto-front.

Note that when  > 1, the resulting Pareto­optimal front is non­convex. It is important to note that when  > 1 is used, the classical weighted sum method cannot find any intermediate Pareto­optimal solution by using any weight vector.

Although there exist other methods (such as  ­perturbation method or goal programming method), they require problem knowledge and, moreover, require multiple application of the single objective optimiser.

Global Pareto-optimal solutions are: 0 x1 1 and xi = 0 for i = 2, 3,…,N as presented in Figure 9. In [79] the proposal is to use N=20 in equation (18), which results in a suffitienty versatile search space.

Figure 9.

10 000 randomly generated individuals and the true non-convex two objectives Pareto-front.

2.1.6.5 Discontinuous Pareto­optimal Front

We have to relax the condition for h being a monotonically decreasing function of f1 to construct multi­objective problems with a discontinuous Pareto­optimal front. In the following, we show one such construction where the function h is a periodic function of f1:

Minimize f1(x) = f1(x1, x2,…,xm),

Minimize f2(x) = g(xm+1,…,xN).h(f1(x1,…,xm),g(xm+1,…,xN)),

h(f1, g) = 1 – (f1/g)- (f1/g)sin(2qf1) (19) where g may be any function 0<g. The parameter q is the number of discontinuous regions in an unit interval of f1. Since the h (and hence f2) function is periodic to x1 (and hence to f1), we generate discontinuous Pareto­optimal regions.

Global Pareto-optimal solutions are: 0 x1 1 and xi = 0 for i = 2, 3,…,N as in Figure 10. In [79] the proposal is N=20.

Figure 10.

10 000 randomly generated individuals and the true discontinuous two objectives Pareto-front along the non-dominated region.

2.1.6.6 Biased Search Space

The function g makes a major role in introducing difficulty to a multi­objective problem. Even though the function g is not chosen to be a multi­modal function nor to be a deceptive function, with a simple monotonic g function the search space can have adverse density of solutions towards the Pareto­optimal region. Consider the following function for g:

Minimize f1(x) = f1(x1, x2,…,xm),

Minimize f2(x) = g(xm+1,…,xN).h(f1(x1,…,xm),g(xm+1,…,xN)),

h = g/f1, where (20)

g(xm+1,…xN) = gmin + (gmax - gmin ) ((sum(xi) – sum(xi,min)) / (sum(xi,max) – sum(xi,min))) where gmin and gmax are minimum and maximum function values that the function g may take. The values of xi,max and xi,min are minimum and maximum values of variable xi. It is important to note that the Pareto­optimal region occurs when g takes the value gmin. The parameter  controls the biasness in the search space. If  < 1 then the density of solutions increases while getting further away from the Pareto­optimal front.

Global Pareto-optimal solutions are: 0 x1 1 and xi = 0 for i = 2, 3,…,N as in Figure 11. In [79] the proposal is N=20.

Figure 11.

10 000 randomly generated individuals and the true biased two objectives Pareto-front.

Random search methods are likely to face difficulty in finding the Pareto­optimal front in the case with  close to zero, mainly due to the low density of solutions towards the Pareto­optimal region.

Although multi­objective GAs, in general, will progress towards the Pareto­optimal front, a different scenario may emerge. Although for values of  greater than one, the search space is biased towards the Pareto­optimal region, the search in a multi­objective GA with proportionate selection and without mutation or without elitism is likely to slow down near the Pareto­optimal front. In such cases, the multi­objective GAs may prematurely converge to a front near the true Pareto­optimal front. This is because the rate of improvement in g value near the optimum (x20) is small with 1.

Nevertheless, a simple change in the function g with a change in  suggested above will change the landscape drastically and multi­objective optimisation algorithms may face difficulty in converging to the true Pareto­optimal front.

2.1.6.7 Generalisation of Two Objectives to Four Objectives

The idea is to simply introduce a third function f3 that is similar to the first one f1 but having different domains. The fourth function f4 is chosen to be similar to the second function f2, but also having different domains. Thus a general 4 objective problem is:

Minimize f1(x) = f1(x1, x2,…, xm),

Minimize f2(x) = g(xm+1,…, xN).h(f1(x1,…, xm),g(xm+1,…, xN)),

Minimize f3(x) = f1(xN+1, xN+2,…, xN+m), (21) Minimize f4(x) = g(xN+m+1,…, xN+N).h(f3(xN+1,…, xN+m),g(xN+m+1,…, xN+N)),

where function f1, g and h can be chosen identically as in equations (15)-(20). With the same approach we can duplicate the four objective problems to eight competing objectives and further on to 16 competing objectives.

In document Óbuda University PhD Dissertation (Pldal 31-40)