• No results found

Aspects of approximate optimisation: overcoming the curse of dimensionality and design of experiments

N/A
N/A
Protected

Academic year: 2021

Share "Aspects of approximate optimisation: overcoming the curse of dimensionality and design of experiments"

Copied!
8
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

ASPECTS OF APPROXIMATE OPTIMISATION: OVERCOMING THE CURSE

OF DIMENSIONALITY AND DESIGN OF EXPERIMENTS

Sophie Trichon

Department of Aerospace and Mechanical Engineering

University of Liège Liège, Belgium sophie.trichon@ulg.ac.be

Martijn Bonte

Faculty of Engineering Technology University of Twente Enschede, The Netherlands

m.h.a.bonte@utwente.nl

Jean-Philippe Ponthot

Department of Aerospace and Mechanical Engineering

University of Liège Liège, Belgium jp.ponthot@ulg.ac.be

Ton van den Boogaard

Faculty of Engineering Technology University of Twente Enschede, The Netherlands a.h.vandenboogaard@utwente.nl

EXTENDED ABSTRACT

Coupling optimisation algorithms to Finite Element Methods (FEM) is a very promising way to achieve optimal metal forming processes. However, many optimisation algorithms exist and it is not clear which of these algorithms to use. This paper investigates the sensitivity of a Sequential Approximate Optimisation algorithm (SAO) proposed in [1-4] to an increasing number of design variables and compares it with two other algorithms: an Evolutionary Strategy (ES) and an Evolutionary version of the SAO (ESAO). In addition, it observes the influence of different Designs Of Experiments used with the SAO. It is concluded that the SAO is very capable and efficient and its combination with an ES is not beneficial. Moreover, the use of SAO with Fractional Factorial Design is the most efficient method, rather than Full Factorial Design as proposed in [1-4].

INTRODUCTION

During the last decades, Finite Element (FEM) simulations of metal forming processes have become important tools for designing feasible production processes. More recently, several authors recognised the potential of coupling FEM simulations to mathematical optimisation algorithms to design optimal metal forming processes instead of only feasible ones. However, many optimisation algorithms exist and it is not clear which of these algorithms to use.

The present research is based on the Sequential Approximate Optimisation (SAO) algorithm which uses both Response Surface Methodology (RSM) and Design and Analysis of Computer Experiments (DACE). The efficiency of this algorithm decreases with increasing number of design variables because of the high number of FEM simulations required. To overcome such a problem two solutions have been investigated.

The first one is to use an Evolutionary Strategy (ES), which is good to find the global optimum and is able to deal with a higher number of design variables. Unfortunately, it requires many function evaluations. Behind this idea was to create an algorithm called ESAO, which is an Evolutionary version of the SAO, to keep the benefit of the ES while using the metamodels of the SAO to avoid too many function evaluations.

The second solution consists in comparing different Design Of Experiments (DOE) for a fixed number of design variables of the SAO. The DOE used are: Full Factorial Design, Fractional Factorial Design or Latin Hypercube Design.

In the first part of this paper, the three algorithms will be briefly explained: SAO, ES and ESAO. In Section II the sensitivity of the three algorithms to an increasing number of design variables taken into account in optimisation problems is investigated. Section III is dedicated to a comparison between the different DOE’s and conclusions are presented in Section IV.

(2)

I THE OPTIMISATION ALGORITHMS

In this first part the three optimisation techniques used in the present research are briefly described.

I-1. Sequential Approximate Optimisation algorithm (SAO)

The Sequential Approximate Optimisation algorithm uses both Response Surface Methodology (RSM) and Design and Analysis of Computer Experiments (DACE) metamodelling techniques [1-4]. SAO is summarised in Fig. 1.

The first step, modelling, consists of carefully choosing the design variables, constraints, and objective function.

Then it uses a Design Of Experiment (DOE) strategy to define starting points. The two designs combined in the SAO are Latin Hypercube Design (LHD), defining points in a spacefilling way, and Full Factorial Design (FFD) selecting all the corner points of the domain as illustrated in Fig. 2.

The third step is running the FEM simulations for each of the DOE points. Once the objective function values are known for each point, the algorithm fits metamodels based on polynomial regression (RSM) or Kriging interpolation (DACE). All of these metamodels are then validated using cross validation before the best one is selected. This final metamodel is used to run a multistart SQP algorithm. This is the optimisation step.

Finally, if the evaluation of the optimum, i.e. FEM calculation of the optimal design variables found, is not satisfying we can improve the metamodel by adding DOE points to the previous ones. This is the sequential improvement. It can be done using several options. In our case we will minimise a merit function based on the parameters of the metamodels [4].

Step 1 - Modelling, DOE Step 2 - Run FEM simulations

Step 3 - Fit metamodels (RSM and DACE) Step 4 - Validate and select best metamodel Step 5 - Optimise

Step 6 - Evaluate optimum Step 7 - Sequential improvement.

Figure 1. The Sequential Approximate Optimisation Algorithm

Figure 2. Latin Hypercube Design (inside points) + Full factorial design (corner points) The SAO algorithm has been implemented in MATLAB [1-4] and can be used in combination with any FEM code. Furthermore, the fitting of the DACE/Kriging metamodels uses the MATLAB Kriging toolbox implemented by Lophaven, Nielsen and Søndergaard [5].

I-2. Evolution Strategy with Covariance Matrix Adaptation (CMA-ES)

The employed ES has been developed to solve optimisation problems with a nonlinear objective function. It is summarised in Fig. 3. In general, the method should be applied if derivative based methods, e.g. quasi-Newton BFGS or conjugate gradient, (supposedly) fail due to a rugged search landscape (e.g. discontinuities, sharp bends and ridges, noise, local optima, outliers).

The CMA-ES has been developed in such a way that we almost have no parameters to define. Basically one just needs to choose the size of the population λ.

When the algorithm starts, a population of new search points (λ individuals) is generated by sampling a multivariate normal distribution based on the covariance matrix. Normal distribution is used for many reasons; one of them is that it implies invariance (does not favour any direction). The normal distribution is based on the eigenvectors and deviations of the covariance matrix in order to introduce mutation in the new generation.

Step 1 - Initialisation, parameters settings Step 2 - Sample new λ search points Step 3 - Selection of µ parents Step 4 - Recombination Step 5 - Step size control

Step 6 - Covariance matrix adaptation Figure 3. The Covariance Matrix Adaptation

(3)

Once this population is created, it evaluates the λ objective function values and selects the best µ points. These µ parents are recombined using weighted recombination. In order to update the covariance matrix, a step size control has been introduced that prevents premature convergence. The last step before sampling the next generation is the covariance matrix adaptation that increases the probability of successful steps to appear again.

The CMA-ES algorithm has been implemented in MATLAB by N. Hansen, from Technical University of Berlin [6] and can also be used in combination with any FEM code to evaluate the λ children.

I-3. Evolutionary SAO (ESAO)

ES are known to be effective optimisation techniques but they involve a large number of FEM simulations what we absolutely want to avoid. The idea here is to combine both the SAO and the CMA-ES so that the advantages of one algorithm counterbalance the disadvantages of the other. The new algorithm, ESAO, is described in Fig. 4 and resembles the Metamodel Assisted Evolutionary Strategy proposed by Emmerich et al. [7].

The ESAO algorithm starts with the first steps of the SAO: choosing the design variables, constraints, and objective function, creating the DOE, running FEM simulation and fitting metamodels.

Then, on the selected metamodel, we introduce the CMA-ES. The starting µ parents of the ES are chosen to be the best µ points of the initial DOE. Then we generate λ offspring individuals and evaluate them on the metamodel. Following with the CMA-ES, we select the best µ points based on the metamodel, add them to the DOE points and run the corresponding FEM simulations in order to update the metamodels and increase their accuracy.

Step 1 - Modelling Step 2 - DOE

Step 3 - Run FEM simulations

Step 4 - Fit metamodels and select best one Step 5 - Generate λ offspring individuals (weighted recombination of µ parents, CMA, normal sampling)

Step 6 - Evaluation on metamodel Step 7 - Select best µ parents.

Figure 4. The Evolutionary Sequential Approximate Optimisation algorithm

Then the recombination step occurs, followed by covariance matrix adaptation and new population sampling. The new generation will be evaluated on the updated metamodels again and so forth until a stopping criterion is met.

We note that the evaluation of the λ individuals is made on the metamodel thus avoiding λ-µ highly time consuming FEM simulations. Then we run the FEM simulations only on the best µ selected points, becoming the parents of the new generation. This is the key difference between the CMA-ES and the ESAO. In the CMA-ES all of the λ search points are evaluated by FEM simulations.

II Comparison between SAO, CMA-ES and ESAO

In the first part of this section one can find the description of the test function and parameters used to obtain the results exposed in the second section. The results are used to compare the algorithms and investigate their sensitivity with respect to the number of variables.

II-1. Test function

All simulations presented in this part are realised with the following Generalised Rosenbrock Function [Equation (1)] with global optimum described in [Equation (2)]:

(

)

(

)

(

)

− = +

+

=

1 1 2 2 1 2

1

100

)

(

N i i i i

x

x

x

f x

(1)

0

)

(

x

=

f

for xi =1, i=1,K,N (2)

Where x is the vector containing the design variables and N is the size of x, i.e. the number of design variables. Figure 5 illustrates the 2-dimensional case.

(4)

The Rosenbrock function is considered to be difficult, because it has a ridge. The top of the ridge is very sharp, and it runs around a parabola. To find the valley is trivial, however convergence to the global optimum is difficult.

The problem we choose to optimise is the following:

Minimise: f (x) With: xi in [0 1.5],

For i = 1,…,N and N = [2, 3, 5, 7, 10] In our method we also need to define the number of starting DOE points and the maximum number of points we can use during the improvement part. In this test we decide to start with 10N DOE points and improve until 20N points.

Regarding the ES strategy we also have to define λ and µ. These parameters change depending on N. Table 1 represents the corresponding values. To summarise, we keep the best 25% of the λ search points to generate the next population.

Table 1. Values of λ and µ depending on the number of design variables.

N λ µ 2 12 3 3 14 3 5 16 4 7 18 4 10 20 5 II-2. Results

First we want to investigate the sensitivity of the three algorithms from two to ten variables. Up

to five variables we can use the SAO algorithm, with LHD and FFD but for more than five variables it becomes impossible with the condition we defined: 10N DOE points to start. One difficulty with factorial designs is that the number of combinations increases exponentially with the number of variables one wants to

manipulate. FFD is made of 2N corner points.

Hence in 10 dimensions it would be 210 = 1024

corner points, which is not an option to consider. For this reason we do not use the FFD for more than five variables and consider LHD only. The corresponding results will be denoted as LHD (Latin Hypercube Design).

Figure 6 represents the objective function improvement vs. the number of variables. For each of the three algorithms, we use the same starting DOE and the same starting objective function value. One can see in solid line the results with standard DOE’s (FFD + LHD), and with dashed lines we use only LHD. The same results are reported in Table 2.

By observing these first results, we can see that the SAO algorithm is quite constant and efficient when using only LHD. Standard SAO, using LHD and FFD, is very effective when the number of variables is below five. Once again, this is totally related to the number of corner points. A lot of corner points implies fewer points inside the domain, and consequently a less accurate metamodel.

CMA-ES also gives good results but it is not constant. This is the hazard of ES’s. We will see later that it needs more iterations than SAO to obtain the same results.

0% 20% 40% 60% 80% 100% 1 2 3 4 5 6 7 8 9 10 11

Num ber of Variables

O b j Fun c ti on im pr ov e m e n t ESA O CM A -ES SA O LHD LHD LHD

(5)

Table 2. Values of λ and µ depending on the number of design variables

Initial Optimal values & Improvement (%) Initial Optimal values & Improvement (%)

Var.

Values SAO CMA-ES ESAO Var. Values SAO CMA-ES ESAO

2 0,27 0 100% 0,1 60% 0,05 82% 3 1,6 0,03 98% 1,03 37% 0,9 45% 3 2,00 0,02 99% 1,9 5% 0,9 52% 5 23,6 4 83% 4 83% 10,3 56% 5 4,00 3,1 20% 3,7 6% 3,2 18% 7 78,3 5,3 93% 42 46% 36,6 53%

10 172,9 29,7 83% 40 76% 109,6 37%

(a) (b)

ESAO is not as good as expected. One of the reasons is the combination of a not very accurate metamodel with random sampling by CMA-ES.

In the following figures we can see the evolution of the objective function values from 10N to 20N DOE points.

Figure 7 shows the results for three and five variables. The graphs represent the evolution of the optimal values for standard DOE (FFD+LHD) and only for LHD for each of the three algorithms. Figure 8 corresponds to seven and ten variables and contains only LHD as explained previously.

In order to be able to compare the algorithms, for each number of variables the three algorithms start with the same 10N DOE points and the same objective function value. We notice two different starting values for the first two graphs. This is because the two DOE’s used do not necessarily contain points in common and then do not give the same first optimal value.

In addition we can observe that some curves stop before 20N DOE points. This is because they give very good results very quickly. What we want to know is just which of the methods is the

most efficient and for that we do not always need to achieve the 20N FEM simulations.

In each of the graphs, the SAO algorithm is the first one to obtain the best improvement. We are now sure that, in the case of this test function, SAO is better than ES and the combination of both algorithms. Indeed ESAO does not give any improvement.

Moreover for an increasing number of design variables, we definitely have to use LHD alone, rather than its combination with FFD as proposed in [1-4].

III COMPARISON OF DESIGNS OF EXPERIMENT

Results from Section II point out that there are some differences to investigate between the DOE’s. In this part we use the SAO algorithm and new DOE’s to compare their efficiency.

One more thing we have to observe is the influence of the position of the optimum. Indeed if we play with corner points in the DOE, perhaps the position near or far from such corner points may be non-negligible. (a) (b) 0,0 0,5 1,0 1,5 2,0 30 35 40 45 50 55 60

Number of FEM simulations

O b je c tiv e fu n c ti o n v a lu e 0 5 10 15 20 25 50 60 70 80 90 100

Number of FEM simulations

ES Full ES LHD

SAO Full SAO LHD

ESAO Full ESAO LHD

(6)

(a) (b) 0 20 40 60 80 70 90 110 130

Num ber of FEM sim ulations

Ob je c tiv e f u n c ti o n v a lue 0 30 60 90 120 150 180 100 120 140 160 180 200

Number of FEM simulations

ES SAO ESAO

Figure 8. Evolutions of Optimal Values: (a) Seven variables, (b) Ten variables.

III-1. Design Of Experiment (DOE)

We already explained that the SAO algorithm is based on both Latin Hypercube Design (LHD) and Full Factorial Design (FFD). In this part we will use another design called Fractional Factorial Design illustrated in Fig. 9-b. The difference between Full and Fractional is that we do not use all the corner points but only a fraction, half of them for instance.

In order to be able to compare the different DOE’s we use five variables in the new experiments. Hence we will have the following designs:

- Full Factorial 25 = 32 corner points,

- Fractional Factorial 25-1 = 16 corner points,

- Fractional Factorial 25-2 = 8 corner points,

- Latin Hypercube Design 0 corner points. Both Fractional Factorial Designs will be denoted as ‘Frac 8’, eight corner points, and ‘Frac16’, sixteen corner points.

(a) (b)

Figure 9. (a) 23 full factorial design, (b) 23-1

fractional factorial design

Each of the Full and Fractional Factorial Designs are combined with LHD in order to obtain the desired number of starting DOE points: 10*5 = 50 points.

III-2. Results

In this section we use the same test function as previously, the Rosenbrock function. But the domain of the variables will change. We will compare the three following cases:

- 5 variables in [0 1,5]

- 4 variables in [0 1,5], 1 in [-1 1] (2nd variable)

- 4 variables in [0 1,5], 1 in [-1 1] (5th variable)

- 3 variables in [0 1,5], 2 in [-1 1] (2nd & 5th

variables)

In Fig. 10 one can observe the evolution of the objective function values vs. the number of DOE points for each of the four different DOE’s. It is important to note that in each picture, four different DOE’s are compared, consequently the optimisation starts with different objective function values.

Even if SAO gives quite good results with all of the DOE’s, the ‘Frac 8’ (Fractional Factorial Design 25-2) is always the first one to find the best

results.

When the optimum is inside the domain the FFD is not accurate enough because of the use of too many corner points. We can see it starts with a good optimal value, lucky choice of the DOE points, but it does not improve it whereas in only 20 FEM simulations ‘Frac 8’ finds better optimal values.

(7)

5 variables inside the domain 0 20 40 60 50 60 70 80 90 100

Number of FEM simulations

O b ject ive f u n c ti o n val u e 2nd variable on boundary 0 10 20 30 40 50 60 70 80 90 100

Number of FEM simulations

(a) (b)

Number of FEM simulations

LHD Full Frac 8 Frac 16

5th variable on boundary 0 15 30 45 60 75 50 60 70 80 90 100

Number of FEM simulations

O b jec ti v e f u n c ti o n valu e

2nd & 5th variables on boundaries

0 10 20 30 40 50 60 70 80 90 100

Number of FEM simulations

(c) (d)

Figure 10. Comparison of the DOE’s: (a) Optimum inside the domain, (b) 2nd variable on the boundary,

(c) 5th variable on the boundary, (d) Two variables on the boundary. IV CONCLUSIONS

In the scope of a project called Optimisation of Forming Processes, an algorithm to solve optimisation problems for metal forming processes using time consuming FEM simulations has been developed.

The efficiency of this Sequential Approximate Optimisation (SAO) algorithm decreases with increasing number of design variables because of the high number of FEM simulations required. To overcome such a problem two solutions have been investigated in the present research.

The first one was the combination of the SAO algorithm with an ES to avoid too many function evaluations; the second one was to use different

DOE’s in order to find more accurate metamodels using less function evaluations.

Even if these comparisons have not yet been applied to a real industrial study, it can be concluded that the SAO algorithm outperforms both the ES and the ESAO algorithm, i.e. it gives more accurate results for less iterations. This is due to the accuracy of the metamodels which is not precise enough to be combined with a random method; it alters too much the evolution of the CMA-ES.

In the third section, it is especially emphasised that when the number of design variables is larger than five in the SAO algorithm, then the choice of the DOE becomes crucial. The comparison of different DOE’s shows that the SAO algorithm

(8)

associated with a fractional factorial design is the most efficient method.

To conclude SAO is very capable and efficient and the use of corner points or the combination with an ES is not beneficial in the case of the Rosenbrock function. It is recommended that some other test are performed to validate this work and in addition a comparison to an industrial metal forming case using FEM simulations as function evaluations would be of great importance.

ACKNOWLEDGEMENTS

This research has been carried out in the framework of the projects “Optimisation of Forming Processes MC1.03162” and the First Europe “METAOPT”.

The first project is part of the research programme of the Netherlands Institute for Metals Research (NIMR). The industrial partners co-operating in this project are gratefully acknowledged for their useful contributions to this research.

The support of the Région Wallonne and the Fonds Social Européen under the project First-Europe EPH3310300R075F/415823 is also gratefully acknowledged.

REFERENCES

1. M. Bonte, A. van den Boogaard and J. Huétink, Metamodelling techniques for the

optimisation of metal forming processes,

Proceedings of ESAFORM, Cluj-Napoca,: Romania, 2005, pp. 155-158.

2. M. Bonte, A. van den Boogaard and J. Huétink, Solving optimisation problems in metal

forming using Finite Element simulation and metamodelling techniques, Proceedings of

APOMAT, Morschach, Switzerland, 2005, pp. 242-251.

3. M. Bonte, A. van den Boogaard and J. Huétink, A metamodel based optimisation

algorithm for metal forming processes, accepted

for Advanced Methods in Material Forming. 4. M. Bonte, T. Do, L. Fourment, A. van den Boogaard, J. Huétink and A. Habbal, A

comparison of optimisation algorithms for metal forming processes, Proceedings of ESAFORM,

Glasgow, UK, 2006, pp 883-886.

5. H.B. Nielsen, Dace, A Matlab Kriging

toolbox, Technical University of Denmark, http://www.imm.dtu.dk/~hbn/dace.

6. N. Hansen, The CMA Evolution Strategy, Technical University of Berlin, Germany,

http://lautaro.bionik.tu-berlin.de/user/niko.

7. M. Emmerich, A. Giotis, M. Özdemir, T. Bäck and K. Giannakoglou, Metamodel-Assisted

Evolution Strategies, in Parallel Problem Solving

from Nature - PPSN VII, Springer-Verlag, 2002, pp 361-370.

Referenties

GERELATEERDE DOCUMENTEN

Het verschil in effect op de rijsnelheid van de twee matrixborden is voor richting 1 niet erg groot en voor richting 2 bleek het continu brandende bord

If the publication is distributed under the terms of Article 25fa of the Dutch Copyright Act, indicated by the “Taverne” license above, please follow below link for the End

Het aandeel van de Nederlandse aanlandingen in de totale internationale aanlandingen van de Noordzee voor schol, tong, kabeljauw en wijting bedraagt tesamen 42%.. Het aandeel van

Wanneer verbanden kunnen worden gelegd tussen deze verschillende gegevens én er een vertaalslag kan worden gemaakt naar concrete acties, is ondersteuning door middel

We evaluated the impact of Prosopis invasion and clearing on vegetation species composition, diversity (alien and indigenous species richness), and structure (alien and

Verder maakt Overdiep in zijn tabel (wel af en toe in de tekst) geen onderscheid tussen gevallen waarin het deelwoord wordt gevolgd door een voorzetselgroep, en gevallen waar- in

Per slot van rekening lijkt het dat de handelsbalans tussen Noord-Amerika en de Lage Landen in de historische taalkunde een beetje asymmetrisch is: Noord- Amerika heeft veel