What is Optimization?
- “The selection of a best element (with regard to some criteria) from some set of available alternatives” {Source: Wikipedia}
- Picking the “best” option from several ways of accomplishing the same task.
- We require a model that can give us the best result.
Global vs. Local Optimization
Consider a hypothetical multi-modal function:
Genetic Algorithm
Genetic algorithm (GA), first proposed by John Holland in 1975 and described in Goldberg (1988), is based on analogies with biological evolution processes. The principle of GA is quite simple and mirrors what is perceived to occur in evolutionary mechanisms. The idea is to start with a given set of feasible trial solutions (either constrained or unconstrained) and iteratively evaluate objective (or fitness) function by keeping only those solutions that give the optimal value of the objective function and randomly mutate them to try and do even better and successive iterations. Thus, the beneficial mutations (that lead to optimal values) are kept while those that perform poorly are discarded, i.e., the survival of the fittest.
Consider the unconstrained optimization problem with the objective function, where is an dimensional vector. Suppose initial guesses are given for the values of so that
Thus, solutions are evaluated and compared with each other to see which of the solutions generate the smallest objective function since our goal (in general) is to minimize it. We can order the guesses such that the first give the smallest values of . Arranging our data, we have:
Since the first solutions are the best, these are kept in the next generation. In addition, we now generate new trial solutions that are randomly mutated from the best solutions. This process is repeated through a finite number of iterations with the hope that convergence to the optimal solution is achieved.
Scheme
The steps involved in the Genetic Algorithm are:
- The initial population of model parameters is randomly chosen within the given search range.
- The simple GA undergoes a set of operations on the model population to produce the next generation: selection, coding, crossover, and mutation.
- In the selection scheme, the model parameters exhibiting the higher fitness value (lower objective function value relative to others) are selected and replicated with the given probability. The total population size remains constant; then, the population members are randomly paired among themselves.
- In the coding scheme, each population’s decimal values are converted to a binary system, forming a long bit string (analogous to a chromosome).
- In the crossover scheme, some part of the long bit string of binary model parameters is exchanged with their corresponding pair to produce a new population.
- In the mutation scheme, some randomly selected sites (with a given probability) of the new set of the binary model population are switched.
- These sets of operations will continue until some pre-defined termination criteria for the technique are satisfied.
The Optimum Size of a Coffee Mug for a Café Owner?
Here, the objective function is to minimize the loss of the Café owner. Let us make a very simple function for that.
where is the size of the coffee mug. If the coffee mug’s size is too small, then the customers may not like it, and if the size is too large, then the owner will have to endure loss, hence the hypothetical function. We assumed that the price of the coffee is fixed.
Solution
Let us solve this iteratively:
Generation 1
The average fitness in this generation is: 52.55
.
Generation 2
The average fitness in this generation is 29.17
. We can continue iterating as long as our termination criteria are satisfied. Some of the standard termination criteria are the tolerance value (the difference in the fitness of the generation compared to the previous generation), the maximum number of generations, etc.
Fitness with Generations
Earthquake Location Using Genetic Algorithm
In this series of earthquake location problems, we have used the Monte Carlo method for the earthquake location that gives us reliable results. The earthquake location problem was introduced in the previous post.
Now, I applied GA from the direct search toolbox of MATLAB to find the solution of the best earthquake location parameters for the earthquake problem defined in previous post for 30 stations. The GA does not improve the solutions but similar to the Monte Carlo method; it allows a large model space to be searched for solutions. I used the lower and upper limits for the search of best parameters as defined in Table 1. After 85 generations of a run and terminated based on the tolerance criterion, the best parameters found by the GA are [2.9750 2.5202 -2.5535 6.6696 -0.0839]. Since GA is a stochastic optimization scheme, it is best practice to make a reasonable number of acceptable solutions and consider its mean as the final solution. For the 50 runs of the GA for the earthquake location problem, the estimated mean and std of the model parameters are shown in Table 2. The difference from the actual model parameters arises because of random noise in the observed arrival time data.
lower=[-3 -3 -3 5 -1];
upper=[3 3 0 7 1];
options = optimoptions('ga','PlotFcn', @gaplotbestf,'Display','iter'); 6. x=ga(@(pp)fit_arrival_times(pp),5,[],[],[],[],lower,upper,[],options)
function E=fit_arrival_times(pp)
filename = 'arrival_times.csv';
[dobs, dpre] = read_arrivaltimes(filename);
stationlocations = read_stnloc('station_locations.csv', 2, 31);
d_pre = zeros(length(stationlocations),1);
for i=1:length(stationlocations)
dist = sqrt((pp(1) - stationlocations(i,1))^2 + (pp(2) - stationlocations( i,2))^2 + (pp(3) - stationlocations(i,3))^2);
arr = dist/pp(4) + pp(5);
d_pre(i) = arr;
end
E=sum((dobs-d_pre).^2);
Complete Codes
Complete codes can be downloaded from my GitHub repository.