Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Meta- Heuristic Algorithms Ant Colony Genetic Algorithm Simulated Annealing Ant Colony Optimization ACO Concept • Ants (blind) navigate from nest to food source • Shortest path is discovered via pheromone trails – – – – each ant moves at random pheromone is deposited on path ants detect lead ant’s path, inclined to follow more pheromone on path increases probability of path being followed ACO System • Virtual “trail” accumulated on path segments • Starting node selected at random • Path selected at random – based on amount of “trail” present on possible paths from starting node – higher probability for paths with more “trail” • Ant reaches next node, selects next path • Continues until reaches starting node • Finished “tour” is a solution ACO System, cont. • A completed tour is analyzed for optimality • “Trail” amount adjusted to favor better solutions – better solutions receive more trail – worse solutions receive less trail – higher probability of ant selecting path that is part of a better-performing tour • New cycle is performed • Repeated until most ants select the same tour on every cycle (convergence to solution) ACO System, cont. • Often applied to TSP (Travelling Salesman Problem): shortest path between n nodes • Algorithm in Pseudocode: – Initialize Trail – Do While (Stopping Criteria Not Satisfied) – Cycle Loop • Do Until (Each Ant Completes a Tour) – Tour Loop • Local Trail Update • End Do • Analyze Tours • Global Trail Update – End Do Background • Discrete optimization problems difficult to solve • “Soft computing techniques” developed in past ten years: – Genetic algorithms (GAs) • based on natural selection and genetics – Ant Colony Optimization (ACO) • modeling ant colony behavior Background, cont. • Developed by Marco Dorigo (Milan, Italy), and others in early 1990s • Some common applications: – Quadratic assignment problems – Scheduling problems – Dynamic routing problems in networks • Theoretical analysis difficult – algorithm is based on a series of random decisions (by artificial ants) – probability of decisions changes on each iteration Implementation Ant Algorithms Ant Algorithms Implementation • Can be used for both Static and Dynamic Combinatorial optimization problems • Convergence is guaranteed, although the speed is unknown – Value – Solution The Algorithm • Ant Colony Algorithms are typically use to solve minimum cost problems. • We may usually have N nodes and A undirected arcs • There are two working modes for the ants: either forwards or backwards. • Pheromones are only deposited in backward mode. The Algorithm • The ants memory allows them to retrace the path it has followed while searching for the destination node • Before moving backward on their memorized path, they eliminate any loops from it. While moving backwards, the ants leave pheromones on the arcs they traversed. The Algorithm • The ants evaluate the cost of the paths they have traversed. • The shorter paths will receive a greater deposit of pheromones. An evaporation rule will be tied with the pheromones, which will reduce the chance for poor quality solutions. The Algorithm • At the beginning of the search process, a constant amount of pheromone is assigned to all arcs. When located at a node i an ant k uses the pheromone trail to compute the probability of choosing j as the next node: pijk ij k if j N i lNik il if j N ik 0 • where N is the neighborhood of ant k when in node i. k i The Algorithm • When the arc (i,j) is traversed , the pheromone value changes as follows: ij ij k • By using this rule, the probability increases that forthcoming ants will use this arc. The Algorithm • After each ant k has moved to the next node, the pheromones evaporate by the following equation to all the arcs: ij (1 p) ij , (i, j ) A • where p (0,1]is a parameter. An iteration is a completer cycle involving ants’ movement, pheromone evaporation, and pheromone deposit. Steps for Solving a Problem by ACO 1. 2. 3. 4. 5. 6. Represent the problem in the form of sets of components and transitions, or by a set of weighted graphs, on which ants can build solutions Define the meaning of the pheromone trails Define the heuristic preference for the ant while constructing a solution If possible implement a efficient local search algorithm for the problem to be solved. Choose a specific ACO algorithm and apply to problem being solved Tune the parameter of the ACO algorithm. Applications Efficiently Solves NP hard Problems • Routing – TSP (Traveling Salesman Problem) – Vehicle Routing – Sequential Ordering • Assignment – – – – – QAP (Quadratic Assignment Problem) Graph Coloring Generalized Assignment Frequency Assignment University Course Time Scheduling 1 5 2 3 4 Applications • Scheduling – – – – – – Job Shop Open Shop Flow Shop Total tardiness (weighted/non-weighted) Project Scheduling Group Shop • Subset – – – – – – – Multi-Knapsack Max Independent Set Redundancy Allocation Set Covering Weight Constrained Graph Tree partition Arc-weighted L cardinality tree Maximum Clique Applications • Other – – – – Shortest Common Sequence Constraint Satisfaction 2D-HP protein folding Bin Packing • Machine Learning – Classification Rules – Bayesian networks – Fuzzy systems • Network Routing – Connection oriented network routing – Connection network routing – Optical network routing Genetic Algorithms Part I: GA Theory What are genetic algorithms? How to design a genetic algorithm? Genetic Algorithm Is Not... Gene coding... Genetic Algorithm Is... … Computer algorithm That resides on principles of genetics and evolution Instead of Introduction... Hill climbing • global local Instead of Introduction…(2) Multi-climbers • Instead of Introduction…(3) Genetic algorithm • I am not at the top. My high is better! I am at the top Height is ... I will continue Instead of Introduction…(3) Genetic algorithm - few microseconds • after GA Concept • Genetic algorithm (GA) introduces the principle of evolution and genetics into search among possible solutions to given problem. • The idea is to simulate the process in natural systems. • This is done by the creation within a machine of a population of individuals represented by chromosomes, in essence a set of character strings, that are analogous to the DNA, that we have in our own chromosomes. Survival of the Fittest • The main principle of evolution used in GA is “survival of the fittest”. • The good solution survive, while bad ones die. Nature and GA... Nature Genetic algorithm Chromosome String Gene Character Locus String position Genotype Population Phenotype Decoded structure The History of GA • Cellular automata – John Holland, university of Michigan, 1975. • Until the early 80s, the concept was studied theoretically. • In 80s, the first “real world” GAs were designed. Algorithmic Phases Initialize the population Select individuals for the mating pool Perform crossover Perform mutation Insert offspring into the population no Stop? yes The End Designing GA... How to represent genomes? How to define the crossover operator? How to define the mutation operator? How to define fitness function? How to generate next generation? How to define stopping criteria? Representing Genomes... Representation Example string 1 array of strings 0 1 http avala 1 1 yubc 0 0 net ~apopovic or > c tree - genetic programming xor a b b 1 Crossover • Crossover is concept from genetics. • Crossover is sexual reproduction. • Crossover combines genetic material from two parents, in order to produce superior offspring. • Few types of crossover: – One-point – Multiple point. One-point Crossover 0 7 1 6 2 5 3 4 4 3 5 2 6 1 7 0 Parent #1 Parent #2 One-point Crossover 0 7 1 6 5 2 3 4 4 3 5 2 6 1 7 0 Parent #1 Parent #2 Mutation • Mutation introduces randomness into the population. • Mutation is asexual reproduction. • The idea of mutation is to reintroduce divergence into a converging population. • Mutation is performed on small part of population, in order to avoid entering unstable state. Mutation... Parent 1 1 0 1 0 0 0 1 Child 0 1 0 1 0 1 0 1 About Probabilities... • Average probability for individual to crossover is, in most cases, about 80%. • Average probability for individual to mutate is about 1-2%. • Probability of genetic operators follow the probability in natural systems. • The better solutions reproduce more often. Fitness Function • Fitness function is evaluation function, that determines what solutions are better than others. • Fitness is computed for each individual. • Fitness function is application depended. Selection • The selection operation copies a single individual, probabilistically selected based on fitness, into the next generation of the population. • There are few possible ways to implement selection: – “Only the strongest survive” • Choose the individuals with the highest fitness for next generation – “Some weak solutions survive” • Assign a probability that a particular individual will be selected for the next generation • More diversity • Some bad solutions might have good parts! Selection - Survival of The Strongest Previous generation 0.93 0.51 0.72 0.31 0.12 Next generation 0.93 0.72 0.64 0.64 Selection - Some Weak Solutions Survive Previous generation 0.93 0.51 0.72 0.31 0.12 0.64 Next generation 0.93 0.72 0.64 0.12 Mutation and Selection... D Phenotype D D Solution distribution Phenotype Phenotype Selection Mutation Stopping Criteria • Final problem is to decide when to stop execution of algorithm. • There are two possible solutions to this problem: – First approach: • Stop after production of definite number of generations – Second approach: • Stop when the improvement in average fitness over two generations is below a threshold GA Vs. Ad-hoc Algorithms Speed Genetic Algorithm Ad-hoc Algorithms Slow * Generally fast Minimal Long and exhaustive Applicability General There are problems that cannot be solved analytically Performance Excellent Depends Human work * Not necessary! Problems With Gas • Sometimes GA is extremely slow, and much slower than usual algorithms Advantages of Gas • Concept is easy to understand. • Minimum human involvement. • Computer is not learned how to use existing solution, but to find new solution! • Modular, separate from application • Supports multi-objective optimization • Always an answer; answer gets better with time !!! • Inherently parallel; easily distributed • Many ways to speed up and improve a GA-based application as knowledge about problem domain is gained • Easy to exploit previous or alternate solutions GA: An Example - Diophantine Equations Diophantine equation (n=4): • A*x + b*y + c*z + d*q = s y z q For given a, b, c, d, and s - find x, y, z, q • x Genome: • GA:An Example - Diophantine Equations(2) Crossover • ( 1, 2, 3, 4 ) ( 1, 6, 3, 4 ) ( 5, 6, 7, 8 ) ( 5, 2, 7, 8 ) ( 1, 2, 3, 4 ) ( 1, 2, 3, 9 ) GA:An Example - Diophantine Equations(3) • First generation is randomly generated of numbers lower than sum (s). • Fitness is defined as absolute value of difference between total and given sum: Fitness = abs ( total - sum ) , • Algorithm enters a loop in which operators are performed on genomes: crossover, mutation, selection. • After number of generation a solution is reached. Some Applications of Gas Control systems design Software guided circuit design Optimization Internet search GA search Data mining Path finding Trend spotting Stock prize prediction Mobile robots Part II: Applications of GAs GA and the internet GA and image segmentation GA and system design Simulated Annealing Simulated Annealing(1) • What is annealing? • Process of slowly cooling down a compound or a substance • Slow cooling let the substance flow around thermodynamic equilibrium • Molecules get obtimum conformation contraction : cause stress Simulated Annealing(2) • What is simulated annealing? – Stochastic optimization method – Simulates slow cooling of annealing process – Implemented by Metropolis algorithm, Monte Carlo step – Analogy • • • • Current thermodynamic state = solution Energy equation = objective function Ground state = global optimum Temperature T = ??? Simulated Annealing(3) • Charateristics – Take some upnhill steps to escape the local minimum – Instead of picking the best move, it picks a random move – If the move improves the situation, it is executed. Otherwise, move with some probability less than 1. Simulated Annealing(4) • Simple Iterative Algorithm 1. find a path p 2. make p’, a variation of p 3. if p’ is better than p, keep p’ as p 4. goto 2 • Metropolis Algorithm – 3’ : if (p’ is better than p) or ( within Prob), then keep p’ as p • Simulated Annealing – T is reduced to 0 by schedule as time passes Simulated Annealing(5) • Algorithm function Simulated-Annealing(problem, schedule) returns a solution state inputs: problem, a problem local variables: current, a node next, a node T, a “temperature” controlling the probability of downward steps current Make-Node(Initial-State[problem]) for t1 to infinity do T schedule[t] if T=0 then return current next a randomly selected successor of current DE Value[next] – Value[current] if DE>0 then currentnext else currentnext only with probability eDE/T Simulated Annealing(6) • Schedule • Determines rate at which the temperature is lowered • Lowers T slowly enough, the algorithm will find a global optimum • Temperature T : control parameter • Used to determine the probability • High T : large changes • Low T : small changes Simulated Annealing(7) • Probability distribution p 1 eE/T 0 E Simulated Annealing(8) • Schedule example T0 T(n) Tf moves n Simulated Annealing(9) higher probability of escaping local maxima Little chance of escaping local maxima, but local maxima may be good enough in practical problmes. Simulated Annealing(10) • To avoid of entrainment in local minima – Anealing schedule : by trial and error • Choice of initial temperature • How many iterations are performed at each temperature • How much the temperature is decremented at each step as cooling proceeds Simulated Annealing(11) • Difficulties – Determination of parameters • Schedule T • # of moves at each temperature – If cooling is too slow • Too much time to get solution – If cooling is too rapid • Solution may not be the global optimum Simulated Annealing(12) • Application – VLSI layout – Signal and image processing – Task allocation – Network design – Molecular analysis Application in CSP • Can be solved by IIA – Allow states with unsatisfied constraints operators reassign variable values • Heuristic repair method – Assigning values to all the variables – Applying modification operators which simply assign a different value to a variable • Min-conflicts heuristic – Choose value that violates the fewest constraints i.e., hillclimb with h(n) = total number of violated constraints – Surprisingly effective (ex:million-queens in less than 50 steps) Min-conflicts heuristics • Example : 4-Queens – States : 4 queens in 4 columns(44 = 256 states) – Operators : move queen in column – Goal test : no attacks – Evaluation : h(n) = number of attacks h=5 h=2 h=0 Summary • Iterative improvement algorithms – Keep only single state in memory – Can get stuck on local maxima – Simulated annealing to escape local maxima • Complete, optimal given a long enough cooling schedule • For CSP – Variable and value ordering heuristics can provide huge performance gains – Current algorithms often solve very large problems very quickly