Evolutionary Search. Announcements. Announcements. Homework #1 Clarifications. Genetic Algorithms. Homework #1 Clarifications

Similar documents
Laboratorio di Intelligenza Artificiale e Robotica

Seminar - Organic Computing

Reinforcement Learning by Comparing Immediate Reward

While you are waiting... socrative.com, room number SIMLANG2016

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

Laboratorio di Intelligenza Artificiale e Robotica

Lecture 1: Machine Learning Basics

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

A simulated annealing and hill-climbing algorithm for the traveling tournament problem

LEGO MINDSTORMS Education EV3 Coding Activities

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

The dilemma of Saussurean communication

A Comparison of Annealing Techniques for Academic Course Scheduling

Agents and environments. Intelligent Agents. Reminders. Vacuum-cleaner world. Outline. A vacuum-cleaner agent. Chapter 2 Actuators

TABLE OF CONTENTS TABLE OF CONTENTS COVER PAGE HALAMAN PENGESAHAN PERNYATAAN NASKAH SOAL TUGAS AKHIR ACKNOWLEDGEMENT FOREWORD

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Lecture 10: Reinforcement Learning

Discriminative Learning of Beam-Search Heuristics for Planning

Evolution of Symbolisation in Chimpanzees and Neural Nets

Intelligent Agents. Chapter 2. Chapter 2 1

BMBF Project ROBUKOM: Robust Communication Networks

Research Article Hybrid Multistarting GA-Tabu Search Method for the Placement of BtB Converters for Korean Metropolitan Ring Grid

CSC200: Lecture 4. Allan Borodin

On the Combined Behavior of Autonomous Resource Management Agents

Getting Started with Deliberate Practice

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Rule Learning with Negation: Issues Regarding Effectiveness

Spring 2015 IET4451 Systems Simulation Course Syllabus for Traditional, Hybrid, and Online Classes

FF+FPG: Guiding a Policy-Gradient Planner

UC Merced Proceedings of the Annual Meeting of the Cognitive Science Society

An Investigation into Team-Based Planning

Rule Learning With Negation: Issues Regarding Effectiveness

Evolution in Paradise

TD(λ) and Q-Learning Based Ludo Players

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Implementation of Genetic Algorithm to Solve Travelling Salesman Problem with Time Window (TSP-TW) for Scheduling Tourist Destinations in Malang City

Knowledge-Based - Systems

Axiom 2013 Team Description Paper

Artificial Neural Networks written examination

A Case-Based Approach To Imitation Learning in Robotic Agents

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

An Introduction to Simio for Beginners

Parsing of part-of-speech tagged Assamese Texts

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Visual CP Representation of Knowledge

A Reinforcement Learning Variant for Control Scheduling

Alpha provides an overall measure of the internal reliability of the test. The Coefficient Alphas for the STEP are:

Python Machine Learning

EVOLVING POLICIES TO SOLVE THE RUBIK S CUBE: EXPERIMENTS WITH IDEAL AND APPROXIMATE PERFORMANCE FUNCTIONS

A Version Space Approach to Learning Context-free Grammars

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

AMULTIAGENT system [1] can be defined as a group of

SARDNET: A Self-Organizing Feature Map for Sequences

College Pricing. Ben Johnson. April 30, Abstract. Colleges in the United States price discriminate based on student characteristics

SCORING KEY AND RATING GUIDE

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots

Ordered Incremental Training with Genetic Algorithms

The Nature of Exploratory Testing

Course Content Concepts

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Learning From the Past with Experiment Databases

Concept Acquisition Without Representation William Dylan Sabo

Navigating the PhD Options in CMS

Ricochet Robots - A Case Study for Human Complex Problem Solving

Radius STEM Readiness TM

Cognitive Modeling. Tower of Hanoi: Description. Tower of Hanoi: The Task. Lecture 5: Models of Problem Solving. Frank Keller.

Liquid Narrative Group Technical Report Number

Learning Prospective Robot Behavior

Comparison of network inference packages and methods for multiple networks inference

Learning and Transferring Relational Instance-Based Policies

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

Modeling user preferences and norms in context-aware systems

Computers Change the World

BIOS 104 Biology for Non-Science Majors Spring 2016 CRN Course Syllabus

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Major Milestones, Team Activities, and Individual Deliverables

Learning Methods for Fuzzy Systems

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

A Genetic Irrational Belief System

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Learning goal-oriented strategies in problem solving

Natural Language Processing. George Konidaris

A General Class of Noncontext Free Grammars Generating Context Free Languages

Software Development: Programming Paradigms (SCQF level 8)

Multimedia Application Effective Support of Education

Go fishing! Responsibility judgments when cooperation breaks down

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

CS 100: Principles of Computing

Classification Using ANN: A Review

An Introduction to the Minimalist Program

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

Examining the Structure of a Multidisciplinary Engineering Capstone Design Program

Managerial Decision Making

The Strong Minimalist Thesis and Bounded Optimality

Evolutive Neural Net Fuzzy Filtering: Basic Description

Thesis-Proposal Outline/Template

ABSTRACT. A major goal of human genetics is the discovery and validation of genetic polymorphisms

Designing a Computer to Play Nim: A Mini-Capstone Project in Digital Design I

Learning to Schedule Straight-Line Code

Transcription:

Evolutionary Search Burr H. Settles CS-540, UW-Madison www.cs.wisc.edu/~cs540-1 Summer 2003 1 Announcements This week s mailing list topic: think of a realworld problem where we could apply an optimization search You may not repeat someone else s answer! What are the states? What are the actions? What is the objective function? Read Chapter 6 in AI: A Modern Approach for next time 2 Announcements Homework #1 Clarifications If you choose to do a paper instead of a programming project, 6 pages is a minimum you may write more if you feel that there is too much material (but please no more than 10) Keep in mind each group will only have 15 minutes to present on the last week The presentations don t have to go into as much detail as the paper, though For problem 1, part b, your heuristics are for the 2-letter change problem, and don t have to be admissible But do think about if they are or aren t, and explain Even though hill-climbing and beam searches are local strategies (and often used for complete searching), you can still use them to solve partial search problems like the graph in problem 2 3 4 Homework #1 Clarifications Genetic Algorithms For problem 3, let s say you have the initial state ABC, and are doing standard HC what are the neighbors you need to consider? So far the optimization strategies we ve discussed search for a single solution, one state at a time within a neighborhood BC, AC, AB, ABCD, ABCE, ABCF So to do hill-climbing, you will generate all these states, score them all, and choose the best one (since we are maximizing the objective function) Genetic algorithms (GAs) are a unique search approach that maintains a population of states, or individuals, which evolves Also called evolutionary search 5 6 1

Evolutionary Analogy Evolutionary Analogy Consider a population of rabbits: some individuals are faster and smarter than others Slower, dumber rabbits are likely to be caught and eaten by foxes Fast, smart rabbits survive to do what rabbits to best: make more rabbits!! The rabbits that survive breed with each other to generate offspring, which starts to mix up their genetic material Fast rabbits might breed with fast rabbits Fast rabbits with slow rabbits Smart with not-so-smart, etc Furthermore, nature occasionally throws in a wild hare because genes can mutate 7 8 Evolutionary Analogy Core Elements of GAs In this analogy, an individual rabbit represents a solution to the problem (i.e. a single point in the state space) The state description is its DNA, if you will The foxes represent the problem constraints Solutions that do well are likely to survive For selection, we use a fitness function to rank the individuals of the population For reproduction, we define a crossover operator which takes state descriptions of individuals and combines them to create new ones What advantages does this present over local search? What we need to create are notions of natural selection, reproduction, and mutation 9 For mutation, we can merely choose individuals in the population and alter part of its state 10 Genetic Algorithm Example Genetic Algorithm Example POP = initialpopulation // build a new population repeat { // with every generation NEW_POP = empty for I = 1 to POP_SIZE { X = fit individual // natural selection Y = fit individual CHILD = crossover(x,y) // reproduction if small random probability then mutate(child) // mutation add CHILD to NEW_POP } POP = NEW_POP } until solution found or enough time elapsed return most fit individual in POP 11 The previous algorithm completely replaces the population for each new generation but we can allow individuals from older generations to live on Reproduction here is only between two parents (as in nature), but we can allow for more!! The population size also is fixed but we can have this vary from one generation to the next 12 2

Genetic Algorithm Example Selection Basically, there is no one GA, we can devise many variants of these 3 principles for nearly any problem!! Chapters 7 & 8 in How to Solve It: Modern Heuristics have a very thorough presentation of how to design genetic algorithms for particular problems 13 Selection (either to reproduce or live on) from one generation to the next relies on the fitness function We can usually think of the fitness function as being a heuristic, or the objective function We want to apply pressure that good solutions survive and bad solutions die Too much and we converge to sub-optimal solutions Too little and we don t make much progress 14 Selection Selection Deterministic selection 1. Rank all the individuals using the fitness function and choose the best k to survive 2. Replace the rest with offspring Can lead fast convergence (and local optima) Stochastic selection Instead of selecting the best k, we could select each individual in proportion to its relative fitness to the population Slower to converge, but could lose good solutions 15 Tournament selection 1. For each individual i, create a subset q of the population by random selection 2. Assign a point to i for each individual in q that it beats (is less fit than itself) 3. Rebuild the population based not on fitness scores, but the points accumulated in the tournament As the size of q increases, though, it becomes more like deterministic selection 16 Reproduction Reproduction The unique thing about GAs is the ability of solutions to inherit properties from other solutions in the population The basic way to perform a crossover operation is to splice together parts of the state description from each parent for example, in SAT: 1 0 0 1 1 1 0 1 0 1 0 0 0 1 1 0 1 0 0 1 0 1 1 0 0 1 0 0 1 1 0 1 There are many different ways to choose crossover point(s) for reproduction: Single-point: choose the center, or some optimal point in the state description take the first half of one parent, the second half of the other Random: choose the split point randomly (or proportional to the parents fitness scores) n-point: make not 1 split point, but n different ones Uniform: choose each element of the state description independently, at random (or proportional to fitness) parents children 17 18 3

Reproduction for 8-Queens Reproduction for TSP For the 8-queens problem, we could choose a crossover point with the same number of queens on either side What else could we do? For TSP, our individuals are n-city tours (e.g. 1-4-6-2-5-3 or 3-5-1-2-4-6 for 6 cities) Can we do a simple point-crossover like we could for SAT and 8-Queens? 1-4-6-2-5-3 3-5-1-2-4-6 parents 1-4-6-2-4-6 3-5-1-2-5-3 children These aren t } legal tours! 19 20 Reproduction for TSP Reproduction for TSP One option is to have an ordered master queue of all the cities in the problem: e.g. 1,2,3,4,5,6 Each individual, then, is a code that corresponds to dequeuing instructions from this master queue Numbers in the code are the relative position of the appropriate city left in the queue These codes can then be mated with point-crossovers actual parents 1-4-6-2-5-3 3-5-2-1-4-6 GA parent codes 1 3 4 1 2 1 3 4 1 1 1 1 GA child codes 1 3 4 1 1 1 3 4 1 1 2 1 actual children 1-4-6-2-3-5 3-5-2-1-6-4 21 We can try something even simpler to try and conserve information from both parents Pick an block of contiguous cities in the tour to pass from one parent to a child Remove all the cities in block from the other parent Add the remaining cities to the child in their preserved order, after the other block 1-4-6-2-5-3 3-5-2-1-4-6 conserve blocks remove cities from other parent 4-6-2-x-x-x 5-2-1-x-x-x 1-4-6-2-5-3 3-5-2-1-4-6 cross over the remainders 4-6-2-3-5-1 5-2-1-4-6-3 22 Mutation GAs and Creativity There are also a variety of ways to mutate individuals in the population The first question to consider it who to mutate Alter the most fit? Least fit? Random? Mutate children only, or surviving parents as well? How many to mutate? The second question is how to mutate Totally arbitrarily? Mutate to a better neighbor? GAs can be thought of as a simultaneous, parallel hill-climbing search The population as a whole is trying to converge to an optimal solution Because solutions can evolve from a variety of factors, without prodding from us as to which direction to go (as in local search), very novel problem solutions can be found discovered 23 24 4

GAs and Creativity GAs and Creativity Sensor-actuator networks (SANs) are structures that model connections between sensors and actuators (motors and muscles) in simple robots GAs can learn parameters for SANs that solve locomotion problems in a variety of ways M. van de Panne, Control Techniques For Physically-Based Animation, Third Eurographics Workshop on Animation and Simulation, 1994 25 26 So far, we ve talked about GAs as a search strategy for a problem solving (in which case, there is an agent conducting the GA search) Recall from the second lecture about multi-agent environments Now consider a GA that evolves a population of agents!! Now, our GA population is a virtual multi-agent environment 27 Let s say we want to grow agents to predict stock market trends Each agent might be some statistical function that maps a stock s history to its predicted future performance: (α todaysprice) + (β yesterdaysprice) + (γ relativevalue) + (δ 1-monthStdDev) + 28 We could maintain a population of these agents, where each agent s state ( DNA ) is its set of coefficients (α, β, γ, δ, etc.) Over time, the population should converge on a population of individuals that reflect the current stock market trends Now let s think about a GA for them: What is our fitness function? What is a good crossover? How can we mutate them? But will the most fit individual become a universally good stock-picker? Perhaps not! 29 30 5

It is possible that different agents in the population specialize to aspect of the task Some agents predict well for the Fortune-500 stocks Others predict well for sports companies Still others pick non-profits well, etc If there is no one universally intelligent agent in the population, perhaps we can let them all predict or vote on a predictions So our population of stock-pickers is a multi-agent environment is it cooperative or competitive? This phenomenon is what we might refer to as emergent intelligence No single agent in the environment is, taken individually, all that intelligent Taken together, though, the entire population possesses a more global intelligence that emerges from its constituent agents 31 32 Genetic programming is a field related to genetic algorithms (surprise!) Expression trees are graphical, relational representations of functions or programs Instead of maintaining and manipulating a population of strings, however, we use expression trees the goal is to evolve programs Section 9.5 in Machine Learning provides a nice overview of this (slightly more advanced) topic 33 Programming language compilers convert code to such trees before writing out machine-level instructions (CS 536) For example: sin(x) + x 2 + y sin + x + x ^ 2 y 34 Populations are initialized with randomized, wellformed expressions build up from: Operators (e.g. +, sin,, etc.) Terminals (x, y, 2, etc.) sin(x) + x 2 + y + sin x + ^ x 2 y parents + sin(x) + 2 x+y sin ^ x 2 + x y Fitness is evaluated on how well its encoded function/algorithm performs the task sin(x) + 2 x2 + + sin(x) + (x + y) + y Crossover is applied by swapping subtrees in the parent expressions 35 x sin 2 ^ ^ x 2 children sin x + + x y y 36 6

Views of Evolution J. Koza, Genetic programming: On the programming of computers by means of natural selection, MIT Press, 1992 300 random programs were initialized with primitives to solve block-stacking problems with the goal of spelling UNIVERSAL After 10 generations, a program evolved that solved all of 166 initial configurations The Lamarkian Theory Popular in 1800s An individual s genetic makeup is altered as it learns through life experience Today s biological evidence contradicts this 37 38 Views of Evolution Last Thoughts on GAs The Baldwin Effect Learning has no effect on genetic makeup However, ability to learn reduces the need for hardwired functions Therefore, individual learning allows for a more diverse gene pool (less hard-wiring) and more adaptable populations Example: a new predator appears in an environment; individuals who can learn to avoid it live on, resulting in more adaptable gene pool 39 Evolutionary algorithms are simulations of what we perceive happening in nature, but we don t have to follow the laws of nature Lamarkian GAs have been experimented with, and shown successful on some problems Since we get to design the framework for the simulation, there is a wide margin for creative license in the framework we create! Concepts of age/gender/politics? Variety of fitness functions? 40 Partial Search Look through state space for a goal from which a solution can be found Node: state description Edge: action that changes state at some cost Path: sequence of actions that change from the start state to the goal state 41 Uninformed Search: no domain information Complete/optimal if costs uniform: BFS, IDS Complete/optimal with costs: UCS Not complete/optimal: DFS, DLS Informed Search: use heuristics to guide search g(n): cost from start to n h(n): estimates cost from n to goal (heuristic) f(n): g(n) + h(n): estimated cost of searching through n Complete/optimal: A* = h(n) is admissible Not complete/optimal: Greedy, A 42 7

Optimization Search Look through solution space for better solutions to the problem Node: complete solution Edge: operator changes to a new solution Can stop anytime Well-suited for NP-Complete problems, optimization problems Local Search Focus on a local part of the search space rather than exploring it all Beam search limits the list of candidate states Hill-climbing follows a single path of promising successor states Solution heavily dependent on the initial state Can get stuck in local optima 43 44 Escaping Local Optima Ways to avoid the traps into which local search methods tend to fall Random Restarting Simulated Annealing Tabu Search Evolutionary Search Unique, non-local, parallel optimization search 45 8