Jim Tørresen Chapter 9: Working with Evolutionary Algorithms. INF Biologically inspired computing Lecture 4: Eiben and Smith,

Similar documents
Artificial Neural Networks written examination

CS Machine Learning

Laboratorio di Intelligenza Artificiale e Robotica

Laboratorio di Intelligenza Artificiale e Robotica

Learning From the Past with Experiment Databases

Tun your everyday simulation activity into research

Evolutive Neural Net Fuzzy Filtering: Basic Description

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Lecture 1: Machine Learning Basics

TABLE OF CONTENTS TABLE OF CONTENTS COVER PAGE HALAMAN PENGESAHAN PERNYATAAN NASKAH SOAL TUGAS AKHIR ACKNOWLEDGEMENT FOREWORD

On-the-Fly Customization of Automated Essay Scoring

Probability and Statistics Curriculum Pacing Guide

Discriminative Learning of Beam-Search Heuristics for Planning

Major Milestones, Team Activities, and Individual Deliverables

A simulated annealing and hill-climbing algorithm for the traveling tournament problem

An empirical study of learning speed in backpropagation

(Sub)Gradient Descent

Reinforcement Learning by Comparing Immediate Reward

Software Maintenance

On the Combined Behavior of Autonomous Resource Management Agents

Solving Combinatorial Optimization Problems Using Genetic Algorithms and Ant Colony Optimization

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Learning and Transferring Relational Instance-Based Policies

Diagnostic Test. Middle School Mathematics

Softprop: Softmax Neural Network Backpropagation Learning

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

TD(λ) and Q-Learning Based Ludo Players

Investigating Ahuja-Orlin s Large Neighbourhood Search Approach for Examination Timetabling

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

The Strong Minimalist Thesis and Bounded Optimality

Seminar - Organic Computing

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Lecture 2: Quantifiers and Approximation

While you are waiting... socrative.com, room number SIMLANG2016

An Introduction to Simio for Beginners

The Evolution of Random Phenomena

DIANA: A computer-supported heterogeneous grouping system for teachers to conduct successful small learning groups

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

Research Article Hybrid Multistarting GA-Tabu Search Method for the Placement of BtB Converters for Korean Metropolitan Ring Grid

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Ordered Incremental Training with Genetic Algorithms

Statistical Studies: Analyzing Data III.B Student Activity Sheet 7: Using Technology

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

ECE-492 SENIOR ADVANCED DESIGN PROJECT

Cooperative evolutive concept learning: an empirical study

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

Assignment 1: Predicting Amazon Review Ratings

Modeling user preferences and norms in context-aware systems

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

The dilemma of Saussurean communication

Evolution of Symbolisation in Chimpanzees and Neural Nets

Degeneracy results in canalisation of language structure: A computational model of word learning

LEGO MINDSTORMS Education EV3 Coding Activities

BMBF Project ROBUKOM: Robust Communication Networks

Managerial Decision Making

Alpha provides an overall measure of the internal reliability of the test. The Coefficient Alphas for the STEP are:

Evidence for Reliability, Validity and Learning Effectiveness

OVERVIEW OF CURRICULUM-BASED MEASUREMENT AS A GENERAL OUTCOME MEASURE

Chapter 2 Rule Learning in a Nutshell

School Inspection in Hesse/Germany

Rover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes

Cognitive Thinking Style Sample Report

Shockwheat. Statistics 1, Activity 1

Development of Multistage Tests based on Teacher Ratings

Learning to Schedule Straight-Line Code

Leveraging MOOCs to bring entrepreneurship and innovation to everyone on campus

Executive Guide to Simulation for Health

Introduction to Simulation

Redirected Inbound Call Sampling An Example of Fit for Purpose Non-probability Sample Design

Knowledge-Based - Systems

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Speeding Up Reinforcement Learning with Behavior Transfer

Why Did My Detector Do That?!

Conference Presentation

Certified Six Sigma Professionals International Certification Courses in Six Sigma Green Belt

Universityy. The content of

GACE Computer Science Assessment Test at a Glance

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

IMGD Technical Game Development I: Iterative Development Techniques. by Robert W. Lindeman

Acquiring Competence from Performance Data

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations

Student Transportation

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Essentials of Ability Testing. Joni Lakin Assistant Professor Educational Foundations, Leadership, and Technology

A Case Study: News Classification Based on Term Frequency

How to Judge the Quality of an Objective Classroom Test

AUTHOR COPY. Techniques for cold-starting context-aware mobile recommender systems for tourism

Learning Methods for Fuzzy Systems

ABSTRACT. A major goal of human genetics is the discovery and validation of genetic polymorphisms

Rule Learning With Negation: Issues Regarding Effectiveness

ABET Criteria for Accrediting Computer Science Programs

A Pipelined Approach for Iterative Software Process Model

Institutionen för datavetenskap. Hardware test equipment utilization measurement

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

A Comparison of Charter Schools and Traditional Public Schools in Idaho

Transcription:

Chapter 9: Working with Evolutionary Algorithms INF3490 - Biologically inspired computing Lecture 4: Eiben and Smith, Working with evolutionary algorithms (chpt 9) Hybrid algorithms (chpt 10) Multi-objective optimization (chpt 12) Jim Tørresen 1. Experiment design 2. Algorithm design 3. Test problems 4. Measurements and statistics 5. Some tips and summary 2 Experimentation Has a goal or goals Involves algorithm design and implementation Needs problem(s) to run the algorithm(s) on Amounts to running the algorithm(s) on the problem(s) Delivers measurement data, the results Is concluded with evaluating the results in the light of the given goal(s) Is often documented (thesis, papers, web, ) Experimentation: Goals for Research Show that EC is applicable in a (new) problem domain (real-world applications) Show that my_ea is better than benchmark_ea Show that EAs outperform traditional algorithms Optimize or study impact of parameters on the performance of an EA Investigate algorithm behavior (e.g. interaction between selection and variation) See how an EA scales-up with problem size 3 4 1

Example: Repetitive Problems Optimising Internet shopping delivery route Need to run regularly/repetitively Different destinations each day Limited time to run algorithm each day Must always be reasonably good route in limited time Example: Design Problems Optimising spending on improvements to national road network Total cost: billions of Euro Computing costs negligible Six months to run algorithm on hundreds computers Many runs possible Must produce very good result just once 5 6 Algorithm design Design a representation Design a way of mapping a genotype to a phenotype Design a way of evaluating an individual Design suitable mutation operator(s) Design suitable recombination operator(s) Decide how to select individuals to be parents Decide how to select individuals for the next generation (how to manage the population) Decide how to start: initialization method Decide how to stop: termination criterion Test problems 1. Recognized benchmark problem repository (typically challenging ) 2. Problem instances made by random generator 3. Frequently encountered or otherwise important variants of given real-world problems Choice has severe implications on: generalizability and scope of the results 7 8 2

Getting Problem Instances (1/3) Benchmarks Standard data sets in problem repositories, e.g.: OR-Library www.brunel.ac.uk/~mastjjb/jeb/info.html UCI Machine Learning Repository www.ics.uci.edu/~mlearn/mlrepository.html Advantage: Well-chosen problems and instances (hopefully) Much other work on these à results comparable Disadvantage: Not real might miss crucial aspect Algorithms get tuned for popular test suites 9 Getting Problem Instances (2/3) Problem instance generators Problem instance generators produce simulated data for given parameters, e.g.: GA/EA Repository of Test Problem Generators http://vlsicad.eecs.umich.edu/bk/slots/cache/www.cs.uwyo.edu/~wspears/ generators.html Advantage: Allow very systematic comparisons for they can produce many instances with the same characteristics enable gradual traversal of a range of characteristics (hardness) Can be shared allowing comparisons with other researchers Disadvantage Not real might miss crucial aspect Given generator might have hidden bias 10 Getting Problem Instances (3/3) Real-world problems Typical Results from Several EA Runs Testing on (own collected) real data Advantages: Results could be considered as very relevant viewed from the application domain (data supplier) Disadvantages Can be over-complicated Can be few available sets of real data May be commercial sensitive difficult to publish and to allow others to compare Results are hard to generalize Fitness/ Performance 1 2 3 4 5 N Run # 11 12 3

Basic rules of experimentation EAs are stochastic! never draw any conclusion from a single run perform sufficient number of independent runs use statistical measures (averages, standard deviations) use statistical tests to assess reliability of conclusions EA experimentation is about comparison! always do a fair competition use the same amount of resources for the competitors try different comp. limits (to cope with turtle/hare effect) use the same performance measures Things to Measure Many different ways. Examples: Average result in given time Average time for given result Proportion of runs within % of target Best result over n runs Amount of computing required to reach target in given time with % confidence 13 14 What time units do we use? Elapsed time? Depends on computer, network, etc CPU Time? Depends on skill of programmer, implementation, etc Generations? Incomparable when parameters like population size change Evaluations? Evaluation time could depend on algorithm, e.g. direct vs. indirect representation Evaluation time could be small compared to other steps in the EA (e.g. genotype to phenotype translation) 15 Measures Performance measures (off-line) Efficiency (alg. speed, also called performance) Execution time Average no. of evaluations to solution (AES, i.e., number of generated points in the search space) Effectiveness (solution quality, also called accuracy) Success rate (SR): % of runs finding a solution Mean best fitness at termination (MBF) Working measures (on-line) Population distribution (genotypic) Fitness distribution (phenotypic) Improvements per time unit or per genetic operator 16 4

Example: off-line performance measure evaluation Example: on-line performance measure evaluation 30 Populations mean (best) fitness Which algorithm is better? Why? When? Nr. of runs ending with this fitness 25 20 15 10 5 Algorithm A Algorithm B 0-50 51-60 61-70 71-80 Best fitness at termination 81-90 91-100 Alg B Alg A 17 Which algorithm is better? Why? When? 18 Example: averaging on-line measures Example: overlaying on-line measures Averaging can choke interesting information Overlay of curves can lead to very cloudy figures 19 20 5

Statistical Comparisons and Significance Algorithms are stochastic, results have element of luck If a claim is made Mutation A is better than mutation B, need to show statistical significance of comparisons Fundamental problem: two series of samples (random drawings) from the SAME distribution may have DIFFERENT averages and standard deviations Tests can show if the differences are significant or not 21 Example Trial Old Method New Method 1 500 657 2 600 543 3 556 654 4 573 565 5 420 654 6 590 712 7 700 456 8 472 564 9 534 675 10 512 643 Average 545.7 612.3 Is the new method better? 22 Example (cont d) Standard deviations supply additional info T-test (and alike) indicate the chance that the values came from the same underlying distribution (difference is due to random effects) E.g. with 7% chance in this example. 23 Summary of tips for experiments Be organized Decide what you want & define appropriate measures Choose test problems carefully Make an experiment plan (estimate time when possible) Perform sufficient number of runs Keep all experimental data (never throw away anything) Include in publications all necessary parameters to make others able to repeat your experiments Use good statistics ( standard tools from Web, MS, R) Present results well (figures, graphs, tables, ) Watch the scope of your claims Aim at generalizable results (use separate data set for training and testing) Publish code for reproducibility of results (if applicable) 24 Publish data for external validation (open science) 6

Chapter 10: Hybridisation with Other Techniques: Memetic Algorithms 1. Why to Hybridise 2. What is a Memetic Algorithm? 3. Where to hybridise 4. Local Search Lamarckian vs. Baldwinian adaptation 1. Why Hybridise Might be looking at improving on existing techniques (non-ea) Might be looking at improving EA search for good solutions 25 26 1. Why Hybridise Michalewicz s view on EAs in context 2. What is a Memetic Algorithm? The combination of Evolutionary Algorithms with Local Search Operators that work within the EA loop has been termed Memetic Algorithms Term also applies to EAs that use instancespecific knowledge Memetic Algorithms have been shown to be orders of magnitude faster and more accurate than EAs on some problems, and are the state of the art on many problems 27 28 7

3. Where to Hybridise: 3. Where to Hybridise: In initialization Seeding Known good solutions are added Selective initialization Generate kn solutions, keep best N Refined start Perform local search on initial population 29 30 3. Where to Hybridise: Intelligent mutation and crossover Mutation bias Mutation operator has bias towards certain changes Crossover hill-climber Test all 1-point crossover results, choose best 4. Local Search: Local Search Defined by combination of neighbourhood and pivot rule Related to landscape metaphor N(x) is defined as the set of points that can be reached from x with one application of a move operator e.g. bit flipping search on binary problems Repair mutation Use heuristic to make infeasible solution feasible c g d h N(d) = {a,c,h} 31 b f a e 32 8

4. Local Search: Pivot Rules Is the neighbourhood searched randomly, systematically or exhaustively? does the search stop as soon as a fitter neighbour is found (Greedy Ascent) or is the whole set of neighbours examined and the best chosen (Steepest Ascent) of course there is no one best answer, but some are quicker than others to run... 4. Local Search and Evolution Do offspring inherit what their parents have learnt in life? Yes - Lamarckian learning Improved fitness and genotype No - Baldwinian learning: Improved fitness only 33 34 4. Local Search: Induced landscapes Hybrid Algorithms Summary Raw Fitness Lamarckian points Baldwin landscape It is common practice to hybridise EA s when using them in a real world context. This may involve the use of operators from other algorithms which have already been used on the problem, or the incorporation of domain-specific knowledge Memetic algorithms have been shown to be orders of magnitude faster and more accurate than EAs on some problems, and are the state of the art on many problems 35 36 9

Chapter 12: Multiobjective Evolutionary Algorithms Multiobjective optimisation problems (MOP) - Pareto optimality EC approaches - Evolutionary spaces - Preserving diversity Multi-Objective Problems (MOPs) Wide range of problems can be categorised by the presence of a number of n possibly conflicting objectives: buying a car: speed vs. price vs. reliability engineering design: lightness vs. strength Two problems: finding set of good solutions choice of best for the particular application 37 38 An example: Buying a car Two approaches to multiobjective optimisation Weighted sum (scalarisation): speed transform into a single objective optimisation method compute a weighted sum of the different objectives A set of multi-objective solutions (Pareto front): cost The population-based nature of EAs used to simultaneously search for a set of points approximating Pareto front 39 40 10

Comparing solutions Dominance relation Objective space Optimisation task: Minimize both f 1 and f 2 Then: a is better than b a is better than c a is worse than e a and d are incomparable Solution x dominates solution y, (x y), if: x is better than y in at least one objective, x is not worse than y in all other objectives solutions dominating x solutions dominated by x 41 42 Pareto optimality Solution x is non-dominated among a set of solutions Q if no solution from Q dominates x A set of non-dominated solutions from the entire feasible solution space is the Pareto-optimal set, its members Pareto-optimal solutions Illustration of the concepts f 2 (x) min Pareto-optimal front: an image of the Pareto-optimal set in the objective space 43 f 1 (x) min 44 11

Illustration of the concepts A practical example: The beam design problem f 2 (x) min Minimize weight and deflection of a beam (Deb, 2001): d f 1 (x) min 45 46 Formal definition Feasible solutions Minimize minimize subject to where 2 πd f1( d, l) = ρ l 4 3 64Pl f2( d, l) = δ = 4 3Eπ d 0.01 m d 0.05 m 0.2 m l 1.0 m 32Pl σ max = S 3 π d δ δ max ρ = = 3 7800 kg/m, P 2 kn E = 207 GPa S = 300 MPa, δ = 0.005 m y y max (beam weight) (beam deflection) (maximum stress) 47 Decision (variable) space Objective space 48 12

Goal: Finding non-dominated solutions Goal of multiobjective optimisers Find a set of non-dominated solutions (approximation set) following the criteria of: convergence (as close as possible to the Paretooptimal front), diversity (spread, distribution) 49 50 EC approach: Requirements 1. Way of assigning fitness, usually based on dominance 2. Preservation of a diverse set of points similarities to multi-modal problems 3. Remembering all the non-dominated points you have seen usually using elitism or an archive 51 EC approach: 1. Fitness Assignment Could use aggregating approach and change weights during evolution no guarantees Different parts of population use different criteria no guarantee of diversity Dominance (made a breakthrough for MOEA) ranking or depth based fitness related to whole population 52 13

EC approach: 2. Diversity maintenance Usually done by niching techniques such as: fitness sharing adding amount to fitness based on inverse distance to nearest neighbour (minimisation) (adaptively) dividing search space into boxes and counting occupancy All rely on some distance metric in genotype / phenotype space EC approach: 3. Remembering Good Points Could just use elitist algorithm, e.g. ( µ + λ ) replacement crowding distance Common to maintain an archive of nondominated points some algorithms use this as a second population that can be in recombination etc. others divide archive into regions too, e.g. PAES 53 54 Multi objective problems - Summary MO problems occur very frequently EAs are very good in solving MO problems MOEAs are one of the most successful EC subareas 55 14