Outline. Learning from Observations. Learning agents. Learning. Inductive learning (a.k.a. Science) Environment. Agent.

Similar documents
Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Lecture 1: Machine Learning Basics

Proof Theory for Syntacticians

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

MYCIN. The MYCIN Task

CS Machine Learning

Probability and Statistics Curriculum Pacing Guide

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Lecture 10: Reinforcement Learning

Rule Learning With Negation: Issues Regarding Effectiveness

Radius STEM Readiness TM

A Version Space Approach to Learning Context-free Grammars

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

TU-E2090 Research Assignment in Operations Management and Services

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Rule Learning with Negation: Issues Regarding Effectiveness

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Grade 6: Correlated to AGS Basic Math Skills

Chapter 2 Rule Learning in a Nutshell

A Case Study: News Classification Based on Term Frequency

Quantitative analysis with statistics (and ponies) (Some slides, pony-based examples from Blase Ur)

Mathematics Success Grade 7

Cal s Dinner Card Deals

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

IMGD Technical Game Development I: Iterative Development Techniques. by Robert W. Lindeman

Axiom 2013 Team Description Paper

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

DegreeWorks Advisor Reference Guide

SANTIAGO CANYON COLLEGE Reading & English Placement Testing Information

How to set up gradebook categories in Moodle 2.

Intelligent Agents. Chapter 2. Chapter 2 1

re An Interactive web based tool for sorting textbook images prior to adaptation to accessible format: Year 1 Final Report

Pod Assignment Guide

Generating Test Cases From Use Cases

Using dialogue context to improve parsing performance in dialogue systems

Millersville University Degree Works Training User Guide

On the Combined Behavior of Autonomous Resource Management Agents

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Agents and environments. Intelligent Agents. Reminders. Vacuum-cleaner world. Outline. A vacuum-cleaner agent. Chapter 2 Actuators

STA 225: Introductory Statistics (CT)

Physics 270: Experimental Physics

Copyright 2017 DataWORKS Educational Research. All rights reserved.

Language properties and Grammar of Parallel and Series Parallel Languages

Setting Up Tuition Controls, Criteria, Equations, and Waivers

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

RANKING AND UNRANKING LEFT SZILARD LANGUAGES. Erkki Mäkinen DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF TAMPERE REPORT A ER E P S I M S

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Word learning as Bayesian inference

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

(Sub)Gradient Descent

Scientific Method Investigation of Plant Seed Germination

Improving Simple Bayes. Abstract. The simple Bayesian classier (SBC), sometimes called

12- A whirlwind tour of statistics

Eduroam Support Clinics What are they?

Multimedia Application Effective Support of Education

On the Polynomial Degree of Minterm-Cyclic Functions

Action Models and their Induction

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Lecture 1: Basic Concepts of Machine Learning

ELA Grade 4 Literary Heroes Technology Integration Unit

On-Line Data Analytics

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Speech Recognition at ICSI: Broadcast News and beyond

Planning with External Events

A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and

CSL465/603 - Machine Learning

Student Handbook. This handbook was written for the students and participants of the MPI Training Site.

FCE Speaking Part 4 Discussion teacher s notes

Language Acquisition Chart

A Case-Based Approach To Imitation Learning in Robotic Agents

Firms and Markets Saturdays Summer I 2014

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Learning Methods in Multilingual Speech Recognition

Emporia State University Degree Works Training User Guide Advisor

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Reinforcement Learning by Comparing Immediate Reward

The lab is designed to remind you how to work with scientific data (including dealing with uncertainty) and to review experimental design.

The Strong Minimalist Thesis and Bounded Optimality

Tabular and Textual Methods for Selecting Objects from a Group

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

"f TOPIC =T COMP COMP... OBJ

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Generative models and adversarial training

Ph.D. in Behavior Analysis Ph.d. i atferdsanalyse

B. Outcome Reporting Include the following information for each outcome assessed this year:

Lab 1 - The Scientific Method

Individual Differences & Item Effects: How to test them, & how to test them well

BMBF Project ROBUKOM: Robust Communication Networks

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

A Neural Network GUI Tested on Text-To-Phoneme Mapping

How to analyze visual narratives: A tutorial in Visual Narrative Grammar

Python Machine Learning

Transcription:

Outline Learning agents Learning from Observations Inductive learning Decision tree learning Measuring learning performance Chapter 18, Sections 1 3 Chapter 18, Sections 1 3 1 Chapter 18, Sections 1 3 2 Learning Learning is essential for unknown environments, i.e., when designer lacks omniscience Learning is useful as a system construction method, i.e., epose the agent to reality rather than trying to write it down Learning modifies the agent s decision mechanisms to improve performance Performance standard feedback Agent learning goals Critic Learning element Problem generator Learning agents changes knowledge eperiments Sensors Performance element Effectors Environment Chapter 18, Sections 1 3 3 Chapter 18, Sections 1 3 4 Learning element Design of learning element is dictated by what type of performance element is used which functional component is to be learned how that functional compoent is represented what kind of feedback is available Eample scenarios: Performance element Component Representation Alpha beta search Logical agent Utility based agent Simple refle agent Eval. fn. ransition model ransition model Percept action fn Weighted linear function Successor state aioms Dynamic Bayes net Neural net Supervised learning: correct answers for each instance Reinforcement learning: occasional rewards eedback Win/loss Outcome Outcome Correct action Inductive learning (a.k.a. Science) Simplest form: learn a function from eamples (tabula rasa) f is the target function An eample is a pair, f(), e.g., O O X X X, +1 Problem: find a(n) hypothesis h such that h f given a training set of eamples (his is a highly simplified model of real learning: Ignores prior knowledge Assumes a deterministic, observable environment Assumes eamples are given Assumes that the agent wants to learn f why?) Chapter 18, Sections 1 3 5 Chapter 18, Sections 1 3 6

(h is consistent if it agrees with f on all eamples) f() (h is consistent if it agrees with f on all eamples) f() Chapter 18, Sections 1 3 7 Chapter 18, Sections 1 3 8 (h is consistent if it agrees with f on all eamples) f() (h is consistent if it agrees with f on all eamples) f() Chapter 18, Sections 1 3 9 Chapter 18, Sections 1 3 10 (h is consistent if it agrees with f on all eamples) f() (h is consistent if it agrees with f on all eamples) f() Ockham s razor: maimize a combination of consistency and simplicity Chapter 18, Sections 1 3 11 Chapter 18, Sections 1 3 12

Attribute-based representations Eamples described by attribute values (Boolean, discrete, continuous, etc.) E.g., situations where I will/won t wait for a table: Eample Attributes arget Alt Bar ri Hun Pat Price Rain Res ype Est WillWait X 1 Some $$$ rench 0 10 X 2 ull $ hai 30 60 X 3 Some $ Burger 0 10 X 4 ull $ hai 10 30 X 5 ull $$$ rench >60 X 6 Some $$ Italian 0 10 X 7 None $ Burger 0 10 X 8 Some $$ hai 0 10 X 9 ull $ Burger >60 X 10 ull $$$ Italian 10 30 X 11 None $ hai 0 10 X 12 ull $ Burger 30 60 Classification of eamples is positive () or negative () Decision trees One possible representation for hypotheses E.g., here is the true tree for deciding whether to wait: None Some ull Reservation? ri/sat? Bar? WaitEstimate? >60 30 60 10 30 0 10 Alternate? Hungry? Alternate? Raining? Chapter 18, Sections 1 3 13 Chapter 18, Sections 1 3 14 Epressiveness Decision trees can epress any function of the input attributes. E.g., for Boolean functions, truth table row path to leaf: A B A or B B A B rivially, there is a consistent decision tree for any training set w/ one path to leaf for each eample (unless f nondeterministic in ) but it probably won t generalize to new eamples Prefer to find more compact decision trees Chapter 18, Sections 1 3 15 Chapter 18, Sections 1 3 16 = number of distinct truth tables with 2 n rows Chapter 18, Sections 1 3 17 Chapter 18, Sections 1 3 18

Chapter 18, Sections 1 3 19 Chapter 18, Sections 1 3 20 How many purely conjunctive hypotheses (e.g., Hungry Rain)?? How many purely conjunctive hypotheses (e.g., Hungry Rain)?? Each attribute can be in (positive), in (negative), or out 3 n distinct conjunctive hypotheses More epressive hypothesis space increases chance that target function can be epressed increases number of hypotheses consistent w/ training set may get worse predictions Chapter 18, Sections 1 3 21 Chapter 18, Sections 1 3 22 Decision tree learning Aim: find a small tree consistent with the training eamples Idea: (recursively) choose most significant attribute as root of (sub)tree function DL(eamples, attributes, default) returns a decision tree if eamples is empty then return default else if all eamples have the same classification then return the classification else if attributes is empty then return Mode(eamples) else best Choose-Attribute(attributes, eamples) tree a new decision tree with root test best for each value v i of best do eamples i {elements of eamples with best = v i} subtree DL(eamples i, attributes best, Mode(eamples)) add a branch to tree with label v i and subtree subtree return tree Choosing an attribute Idea: a good attribute splits the eamples into subsets that are (ideally) all positive or all negative None Some ull ype? rench Italian hai Burger P atrons? is a better choice gives information about the classification Chapter 18, Sections 1 3 23 Chapter 18, Sections 1 3 24

Information Information answers questions he more clueless I am about the answer initially, the more information is contained in the answer Scale: 1 bit = answer to Boolean question with prior 0.5, 0.5 Informationinananswerwhenprioris P 1,...,P n is H( P 1,...,P n )=Σ n i =1 P i log 2 P i (also called entropy of the prior) Information contd. Suppose we have p positive and n negative eamples at the root H( p/(p+n),n/(p+n) ) bits needed to classify a new eample E.g., for 12 restaurant eamples, p = n =6 so we need 1 bit An attribute splits the eamples E into subsets E i, each of which (we hope) needs less information to complete the classification Let E i have p i positive and n i negative eamples H( p i /(p i +n i ),n i /(p i +n i ) ) bits needed to classify a new eample epected number of bits per eample over all branches is p Σ i + n i i p + n H( p i/(p i + n i ),n i /(p i + n i ) ) or P atrons?, this is 0.459 bits, for ype this is (still) 1 bit choose the attribute that minimizes the remaining information needed Chapter 18, Sections 1 3 25 Chapter 18, Sections 1 3 26 Eample contd. Decision tree learned from the 12 eamples: None Some ull Hungry? Yes No ype? rench Italian hai Burger ri/sat? Substantially simpler than true tree a more comple hypothesis isn t justified by small amount of data Performance measurement Howdoweknowthath f? (Hume sproblem of Induction) 1) Use theorems of computational/statistical learning theory 2) ry h on a new test set of eamples (use same distribution over eample space as training set) Learning curve = % correct on test set as a function of training set size 1 % correct on test set 0.9 0.8 0.7 0.6 0.5 0.4 0 10 20 30 40 50 60 70 80 90 100 raining set size Chapter 18, Sections 1 3 27 Chapter 18, Sections 1 3 28 Performance measurement contd. Learning curve depends on realizable (can epress target function) vs. non-realizable non-realizability can be due to missing attributes or restricted hypothesis class (e.g., thresholded linear function) redundant epressiveness (e.g., loads of irrelevant attributes) % correct 1 realizable redundant nonrealizable Summary Learning needed for unknown environments, lazy designers Learning agent = performance element + learning element Learning method depends on type of performance element, available feedback, type of component to be improved, and its representation or supervised learning, the aim is to find a simple hypothesis that is approimately consistent with training eamples Decision tree learning using information gain Learning performance = prediction accuracy measured on test set # of eamples Chapter 18, Sections 1 3 29 Chapter 18, Sections 1 3 30