Introduction. Notices. A Learning Agent 22/11/2012. COMP219: Artificial Intelligence. COMP219: Artificial Intelligence

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

CS Machine Learning

Lecture 1: Basic Concepts of Machine Learning

Python Machine Learning

Lecture 1: Machine Learning Basics

Rule Learning With Negation: Issues Regarding Effectiveness

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

CSL465/603 - Machine Learning

Exploration. CS : Deep Reinforcement Learning Sergey Levine

MYCIN. The MYCIN Task

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Physics 270: Experimental Physics

Grade 6: Correlated to AGS Basic Math Skills

Rule Learning with Negation: Issues Regarding Effectiveness

Probability and Statistics Curriculum Pacing Guide

Agents and environments. Intelligent Agents. Reminders. Vacuum-cleaner world. Outline. A vacuum-cleaner agent. Chapter 2 Actuators

University of Groningen. Systemen, planning, netwerken Bosman, Aart

A Case Study: News Classification Based on Term Frequency

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Intelligent Agents. Chapter 2. Chapter 2 1

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

STA 225: Introductory Statistics (CT)

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

Learning Methods in Multilingual Speech Recognition

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Lecture 10: Reinforcement Learning

Axiom 2013 Team Description Paper

Learning goal-oriented strategies in problem solving

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Word Segmentation of Off-line Handwritten Documents

STT 231 Test 1. Fill in the Letter of Your Choice to Each Question in the Scantron. Each question is worth 2 point.

Learning From the Past with Experiment Databases

LEGO MINDSTORMS Education EV3 Coding Activities

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Universidade do Minho Escola de Engenharia

12- A whirlwind tour of statistics

Winter School, February 1 to 5, 2016 Schedule. Ronald Schlegel, December 10, 2015

A Reinforcement Learning Variant for Control Scheduling

Artificial Neural Networks written examination

Chapters 1-5 Cumulative Assessment AP Statistics November 2008 Gillespie, Block 4

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

Speech Recognition at ICSI: Broadcast News and beyond

CS 446: Machine Learning

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Generative models and adversarial training

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Using focal point learning to improve human machine tacit coordination

Abstractions and the Brain

Assignment 1: Predicting Amazon Review Ratings

(Sub)Gradient Descent

Reinforcement Learning by Comparing Immediate Reward

Proof Theory for Syntacticians

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Using dialogue context to improve parsing performance in dialogue systems

Linking Task: Identifying authors and book titles in verbose queries

Stacks Teacher notes. Activity description. Suitability. Time. AMP resources. Equipment. Key mathematical language. Key processes

A Version Space Approach to Learning Context-free Grammars

Chapter 2 Rule Learning in a Nutshell

While you are waiting... socrative.com, room number SIMLANG2016

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Mathematics subject curriculum

Radius STEM Readiness TM

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Measures of the Location of the Data

Mathematics process categories

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

TU-E2090 Research Assignment in Operations Management and Services

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

DOCTORAL SCHOOL TRAINING AND DEVELOPMENT PROGRAMME

Learning Methods for Fuzzy Systems

AQUA: An Ontology-Driven Question Answering System

The lab is designed to remind you how to work with scientific data (including dealing with uncertainty) and to review experimental design.

Individual Component Checklist L I S T E N I N G. for use with ONE task ENGLISH VERSION

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and

CLASS EXODUS. The alumni giving rate has dropped 50 percent over the last 20 years. How can you rethink your value to graduates?

Laboratorio di Intelligenza Artificiale e Robotica

Learning and Transferring Relational Instance-Based Policies

Statistical Analysis of Climate Change, Renewable Energies, and Sustainability An Independent Investigation for Introduction to Statistics

Math Placement at Paci c Lutheran University

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Word learning as Bayesian inference

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Applications of data mining algorithms to analysis of medical data

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots

4 th Grade Number and Operations in Base Ten. Set 3. Daily Practice Items And Answer Keys

We are strong in research and particularly noted in software engineering, information security and privacy, and humane gaming.

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

1.11 I Know What Do You Know?

Results In. Planning Questions. Tony Frontier Five Levers to Improve Learning 1

SOFTWARE EVALUATION TOOL

Algebra 2- Semester 2 Review

Outreach Connect User Manual

Rule-based Expert Systems

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Transcription:

COMP219: Artificial Intelligence COMP219: Artificial Intelligence Dr. Annabel Latham Room 2.05 Ashton Building Department of Computer Science University of Liverpool Lecture 27: Introduction to Learning, Supervised Learning 1 2 Notices The schedule for COMP219 lectures next week is as follows: Tuesday 4 th December: Lecture as usual: SUMMARY & REVISION (the final lecture for the module) Thursday 6 th December: No lecture (- you can use this time for revision) Friday 7 th December: No lecture (- you can use this time for revision) Tuesday 11 th December: Class test 2 during the usual lecture slot at 15.00 No further sessions will take place after the test on Tuesday 11th December Introduction Last time Planning in the real world Scheduling with time and resource constraints Critical path method, minimum slack HTN Today Types of learning Supervised learning Learning decision trees 3 4 Why do we want an agent to learn? Cannot anticipate all situations unknown environments (e.g. navigating new space) Cannot predict changes over time (e.g. react to the stock market) Don t know how to design some solutions (e.g. recognising faces) or it s too time consuming to do so Learning modifies the agent s decision mechanisms to improve performance A Learning Agent Agent is learning if it improves its performance on future tasks after making observations about the world Any component of an agent can be improved by learning, but the choice of technique depends on: What the component is What prior knowledge the agent has How the data and component are represented What feedback is available to learn from 5 6 1

Learning Agents Example: Training a Taxi Driver Agent When the instructor shouts Brake the agent may learn a condition-action rule for when to brake; agent also learns when the instructor does not shout By seeing camera images which it is told are buses, the agent learns to recognise buses By trying actions and observing the results (e.g. braking hard on a wet road), agent can learn effects of actions When it receives no tip from passengers after driving wildly, it can learn a component of its utility function 7 8 Applications of Learning Natural language processing (A.L.I.C.E.) Speech/character/face recognition Spam detection, serving advertisements, Google PageRank Computer vision (iphone barcode app) Recommendation systems (NetFlix prize) Gene discovery Computational finance Robotics firefighting robots 3 Main types of learning Supervised learning Agent learns a function from observing example input-output pairs e.g. taxi agent that s a bus Unsupervised learning Learn patterns in the input without explicit feedback Most common task is clustering e.g. taxi agent notices bad traffic days Reinforcement learning Learns from a series of reinforcements: rewards or punishments e.g. 2 points for a win in chess 9 10 Supervised Learning There are many supervised learning methods: Decision trees Linear regression Linear classification Logistic regression Neural networks Nonparametric models, e.g. nearest neighbours and locally weighted regression Support vector machines We will introduce just a few of these entire courses on machine learning do not cover all of them Supervised learning applications Classification problems Facial recognition Handwriting recognition Speech recognition Spam detection Database marketing 11 12 2

The supervised learning task Given a training set of N example input-output pairs (x 1,y 1 ),(x 2,y 2 ), (x N,y N ), where each y j was generated by an unknown function y=f(x), discover a function h that approximates the true function f. Note: x and y can be any value (not just numbers) h is a hypothesis The supervised learning task 2 Learning is a search through the space of possible hypotheses for one that performs well, even on new beyond the training set Test the function by dividing the into a test set and a training set Learn a function from the training set Test its accuracy by applying to the (unseen) test set Hypothesis generalises well if it correctly predicts y from novel 13 14 Learning problem When y is discrete, i.e. one of a finite set of values (e.g. sunny, cloudy, yes, female) we have a classification problem When y is continuous, e.g. a number (e.g. tomorrow s temperature, age) we have a regression problem Hypothesis space How do we choose between multiple consistent hypotheses? Ockham s razor choose the simplest hypothesis consistent with the data 15 16 Supervised learning problems Lack of labelled data Data noise labels may not be accurate e.g. Learning ages from photos of faces take photos and ask age. Some people may lie about their age systematic inaccuracy not random noise Semi-supervised learning: Agent given a few labelled Must learn a large collection of unlabelled sample Learning Decision Trees (1) A decision tree is a simple representation for classifying which is a natural representation easily understood by humans Decision tree learning is one of the most successful techniques for supervised classification learning A decision tree represents a function that takes input vector of attribute values and returns a decision a single output value or class Input and output values can be discrete or continuous 17 18 3

Learning Decision Trees (2) In a decision tree: Each internal (non-leaf) node is labelled as an input attribute it tests a single attribute value Arcs are labelled with possible attribute values Leaves are labelled with a class/value to return To classify an example, filter it down the tree: For each node, follow the arc representing the example s attribute value When a leaf is reached, return the classification Example Decision Tree Problem Problem: decide whether to wait for a table at a restaurant, based on the following attributes: 1. Alternate: is there an alternative restaurant nearby? 2. Bar: is there a comfortable bar area to wait in? 3. Fri/Sat: is today Friday or Saturday? 4. Hungry: are we hungry? 5. Patrons: number of people in the restaurant (None, Some, Full) 6. Price: price range ($, $$, $$$) 7. Raining: is it raining outside? 8. Reservation: have we made a reservation? 9. Type: kind of restaurant (French, Italian, Thai, Burger) 10. WaitEstimate: estimated waiting time (0-10, 10-30, 30-60, >60) 19 20 Attribute-based representations Examples described by attribute values (Boolean, discrete, continuous) e.g. situations where I will/won't wait for a table: Restaurant Example Decision Tree One possible representation for hypotheses e.g. here is the true tree for deciding whether to wait: Classification of is positive (T) or negative (F) Must learn a definition for the Boolean goal predicate Wait NB doesn t use Price and Type attributes 21 22 Expressiveness Decision trees can express any function of the input attributes e.g. for Boolean functions, truth table row path to leaf: Any function in propositional logic can be expressed as a DT: Goal (Path1 Path2 ) where each path is a conjunction of attribute-value tests: Path=(Patrons=Full WaitEstimate=0-10) Decision Tree Learning (1) Aim: find a small tree consistent with the training Training set for a Boolean DT is (x,y) pair where x is the input vector and y is the Boolean output Greedy divide-and-conquer strategy : (recursively) choose most significant attribute as root of (sub)tree Divides the problem into smaller subproblems Always choose the most significant attribute first: the one that makes the most difference to classification in the training set Hope to classify by the smallest number of tests then the tree will be shallow and all paths short 23 24 4

Decision Tree Learning (2) 4 cases to consider for recursive DT problems: 1. If remaining all one class, STOP 2. If are a mix of class, choose best attribute to split them 3. If no remaining (i.e. no observed for this combination of attribute values) return default value 4. If no attributes left but of each class then the have the same description but different classifications, because: Error or noise in data Non-deterministic domain Can t observe an attribute which distinguishes Choosing an Attribute Idea: a good attribute splits the into subsets that are (ideally) all positive or all negative Patrons? is a better choice as it separates more 25 26 Restaurant Example contd. Decision tree learned from just12 : Evaluating Accuracy: Learning Curve How do we know that h f? Try h on a new test set of : Randomly split the example set into a training set and a test set Learn h then test its accuracy by applying to test set Repeat (e.g. 20 trials) using different size of training set, then plot Learning curve = % correct on test set as a function of training set size Substantially simpler than true tree a more complex hypothesis isn t justified by small amount of data 27 28 Broadening Application of Decision Trees Must consider several issues: Missing data: how to classify? Multivalued attributes: usefulness? Continuous and integer-valued attributes: split point? Continuous-valued output attributes: e.g. for numerical output we use a regression tree: each leaf has a linear function rather than a value Summary Today Learning needed for unknown environments, lazy designers Different types of learning For supervised learning, the aim is to find a simple hypothesis approximately consistent with training Decision tree learning Next time Linear models 29 30 5