Inductive Learning and Decision Trees

Save this PDF as:

Size: px
Start display at page:

Transcription

1 Inductive Learning and Decision Trees Doug Downey EECS 349 Winter 2014 with slides from Pedro Domingos, Bryan Pardo

2 Outline Announcements Homework #1 assigned Have you completed it? Inductive learning Decision Trees 2

3 Outline Announcements Homework #1 assigned Have you completed it? Inductive learning Decision Trees 3

4 Instances E.g. Four Days, in terms of weather: Sky Temp Humid Wind Water Forecast sunny warm normal strong warm same sunny warm high strong warm same rainy cold high strong warm change sunny warm high strong cool change

5 Functions Days on which my friend Aldo enjoys his favorite water sport INPUT OUTPUT Sky Temp Humid Wind Water Forecast f(x) sunny warm normal strong warm same 1 sunny warm high strong warm same 1 rainy cold high strong warm change 0 sunny warm high strong cool change 1 5

6 Inductive Learning! Predict the output for a new instance INPUT OUTPUT Sky Temp Humid Wind Water Forecast f(x) sunny warm normal strong warm same 1 sunny warm high strong warm same 1 rainy cold high strong warm change 0 sunny warm high strong cool change 1 rainy warm high strong cool change? 6

7 General Inductive Learning Task DEFINE: Set X of Instances (of n-tuples x = <x 1,..., x n >) E.g., days decribed by attributes (or features): Sky, Temp, Humidity, Wind, Water, Forecast Target function f : X Y, e.g.: EnjoySport X Y = {0,1} HoursOfSport X Y = {0, 1, 2, 3, 4} InchesOfRain X Y = [0, 10] GIVEN: Training examples D FIND: examples of the target function: <x, f(x)> A hypothesis h such that h(x) approximates f(x).

8 Another example: continuous attributes Learn function from x = (x 1,, x d ) to f (x) {0, 1} given labeled examples (x, f (x))? x 2 x 1

9 Hypothesis Spaces Hypothesis space H is a subset of all f : X Y e.g.: Linear separators Conjunctions of constraints on attributes (humidity must be low, and outlook!= rain) Etc. In machine learning, we restrict ourselves to H The subset aspect turns out to be important

10 Examples Credit Risk Analysis X: Properties of customer and proposed purchase f (x): Approve (1) or Disapprove (0) Disease Diagnosis X: Properties of patient (symptoms, lab tests) f (x): Disease (if any) Face Recognition X: Bitmap image f (x):name of person Automatic Steering X: Bitmap picture of road surface in front of car f (x): Degrees to turn the steering wheel

11 When to use? Inductive Learning is appropriate for building a face recognizer It is not appropriate for building a calculator You d just write a calculator program Question: What general characteristics make a problem suitable for inductive learning?

12 Think/Pair/Share What general characteristics make a problem suitable for inductive learning? Think Start End 12

13 Think/Pair/Share What general characteristics make a problem suitable for inductive learning? Pair Start End 13

14 Think/Pair/Share What general characteristics make a problem suitable for inductive learning? Share 14

15 Appropriate applications Situations in which: There is no human expert Humans can perform the task but can t describe how The desired function changes frequently Each user needs a customized f

16 Outline Announcements Homework #1 assigned Inductive learning Decision Trees 16

17 Task: Will I wait for a table? 17

18 18 Decision Trees!

19 Expressiveness of D-Trees 19

20 A learned decision tree 20

21 Inductive Bias To learn, we must prefer some functions to others Selection bias use a restricted hypothesis space, e.g.: linear separators 2-level decision trees Preference bias use the whole concept space, but state a preference over concepts, e.g.: Lowest-degree polynomial that separates the data shortest decision tree that fits the data 21

22 Decision Tree Learning (ID3) 22

23 Recap Inductive learning Goal: generate a hypothesis a function from instances described by attributes to an output using training examples. Requires inductive bias a restricted hypothesis space, or preferences over hypotheses. Decision Trees Simple representation of hypotheses, recursive learning algorithm Prefer smaller trees! 23

24 Choosing an attribute 24

25 Think/Pair/Share How should we choose which attribute to split on next? Think Start End 25

26 Think/Pair/Share How should we choose which attribute to split on next? Pair Start End 26

27 Think/Pair/Share How should we choose which attribute to split on next? Share 27

28 Information 28

29 H(V) Entropy The entropy H(V) of a Boolean random variable V as the probability of V = 0 varies from 0 to 1 29 P(V=0)

30 Using Information 30

31 Measuring Performance 31

32 What the learning curve tells us 32

33 Overfitting

34 Overfitting is due to noise Sources of noise: Erroneous training data concept variable incorrect (annotator error) Attributes mis-measured Much more significant: Irrelevant attributes Target function not realizable in attributes

35 Irrelevant attributes If many attributes are noisy, information gains can be spurious, e.g.: 20 noisy attributes 10 training examples Expected # of different depth-3 trees that split the training data perfectly using only noisy attributes: 13.4

36 Not realizable In general: We can t measure all the variables we need to do perfect prediction. => Target function is not uniquely determined by attribute values

37 Not realizable: Example Humidity EnjoySport Decent hypothesis: Humidity > 0.70 No Otherwise Yes Overfit hypothesis: Humidity > 0.89 No Humidity > 0.80 ^ Humidity <= 0.89 Yes Humidity > 0.70 ^ Humidity <= 0.80 No Humidity <= 0.70 Yes

38

39 Avoiding Overfitting Approaches Stop splitting when information gain is low or when split is not statistically significant. Grow full tree and then prune it when done 39

40

41 Effect of Reduced Error Pruning 41

42 Cross-validation

43 C4.5 Algorithm Builds a decision tree from labeled training data Generalizes simple ID3 tree by Prunes tree after building to improve generality Allows missing attributes in examples Allowing continuous-valued attributes 43

44 Rule post pruning Used in C4.5 Steps 1. Build the decision tree 2. Convert it to a set of logical rules 3. Prune each rule independently 4. Sort rules into desired sequence for use 44

45

46

47 Other Odds and Ends Unknown Attribute Values?

48

49 Odds and Ends Unknown Attribute Values? Continuous Attributes?

50 Decision Tree Boundaries 50

51

52 Decision Trees Bias How to solve 2-bit parity: Two step look-ahead, or Split on pairs of attributes at once For k-bit parity, why not just do k-step look ahead? Or split on k attribute values? =>Parity functions are among the victims of the decision tree s inductive bias.

53 Take away about decision trees Used as classifiers Supervised learning algorithms (ID3, C4.5) Good for situations where Inputs, outputs are discrete We think the true function is a small tree 53

Inductive Learning and Decision Trees

Inductive Learning and Decision Trees Doug Downey EECS 349 Spring 2017 with slides from Pedro Domingos, Bryan Pardo Outline Announcements Homework #1 was assigned on Monday (due in five days!) Inductive

Machine Learning B, Fall 2016

Machine Learning 10-601 B, Fall 2016 Decision Trees (Summary) Lecture 2, 08/31/ 2016 Maria-Florina (Nina) Balcan Learning Decision Trees. Supervised Classification. Useful Readings: Mitchell, Chapter 3

Decision Tree for Playing Tennis

Decision Tree Decision Tree for Playing Tennis (outlook=sunny, wind=strong, humidity=normal,? ) DT for prediction C-section risks Characteristics of Decision Trees Decision trees have many appealing properties

Supervised learning can be done by choosing the hypothesis that is most probable given the data: = arg max ) = arg max

The learning problem is called realizable if the hypothesis space contains the true function; otherwise it is unrealizable On the other hand, in the name of better generalization ability it may be sensible

CSC 4510/9010: Applied Machine Learning Rule Inference

CSC 4510/9010: Applied Machine Learning Rule Inference Dr. Paula Matuszek Paula.Matuszek@villanova.edu Paula.Matuszek@gmail.com (610) 647-9789 CSC 4510.9010 Spring 2015. Paula Matuszek 1 Red Tape Going

Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University. January 11, 2011

Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 11, 2011 Today: What is machine learning? Decision tree learning Course logistics Readings: The Discipline

18 LEARNING FROM EXAMPLES

18 LEARNING FROM EXAMPLES An intelligent agent may have to learn, for instance, the following components: A direct mapping from conditions on the current state to actions A means to infer relevant properties

Lecture 1: Basic Concepts of Machine Learning

Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010

P(A, B) = P(A B) = P(A) + P(B) - P(A B)

AND Probability P(A, B) = P(A B) = P(A) + P(B) - P(A B) P(A B) = P(A) + P(B) - P(A B) Area = Probability of Event AND Probability P(A, B) = P(A B) = P(A) + P(B) - P(A B) If, and only if, A and B are independent,

Machine Learning. Basic Concepts. Joakim Nivre. Machine Learning 1(24)

Machine Learning Basic Concepts Joakim Nivre Uppsala University and Växjö University, Sweden E-mail: nivre@msi.vxu.se Machine Learning 1(24) Machine Learning Idea: Synthesize computer programs by learning

Machine Learning 2nd Edition

INTRODUCTION TO Lecture Slides for Machine Learning 2nd Edition ETHEM ALPAYDIN, modified by Leonardo Bobadilla and some parts from http://www.cs.tau.ac.il/~apartzin/machinelearning/ The MIT Press, 2010

Machine Learning Tom M. Mitchell Machine Learning Department Carnegie Mellon University. January 12, 2015

Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 12, 2015 Today: What is machine learning? Decision tree learning Course logistics Readings: The Discipline

Outline. Learning from Observations. Learning agents. Learning. Inductive learning (a.k.a. Science) Environment. Agent.

Outline Learning agents Learning from Observations Inductive learning Decision tree learning Measuring learning performance Chapter 18, Sections 1 3 Chapter 18, Sections 1 3 1 Chapter 18, Sections 1 3

PRESENTATION TITLE. A Two-Step Data Mining Approach for Graduation Outcomes CAIR Conference

PRESENTATION TITLE A Two-Step Data Mining Approach for Graduation Outcomes 2013 CAIR Conference Afshin Karimi (akarimi@fullerton.edu) Ed Sullivan (esullivan@fullerton.edu) James Hershey (jrhershey@fullerton.edu)

A Few Useful Things to Know about Machine Learning. Pedro Domingos Department of Computer Science and Engineering University of Washington" 2012"

A Few Useful Things to Know about Machine Learning Pedro Domingos Department of Computer Science and Engineering University of Washington 2012 A Few Useful Things to Know about Machine Learning Machine

IAI : Machine Learning

IAI : Machine Learning John A. Bullinaria, 2005 1. What is Machine Learning? 2. The Need for Learning 3. Learning in Neural and Evolutionary Systems 4. Problems Facing Expert Systems 5. Learning in Rule

Machine Learning and Auto-Evaluation

Machine Learning and Auto-Evaluation In very simple terms, Machine Learning is about training or teaching computers to take decisions or actions without explicitly programming them. For example, whenever

Decision Tree For Playing Tennis

Decision Tree For Playing Tennis ROOT NODE BRANCH INTERNAL NODE LEAF NODE Disjunction of conjunctions Another Perspective of a Decision Tree Model Age 60 40 20 NoDefault NoDefault + + NoDefault Default

COMP 551 Applied Machine Learning Lecture 6: Performance evaluation. Model assessment and selection.

COMP 551 Applied Machine Learning Lecture 6: Performance evaluation. Model assessment and selection. Instructor: Herke van Hoof (herke.vanhoof@mail.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551

CSE 546 Machine Learning

CSE 546 Machine Learning Instructor: Luke Zettlemoyer TA: Lydia Chilton Slides adapted from Pedro Domingos and Carlos Guestrin Logistics Instructor: Luke Zettlemoyer Email: lsz@cs Office: CSE 658 Office

10701/15781 Machine Learning, Spring 2005: Homework 1

10701/15781 Machine Learning, Spring 2005: Homework 1 Due: Monday, February 6, beginning of the class 1 [15 Points] Probability and Regression [Stano] 1 1.1 [10 Points] The Matrix Strikes Back The Matrix

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Todd Holloway Two Lecture Series for B551 November 20 & 27, 2007 Indiana University Outline Introduction Bias and

Applied Machine Learning Lecture 1: Introduction

Applied Machine Learning Lecture 1: Introduction Richard Johansson January 16, 2018 welcome to the course! machine learning is getting increasingly popular among students our courses are full! many thesis

COMP 551 Applied Machine Learning Lecture 6: Performance evaluation. Model assessment and selection.

COMP 551 Applied Machine Learning Lecture 6: Performance evaluation. Model assessment and selection. Instructor: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551 Unless otherwise

Linear Regression. Chapter Introduction

Chapter 9 Linear Regression 9.1 Introduction In this class, we have looked at a variety of di erent models and learning methods, such as finite state machines, sequence models, and classification methods.

CS 354R: Computer Game Technology

CS 354R: Computer Game Technology AI Decision Trees and Rule Systems Fall 2017 Decision Trees Nodes represent attribute tests One child for each outcome Leaves represent classifications Can have same classification

Linear Models Continued: Perceptron & Logistic Regression

Linear Models Continued: Perceptron & Logistic Regression CMSC 723 / LING 723 / INST 725 Marine Carpuat Slides credit: Graham Neubig, Jacob Eisenstein Linear Models for Classification Feature function

LEARNING FROM OBSERVATIONS

1 LEARNING FROM OBSERVATIONS In which we describe agents that can improve their behavior through diligent study of their own experiences. The idea behind learning is that percepts should be used not only

Introduction to Classification

Introduction to Classification Classification: Definition Given a collection of examples (training set ) Each example is represented by a set of features, sometimes called attributes Each example is to

The Health Economics and Outcomes Research Applications and Valuation of Digital Health Technologies and Machine Learning

The Health Economics and Outcomes Research Applications and Valuation of Digital Health Technologies and Machine Learning Workshop W29 - Session V 3:00 4:00pm May 25, 2016 ISPOR 21 st Annual International

Course 395: Machine Learning - Lectures

Course 395: Machine Learning - Lectures Lecture 1-2: Concept Learning (M. Pantic) Lecture 3-4: Decision Trees & CBC Intro (M. Pantic & S. Petridis) Lecture 5-6: Evaluating Hypotheses (S. Petridis) Lecture

Assignment 6 (Sol.) Introduction to Machine Learning Prof. B. Ravindran

Assignment 6 (Sol.) Introduction to Machine Learning Prof. B. Ravindran 1. Assume that you are given a data set and a neural network model trained on the data set. You are asked to build a decision tree

Predictive Analysis of Text: Concepts, Features, and Instances

of Text: Concepts, Features, and Instances Jaime Arguello jarguell@email.unc.edu August 26, 2015 of Text Objective: developing and evaluating computer programs that automatically detect a particular concept

A Combination of Decision Trees and Instance-Based Learning Master s Scholarly Paper Peter Fontana,

A Combination of Decision s and Instance-Based Learning Master s Scholarly Paper Peter Fontana, pfontana@cs.umd.edu March 21, 2008 Abstract People are interested in developing a machine learning algorithm

CS 4510/9010 Applied Machine Learning. Evaluation. Paula Matuszek Fall, copyright Paula Matuszek 2016

CS 4510/9010 Applied Machine Learning 1 Evaluation Paula Matuszek Fall, 2016 Evaluating Classifiers 2 With a decision tree, or with any classifier, we need to know how well our trained model performs on

Big Data Analytics Clustering and Classification

E6893 Big Data Analytics Lecture 4: Big Data Analytics Clustering and Classification Ching-Yung Lin, Ph.D. Adjunct Professor, Dept. of Electrical Engineering and Computer Science September 28th, 2017 1

ANALYZING BIG DATA WITH DECISION TREES

San Jose State University SJSU ScholarWorks Master's Projects Master's Theses and Graduate Research Spring 2014 ANALYZING BIG DATA WITH DECISION TREES Lok Kei Leong Follow this and additional works at:

Learning. Part 6 in Russell / Norvig Book

Wisdom is not the product of schooling but the lifelong attempt to acquire it. - Albert Einstein Learning Part 6 in Russell / Norvig Book Gerhard Fischer AI Course, Fall 1996, Lecture October 14 1 Overview

Lecture 1. Introduction Bastian Leibe Visual Computing Institute RWTH Aachen University

Advanced Machine Learning Lecture 1 Introduction 20.10.2015 Bastian Leibe Visual Computing Institute RWTH Aachen University http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de Organization Lecturer

Let s Learn English Lesson Plan

Let s Learn English Lesson Plan Introduction: Let s Learn English lesson plans are based on the CALLA approach. See the end of each lesson for more information and resources on teaching with the CALLA

Introduction to Classification, aka Machine Learning

Introduction to Classification, aka Machine Learning Classification: Definition Given a collection of examples (training set ) Each example is represented by a set of features, sometimes called attributes

Learning Agents: Introduction

Learning Agents: Introduction S Luz luzs@cs.tcd.ie October 28, 2014 Learning in agent architectures Agent Learning in agent architectures Agent Learning in agent architectures Agent perception Learning

Analysis of Different Classifiers for Medical Dataset using Various Measures

Analysis of Different for Medical Dataset using Various Measures Payal Dhakate ME Student, Pune, India. K. Rajeswari Associate Professor Pune,India Deepa Abin Assistant Professor, Pune, India ABSTRACT

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Version Space Javier Béjar cbea LSI - FIB Term 2012/2013 Javier Béjar cbea (LSI - FIB) Version Space Term 2012/2013 1 / 18 Outline 1 Learning logical formulas 2 Version space Introduction Search strategy

Uninformed Search (Ch )

1 Uninformed Search (Ch. 3-3.4) 2 Announcements Will make homework this weekend (~4 days) due next weekend (~13 days) 3 What did we do last time? Take away messages: Lecture 1: Class schedule (ended early)

Classifying Breast Cancer By Using Decision Tree Algorithms

Classifying Breast Cancer By Using Decision Tree Algorithms Nusaibah AL-SALIHY, Turgay IBRIKCI (Presenter) Cukurova University, TURKEY What Is A Decision Tree? Why A Decision Tree? Why Decision TreeClassification?

TKT CLIL LESSON PLAN

TKT CLIL LESSON PLAN Teacher s name ROSANNA TROTTER Date lesson delivery: 23 rd May 2015 Time Subject 5 hours (observation 1 hour) Geography Class 4 th class, Siror s Primary school Group profile Strengths:

Ogdensburg School. Science Curriculum K-8 September 2015

Ogdensburg School Science Curriculum K-8 September 2015 1 Science Curriculum Grades K 8 th Rationale We believe that our Science program should prepare students to become scientifically literate. There

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

Chapter 1-2: Methods for Describing Sets of Data. Introductory Concepts:

Introductory Concepts: Statistics is the science of data. It involves collecting, classifying, summarizing, organizing, analyzing, and interpreting numerical information. Descriptive Stat: Involves collecting,

Cost-Sensitive Learning and the Class Imbalance Problem

To appear in Encyclopedia of Machine Learning. C. Sammut (Ed.). Springer. 2008 Cost-Sensitive Learning and the Class Imbalance Problem Charles X. Ling, Victor S. Sheng The University of Western Ontario,

Foundations of Intelligent Systems CSCI (Fall 2015)

Foundations of Intelligent Systems CSCI-630-01 (Fall 2015) Final Examination, Fri. Dec 18, 2015 Instructor: Richard Zanibbi, Duration: 120 Minutes Name: Instructions The exam questions are worth a total

Learning Concept Classification Rules Using Genetic Algorithms

Learning Concept Classification Rules Using Genetic Algorithms Kenneth A. De Jong George Mason University Fairfax, VA 22030 USA kdejong@aic.gmu.edu William M. Spears Naval Research Laboratory Washington,

Introduction to Pattern Recognition

Introduction to Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University)

STOR 155 Introductory Statistics. Lecture 11: General Probability Rules

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL STOR 155 Introductory Statistics Lecture 11: General Probability Rules 10/8/09 Lecture 11 1 Review Outcome, Sample space, Event Union (or), Intersection

STOR 155 Introductory Statistics. Lecture 11: General Probability Rules

The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL STOR 155 Introductory Statistics Lecture 11: General Probability Rules 2/24/11 Lecture 11 1 Review Outcome, Sample space, Event Union (or), Intersection

TOWARDS DATA-DRIVEN AUTONOMICS IN DATA CENTERS

TOWARDS DATA-DRIVEN AUTONOMICS IN DATA CENTERS ALINA SIRBU, OZALP BABAOGLU SUMMARIZED BY ARDA GUMUSALAN MOTIVATION 2 MOTIVATION Human-interaction-dependent data centers are not sustainable for future data

Optical Character Recognition Domain Expert Approximation Through Oracle Learning

Optical Character Recognition Domain Expert Approximation Through Oracle Learning Joshua Menke NNML Lab BYU CS josh@cs.byu.edu March 24, 2004 BYU CS Optical Character Recognition (OCR) optical character

ORIE 4741: Learning with Big Messy Data. Introduction

ORIE 4741: Learning with Big Messy Data Introduction Professor Udell Operations Research and Information Engineering Cornell September 15, 2017 1 / 33 Outline Stories Definitions Kinds of learning Syllabus

Data Mining CAP

Data Mining CAP 5771-001 Administrative Details The text is a high-level overview of data mining. You can supplement this by papers from the bibliography available on the Web. They will provide some details.

CSL465/603 - Machine Learning

CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am

Machine Learning for NLP

Natural Language Processing SoSe 2014 Machine Learning for NLP Dr. Mariana Neves April 30th, 2014 (based on the slides of Dr. Saeedeh Momtazi) Introduction Field of study that gives computers the ability

A study of the NIPS feature selection challenge

A study of the NIPS feature selection challenge Nicholas Johnson November 29, 2009 Abstract The 2003 Nips Feature extraction challenge was dominated by Bayesian approaches developed by the team of Radford

Bias and the Probability of Generalization

Brigham Young University BYU ScholarsArchive All Faculty Publications 1997-12-10 Bias and the Probability of Generalization Tony R. Martinez martinez@cs.byu.edu D. Randall Wilson Follow this and additional

CS545 Machine Learning

Machine learning and related fields CS545 Machine Learning Course Introduction Machine learning: the construction and study of systems that learn from data. Pattern recognition: the same field, different

Distinguish Wild Mushrooms with Decision Tree. Shiqin Yan

Distinguish Wild Mushrooms with Decision Tree Shiqin Yan Introduction Mushroom poisoning, which also known as mycetism, refers to harmful effects from ingestion of toxic substances present in the mushroom.

INTRODUCTION TO DATA SCIENCE

DATA11001 INTRODUCTION TO DATA SCIENCE EPISODE 6: MACHINE LEARNING TODAY S MENU 1. WHAT IS ML? 2. CLASSIFICATION AND REGRESSSION 3. EVALUATING PERFORMANCE & OVERFITTING WHAT IS MACHINE LEARNING? Definition:

A case study of a depression weather system, Klaus 2009.

A case study of a depression weather system, Klaus 2009. This series of lessons covers the formation of, weather associated with, and the impacts of a depression. It uses depression Klaus in 2009 as a

Dudon Wai Georgia Institute of Technology CS 7641: Machine Learning Atlanta, GA

Adult Income and Letter Recognition - Supervised Learning Report An objective look at classifier performance for predicting adult income and Letter Recognition Dudon Wai Georgia Institute of Technology

Class Noise vs. Attribute Noise: A Quantitative Study of Their Impacts

Artificial Intelligence Review 22: 177 210, 2004. Ó 2004 Kluwer Academic Publishers. Printed in the Netherlands. 177 Class Noise vs. Attribute Noise: A Quantitative Study of Their Impacts XINGQUAN ZHU*

Introduction to Classification and Clustering

Villanova University Machine Learning Project Introduction to lassification and lustering Overview This module introduces two important machine learning approaches: lassification and lustering. Each approach

El Rancho Unified School District

Grade: Kindergarten Selection Reading: The Wood Cutter s Cap Week 1 Day 1 Common Core Standards RL K.1 With prompting and support, ask and answer questions about key details in a text. RL K.3 With prompting

(Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include

Stay Alert!: Creating a Classifier to Predict Driver Alertness in Real-time Aditya Sarkar, Julien Kawawa-Beaudan, Quentin Perrot Friday, December 11, 2014 1 Problem Definition Driving while drowsy inevitably

Lecture 1: Introduc4on

CSC2515 Spring 2014 Introduc4on to Machine Learning Lecture 1: Introduc4on All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/csc2515_winter15.html

Constrained Dynamic Rule Induction Learning

Constrained Dynamic Rule Induction Learning Fadi Thabtah a, Issa Qabajeh b, Francisco Chiclana c a. Applied Business and Computing, NMIT, Auckland, New Zealand b. School of Computer Sciences and Informatics,

Practical Feature Subset Selection for Machine Learning

Practical Feature Subset Selection for Machine Learning Mark A. Hall, Lloyd A. Smith {mhall, las}@cs.waikato.ac.nz Department of Computer Science, University of Waikato, Hamilton, New Zealand. Abstract

Intelligent Systems. Neural Networks. Copyright 2009 Dieter Fensel and Reto Krummenacher

Intelligent Systems Neural Networks Copyright 2009 Dieter Fensel and Reto Krummenacher 1 Where are we? # Title 1 Introduction 2 Propositional Logic 3 Predicate Logic 4 Theorem Proving, Description Logics

ANNA UNIVERSITY SUBJECT NAME : ARTIFICIAL INTELLIGENCE SUBJECT CODE : CS2351 YEAR/SEM :III / VI QUESTION BANK UNIT I PROBLEM SOLVING 1. What is Intelligence? 2. Describe the four categories under which

WING-NUS at CL-SciSumm 2017: Learning from Syntactic and Semantic Similarity for Citation Contextualization

WING-NUS at CL-SciSumm 2017: Learning from Syntactic and Semantic Similarity for Citation Contextualization Animesh Prasad School of Computing, National University of Singapore, Singapore a0123877@u.nus.edu

Mathematical Sciences

Mathematical Sciences Associate Professors McKenzie R. Lamb (Chair), David W. Scott, Andrea N. Young Visiting Professors Mark A. Krines, William S. Retert Communicating Plus - Mathematical Sciences: Students

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Active Learning Yingyu Liang Computer Sciences 760 Fall 2017 http://pages.cs.wisc.edu/~yliang/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed by Mark Craven,

Intelligent Systems: Reasoning and Recognition

Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 and MoSIG M1 Winter Semester 2014 Lecture 2 7 February 2014 Outline: What is Reasoning?...1 Introduction to Expert Systems...2

On extending F-measure and G-mean metrics to multi-class problems

Data Mining VI 25 On extending F-measure and G-mean metrics to multi-class problems R. P. Espíndola & N. F. F. Ebecken COPPE/Federal University of Rio de Janeiro, Brazil Abstract The evaluation of classifiers

Lecture 1: Machine Learning Basics

1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

HAMLET JERRY ZHU UNIVERSITY OF WISCONSIN

HAMLET JERRY ZHU UNIVERSITY OF WISCONSIN Collaborators: Rui Castro, Michael Coen, Ricki Colman, Charles Kalish, Joseph Kemnitz, Robert Nowak, Ruichen Qian, Shelley Prudom, Timothy Rogers Somewhere, something

A Course in Machine Learning

A Course in Machine Learning Hal Daumé III Copyright 2012 Hal Daumé III http://ciml.info This book is for the use of anyone anywhere at no cost and with almost no restrictions whatsoever. You may copy

Tanagra Tutorials. Figure 1 Tree size and generalization error rate (Source:

1 Topic Describing the post pruning process during the induction of decision trees (CART algorithm, Breiman and al., 1984 C RT component into TANAGRA). Determining the appropriate size of the tree is a

Learning dispatching rules via an association rule mining approach. Dongwook Kim. A thesis submitted to the graduate faculty

Learning dispatching rules via an association rule mining approach by Dongwook Kim A thesis submitted to the graduate faculty in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE

6.034 Notes: Section 13.1

6.034 Notes: Section 13.1 Slide 13.1.1 Now that we have looked at the basic mathematical techniques for minimizing the training error of a neural net, we should step back and look at the whole approach

Learning to Predict Rare Events in Event Sequences

Appears in Proceedings of the 4 th International Conference on Knowledge Discovery and Data Mining, AAAI Press, 1998, 359-363. Learning to Predict Rare Events in Event Sequences Gary M. Weiss * and Haym

RESEARCH METHODOLOGY AND LITERATURE REVIEW ASSOCIATE PROFESSOR DR. RAYNER ALFRED

RESEARCH METHODOLOGY AND LITERATURE REVIEW ASSOCIATE PROFESSOR DR. RAYNER ALFRED WRITING A LITERATURE REVIEW ASSOCIATE PROFESSOR DR. RAYNER ALFRED A literature review discusses

Using Two Variables to Represent a Relationship

Using Two Variables to Represent a Relationship LAUNCH (8 MIN) Before What is generally meant by the phrase think outside the box? During How is it possible that two cardboard boxes can be the same size

Automatic Induction of MAXQ Hierarchies

Automatic Induction of MAXQ Hierarchies Neville Mehta, Mike Wynkoop, Soumya Ray, Prasad Tadepalli, and Tom Dietterich School of EECS, Oregon State University Scaling up reinforcement learning to large

CS 446: Machine Learning

CS 446: Machine Learning Introduction to LBJava: a Learning Based Programming Language Writing classifiers Christos Christodoulopoulos Parisa Kordjamshidi Motivation 2 Motivation You still have not learnt

Grade 1 Transition to New TN Mathematics Standards

Grade 1 Transition to New TN Mathematics Standards 225 Grade 1 Mathematical Processes New Grade Level Expectations GLE 0106.1.1 Use mathematical language, symbols, and definitions while developing mathematical

Cross-Domain Video Concept Detection Using Adaptive SVMs

Cross-Domain Video Concept Detection Using Adaptive SVMs AUTHORS: JUN YANG, RONG YAN, ALEXANDER G. HAUPTMANN PRESENTATION: JESSE DAVIS CS 3710 VISUAL RECOGNITION Problem-Idea-Challenges Address accuracy