Activity: Evaluating Classifiers. Exercise 1. Exercise 1 - Data

Similar documents
Learning From the Past with Experiment Databases

Python Machine Learning

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

CS 446: Machine Learning

Managing the Student View of the Grade Center

Setting Up Tuition Controls, Criteria, Equations, and Waivers

CS Machine Learning

CSL465/603 - Machine Learning

SECTION 12 E-Learning (CBT) Delivery Module

A Case Study: News Classification Based on Term Frequency

Lecture 1: Machine Learning Basics

Reducing Features to Improve Bug Prediction

Rule Learning with Negation: Issues Regarding Effectiveness

Speech Emotion Recognition Using Support Vector Machine

Indian Institute of Technology, Kanpur

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Assignment 1: Predicting Amazon Review Ratings

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Switchboard Language Model Improvement with Conversational Data from Gigaword

Lecture 1: Basic Concepts of Machine Learning

ALEKS. ALEKS Pie Report (Class Level)

Beyond the Pipeline: Discrete Optimization in NLP

Rule Learning With Negation: Issues Regarding Effectiveness

Activity Recognition from Accelerometer Data

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Artificial Neural Networks written examination

Detailed Instructions to Create a Screen Name, Create a Group, and Join a Group

Moodle MyFeedback update April 2017

Experience College- and Career-Ready Assessment User Guide

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Australian Journal of Basic and Applied Sciences

6 Financial Aid Information

USER GUIDANCE. (2)Microphone & Headphone (to avoid howling).

Outreach Connect User Manual

Softprop: Softmax Neural Network Backpropagation Learning

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

SCT Banner Financial Aid Needs Analysis Training Workbook January 2005 Release 7

How to write websites in an essay >>>CLICK HERE<<<

EMPOWER Self-Service Portal Student User Manual

Skyward Gradebook Online Assignments

LMS - LEARNING MANAGEMENT SYSTEM END USER GUIDE

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Issues in the Mining of Heart Failure Datasets

Applications of data mining algorithms to analysis of medical data

Test Effort Estimation Using Neural Network

Student Handbook. This handbook was written for the students and participants of the MPI Training Site.

DegreeWorks Advisor Reference Guide

Time series prediction

Java Programming. Specialized Certificate

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Genre classification on German novels

NCAA Eligibility Center High School Portal Instructions. Course Module

Moodle 2 Assignments. LATTC Faculty Technology Training Tutorial

Training Pack. Kaizen Focused Improvement Teams (F.I.T.)

A new way to share, organize and learn from experiments

Off-line handwritten Thai name recognition for student identification in an automated assessment system

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

An Introductory Blackboard (elearn) Guide For Parents

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

MOODLE 2.0 GLOSSARY TUTORIALS

Universidade do Minho Escola de Engenharia

New Features & Functionality in Q Release Version 3.1 January 2016

How to set up gradebook categories in Moodle 2.

Intel-powered Classmate PC. SMART Response* Training Foils. Version 2.0

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

arxiv: v1 [cs.lg] 15 Jun 2015

TIPS PORTAL TRAINING DOCUMENTATION

Human Emotion Recognition From Speech

EDCI 699 Statistics: Content, Process, Application COURSE SYLLABUS: SPRING 2016

arxiv: v1 [cs.cv] 10 May 2017

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

Municipal Accounting Systems, Inc. Wen-GAGE Gradebook FAQs

Creating Your Term Schedule

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Creating an Online Test. **This document was revised for the use of Plano ISD teachers and staff.

Chamilo 2.0: A Second Generation Open Source E-learning and Collaboration Platform

INSTRUCTOR USER MANUAL/HELP SECTION

The following information has been adapted from A guide to using AntConc.

Test How To. Creating a New Test

Best Colleges Main Survey

Feature Selection based on Sampling and C4.5 Algorithm to Improve the Quality of Text Classification using Naïve Bayes

CORRECT YOUR ENGLISH ERRORS BY TIM COLLINS DOWNLOAD EBOOK : CORRECT YOUR ENGLISH ERRORS BY TIM COLLINS PDF

Beginning Blackboard. Getting Started. The Control Panel. 1. Accessing Blackboard:

Content-based Image Retrieval Using Image Regions as Query Examples

Large-Scale Web Page Classification. Sathi T Marath. Submitted in partial fulfilment of the requirements. for the degree of Doctor of Philosophy

THE FRYDERYK CHOPIN UNIVERSITY OF MUSIC

ecampus Basics Overview

Discriminative Learning of Beam-Search Heuristics for Planning

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Preparing for the School Census Autumn 2017 Return preparation guide. English Primary, Nursery and Special Phase Schools Applicable to 7.

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

READ 180 Next Generation Software Manual

Axiom 2013 Team Description Paper

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Emporia State University Degree Works Training User Guide Advisor

Knowledge-Based - Systems

Transcription:

Exercise 1 % Data for Evaluation exercises. % % We have run a classifier to learn whether someone will pass a course. % In the results below, we have 12 cases with learned % prediction and % actual result. % % %Case # Predicted Actual 01 yes yes 02 yes yes 03 yes yes 04 yes yes 05 yes yes 06 yes yes 07 yes no 08 no yes 09 no yes 10 no no 11 no no 12 no no Instructions Exercise 1 - Data Exercise 1 Data has a set of predicted and actual results for a system which predicts someone will pass a course. Based on this set: 1. What is the accuracy of this classifier? 2. What is the accuracy of the majority classifier for the same data?

3. Draw the confusion matrix for this classifier. 4. If you can admit as many students as you wish, would you use this system to decide who gets in? 5. If you only have room for seven students, would you use this system to decide who gets in?

Exercise 2 using Weka Note Weka provides a lot of information. This module does not go into detail about the more complex data. If students want to explore more on their own, the Weka documentation page (http://www.cs.waikato.ac.nz/ml/weka/documentation.html) provides information. This exercise assumes that the students have access to Weka. The questions can actually be addressed without running Weka by giving the students copies of Figures 3, 4, 5 and 8. Instructions 1. Start Weka and the Explorer. Figure 1: Weka start page

2. In the Preprocess tab, open the file credit-r.arff, which is in the installed Weka data file. This is a set of cases for deciding whether someone has good or bad credit. 3. In the Preprocess tab, with credit-r open: Figure 2: Opening credit-r Figure 3: credit-r open

3a. What is the relation being studied? (german_credit)? 3b. How many instances are there? (1000) 3c. Which attribute number are we trying to predict? In other words, which is the class? (#21) 3d. How many good cases are there? (700) How many bad? (300) 3e. Looking just at the information in the Preprocess tab, what is the accuracy for the majority classifier? (70%. We will predict good for all thousand cases, and be right for the 700 good cases, so 700/1000, or 70%. 4. Choose the Classify tab. Figure 4: Classifier tab with defaults

Look first at the defaults. 4a. What is the default Classifier? (ZeroR, which is what Weka calls the majority classifier). 4b. What is the default test protocol? (Cross-validation with a 10 fold split). What is the default class? (class, #21. Weka defaults to the last attribute) 5. Run the ZeroR classifier by clicking the start button. Figure 5: Result of running ZeroR

5a. How long did it take to run and test the classifier? Answer: no noticeable time, unless your computer is incredibly slow.) 5b. What is the accuracy? (70%) 5c. How many good credit applications are called bad? (0) 5d. How many bad credit applications are called good? (300) 5e: Is this useful? (no) 6. Change the classifier to SMO. (This is the Weka Support Vector Machine method. It s slower and not as good as SVM, but does not involve installing additional software). Turn off all options except Output confusion matrix. Run the SMO classifier by choosing Start. Figure 6: Choosing SMO, under Classifiers->Functions

Figure 7: Setting the options

Figure 8: SMO Classifier results

6a. How long did it take to run and test the classifier? (Typically, a few seconds; much longer than ZeroR, even for a few cases) 6b. What is the accuracy? (75.1%) 6c. How many good credit applications are called bad? (90) 6d. How many bad credit applications are called good? (159) 6e. Is this useful? (Note first that although the accuracy hasn t changed much, the confusion matrix is very different. So maybe. It s certainly more informative than the majority classifier, but you re still going to deny credit to some people who would be good credit risks, and give loans to 159 people who are bad risks. Possibly better than your human evaluators do? Comes down to the cost of making a mistake. 7. Advanced exercise: can the students do better with other classifiers, or other parameters? This can be an interesting exploration exercise. The typical classifiers with Weka defaults (decision tree, regression, neural nets, simple naive Bayes) do about the same, or slightly worse. Also note that neural nets (multilayer perceptrons in Weka) are very slow. With 10- fold cross-validation, this can take minutes for the credit-r data set.