Evaluating the Performance of Classification Algorithms Based on Metrics over Different Datasets

Similar documents
Rule Learning With Negation: Issues Regarding Effectiveness

Rule Learning with Negation: Issues Regarding Effectiveness

Python Machine Learning

CS Machine Learning

Learning From the Past with Experiment Databases

(Sub)Gradient Descent

Lecture 1: Machine Learning Basics

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

A Case Study: News Classification Based on Term Frequency

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Australian Journal of Basic and Applied Sciences

Applications of data mining algorithms to analysis of medical data

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Chapter 2 Rule Learning in a Nutshell

Reducing Features to Improve Bug Prediction

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Generative models and adversarial training

Human Emotion Recognition From Speech

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Evolutive Neural Net Fuzzy Filtering: Basic Description

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

On-Line Data Analytics

Mining Association Rules in Student s Assessment Data

Universidade do Minho Escola de Engenharia

CS 446: Machine Learning

A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and

CSL465/603 - Machine Learning

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Softprop: Softmax Neural Network Backpropagation Learning

Learning Methods for Fuzzy Systems

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

Computerized Adaptive Psychological Testing A Personalisation Perspective

Assignment 1: Predicting Amazon Review Ratings

Fuzzy rule-based system applied to risk estimation of cardiovascular patients

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Speech Recognition at ICSI: Broadcast News and beyond

Improving Simple Bayes. Abstract. The simple Bayesian classier (SBC), sometimes called

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Analyzing sentiments in tweets for Tesla Model 3 using SAS Enterprise Miner and SAS Sentiment Analysis Studio

Learning Methods in Multilingual Speech Recognition

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

Disambiguation of Thai Personal Name from Online News Articles

Word Segmentation of Off-line Handwritten Documents

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Fragment Analysis and Test Case Generation using F- Measure for Adaptive Random Testing and Partitioned Block based Adaptive Random Testing

INPE São José dos Campos

Data Stream Processing and Analytics

Model Ensemble for Click Prediction in Bing Search Ads

Linking Task: Identifying authors and book titles in verbose queries

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Calibration of Confidence Measures in Speech Recognition

Mining Student Evolution Using Associative Classification and Clustering

Grade 6: Correlated to AGS Basic Math Skills

Test Effort Estimation Using Neural Network

Generation of Attribute Value Taxonomies from Data for Data-Driven Construction of Accurate and Compact Classifiers

Handling Concept Drifts Using Dynamic Selection of Classifiers

Semi-Supervised Face Detection

Software Maintenance

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Switchboard Language Model Improvement with Conversational Data from Gigaword

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

WHEN THERE IS A mismatch between the acoustic

An Effective Framework for Fast Expert Mining in Collaboration Networks: A Group-Oriented and Cost-Based Method

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Using dialogue context to improve parsing performance in dialogue systems

K-Medoid Algorithm in Clustering Student Scholarship Applicants

Content-based Image Retrieval Using Image Regions as Query Examples

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Modeling function word errors in DNN-HMM based LVCSR systems

A Case-Based Approach To Imitation Learning in Robotic Agents

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Optimizing to Arbitrary NLP Metrics using Ensemble Selection

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Data Fusion Through Statistical Matching

Lecture 10: Reinforcement Learning

AQUA: An Ontology-Driven Question Answering System

Beyond the Pipeline: Discrete Optimization in NLP

ABSTRACT. A major goal of human genetics is the discovery and validation of genetic polymorphisms

Probability and Statistics Curriculum Pacing Guide

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Preference Learning in Recommender Systems

Delaware Performance Appraisal System Building greater skills and knowledge for educators

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

arxiv: v1 [cs.cl] 2 Apr 2017

Multi-Lingual Text Leveling

Customized Question Handling in Data Removal Using CPHC

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS

Issues in the Mining of Heart Failure Datasets

Transcription:

Evaluating the Performance of Classification Algorithms Based on Metrics over Different Datasets D.Ramya Department of Computer Science & Engineering, Sri Venkateswara College of Engineering & Technology, (India) ABSTRACT Classification is a data mining (machine learning) technique used to predict group membership for data instances. Evaluation of information, gathered almost everywhere in our day to day life can help devising some efficient and personalized strategies. Classification is one of the fundamental tasks in data mining and has also been studied extensively in statistics, machine learning, neural networks and expert systems over decades. Classification is a well known data mining technique that tells the class of an unknown object. For this purpose, classification predicts categorical (discrete, unordered) labels. Many classification algorithms have been proposed by researchers in statistics, machine learning and pattern recognition. In this study, the performance evaluation of BayesNet, LWL, BFTree, J48, LADTree and NBTree classification algorithms are experimented. I am Scheming the Performance of selective classification algorithms over different chosen data sets based on evaluation metrics Precision, F-Measure and ROC Area. Keywords : Classification, BayesNet, LWL, BFTree, J48, LADTree, NBTree I. INTRODUCTION Data mining is the process of exploration and analysis, by automatic and semi automatic means of large quantities of data in order to discover meaningful patterns and rules. The six main data mining activities are classification, estimation, prediction, affinity grouping, clustering, estimation and visualization [1]. From the past few years, the fields of machine learning and data mining have been studied to a great extent and applied in various fields of studies. It is now realized among the research communities that the contribution of machine learning has become immense for the development of science and technology. Classification, which is one of the supervised machine learning methodologies, is related to one of the fundamental tasks in data mining and has also been studied extensively in statistics, neural networks and expert systems over decades [2]. Classification involves two phases-construction of a model for classification/prediction and testing & usage of it for determining the class labels/ prediction. In this paper, performance evaluation of BayesNet, LWL, BFTree, J48, LADTree and NBTree classification algorithms are experimented based on the five different standard UCI data sets. II. LEARNING ALGORITHMS FOR CLASSIFICATION BayesNet: A Bayes Net is a model. It reflects the states of some part of a world that is being modeled and it describes how those states are related by probabilities. The model might be of your house, or your car, your 1070 P a g e

body, your community, an ecosystem, a stock-market, etc. Absolutely anything can be modeled by a Bayes Net. All the possible states of the model represent all the possible worlds that can exist, that is, all the possible ways that the parts or states can be configured. The car engine can be running normally or giving trouble. It's tires can be inflated or flat. Your body can be sick or healthy, and so on. LWL: Locally Weighted Learning is a class of function approximation techniques, where a prediction is done by using an approximated local model around the current point of interest. BFTree: An alternating decision tree (ADTree) is a machine learning method for classification. It generalizes decision trees and has connections to boosting. An ADTree consists of an alternation of decision nodes, which specify a predicate condition, and prediction nodes, which contain a single number. An instance is classified by an ADTree by following all paths for which all decision nodes are true, and summing any prediction nodes that are traversed and Class for building a best-first decision tree classifier is known as BFTree [3]. J48: In case the instances belong to the same class the tree represents a leaf so the leaf is returned by labeling with the same class. The potential information is calculated for every attribute, given by a test on the attribute. Then the gain in information is calculated that would result from a test on the attribute. Then the best attribute is found on the basis of the present selection criterion and that attribute selected for branching [4]. LADTree: LAD Tree builds a classifier for binary target variable based on learning a logical expression that can discriminate between positive and negative samples in a data set. LAD Tree Classifier generates a multi-class alternating decision tree using the Logit Boost strategy. The LAD Tree algorithm applies logistic boosting algorithm in order to induce an alternating decision tree. In this algorithm, a single attribute test is chosen as a splitter node for the tree at each iteration. For each training instance, working response and weights are calculated and stored on a per-class basis. Then, it fits the working response to the mean value of the instances, in a particular subset, by minimizing the least-squares value between them. In this algorithm, trees for the different classes are grown in parallel. Once all the trees have been constructed, then it merges the trees into a final model [5]. NBTree: The NBTree algorithm is a hybrid of the Naïve Bayes and the Decision Tree algorithm. This tree is constructed recursively. But, the leaf nodes are Naive Bayes categorizers. The NBTree algorithm strives to approximate whether the generalization accuracy of Naive Bayes at each leaf is higher than a single Naive Bayes classifier at that node. A split is considered to be significant if relative reduction in the error is greater than 5% and there are a minimum of 30 instances in the node. For discrete valued attributes, the Naive Bayes method performs quite well. With the increase in data size, the performance also improves. But in case of continuous valued attributes, Naive Bayes method does not take into account the attribute interactions. Whereas, the decision trees do not give good performance when the data size is very large. These shortcomings are overcome by the NBTree algorithm [6]. III. DATASETS In the classification algorithms, data sets are transformed into training sets and test sets in order to build a model and use it for the classification purpose respectively. The training set involves the various attributes 1071 P a g e

having one as classifying attribute. On the other hand the test set includes the same attributes with the unseen tuples of data that the model is going to classify the instances [7]. IV. EXPERIMENT To conduct experiment, I used six classifiers namely BayesNet, LWL, BFTree, J48, LADTree and NBTree on the five different UCI data sets (Breast-Cancer, Weather, Labor, Diabetes and Iris2D). The following table shows the description of considered datasets. Table 1: Data Description DATASETS INSTANCES ATTRIBUTES TYPE Breast-Cancer 286 10 Weather 14 5 Nominal Labor 57 17 Diabetes 768 9 Numeric Iris2d 150 43 For the performance issue of classifiers, I focused on the evaluation parameters: precision, F-measure and ROC Area. V. RESULT The Fig. 1 shows the Accuracy comparison of six selective classification algorithms over different chosen data sets. Fig. 1. : Accuracy comparison of different classifiers On the other hand, the figures named as Fig. 2, Fig. 3 and Fig. 4 shows the Performance of selective classification algorithms over different chosen data sets based on evaluation metrics Precision, F-Measure and ROC Area respectively. 1072 P a g e

Fig. 2. : Average precision of learning classifiers over chosen datasets Fig. 3. : Average F-Measure of learning classifiers over chosen datasets Fig. 4. Average ROC Area of learning classifiers over chosen datasets 1073 P a g e

VI. CONCLUSION In this paper, various data classification techniques such as BayesNet, LWL, BFTree, J48, LADTree, NBTree have been discussed. I consider the precision, F-measure and ROC Area as evaluation metrics for conducting my experiment on the performance evaluation of different classifiers over five different UCI datasets (i.e., Breast_Cancer, Weather, Labor, Diabetes and Iris2D). REFERENCES [1] Dr. D. Durga Bhavani, A. Vasavi, P.T. Keshava : Machine Learning: A Critical Review of Classification Techniques, International Journal of Advanced Research in Computer and Communication Engineering (2278-1021) Vol. 5, Special Issue 3, November 2016. [2] D.Lavanya,Dr. K.Usha Rani : Performance Evaluation of Decision Tree Classifiers on Medical Datasets, International Journal of Computer Applications (0975 8887)Volume 26 No.4, July 2011. [3] Abhaya Kumar Samal, Subhendu Kumar Pani, Comparative Study of J48, AD Tree, REP Tree and BF Tree Data Mining Algorithms through Colon Tumour Dataset, IJSRD - International Journal for Scientific Research & Development Vol. 4, Issue 03, 2016 ISSN (online): 2321-0613. [4] Gaganjot Kaur, Amit Chhabra Improved J48 Classification Algorithm for the Prediction of Diabetes,International Journal of Computer Applications (0975 8887) Volume 98 No.22, July 2014. [5] Lakshmi Devasena C, proficiency comparison of ladtree And reptree classifiers for credit Risk forecast, International Journal on Computational Sciences & Applications (IJCSA) Vol.5, No.1, February 2015. [6] Rupali Malviya, Brajesh K. Umrao, Comparison of NBTree and VFI Machine Learning Algorithms for Network Intrusion Detection using Feature Selection, International Journal of Computer Applications (0975 8887) Volume 108 No. 2, December 2014. [7] Solomon Getahun Fentie, Abebe Demessie Alemu, Bhabani Shankar D. M. : A Comparative Study on Performance Evalution of Eager versus Lazy Learning Methods, IJCSMC(ISSN 2320 088X), Vol. 3, Issue. 3, March 2014, pg.562 568. 1074 P a g e