A NEW ALGORITHM FOR GENERATION OF DECISION TREES

Similar documents
The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Lecture 1: Machine Learning Basics

Rule Learning With Negation: Issues Regarding Effectiveness

Applications of data mining algorithms to analysis of medical data

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

Rule Learning with Negation: Issues Regarding Effectiveness

Python Machine Learning

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Learning From the Past with Experiment Databases

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

stateorvalue to each variable in a given set. We use p(x = xjy = y) (or p(xjy) as a shorthand) to denote the probability that X = x given Y = y. We al

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

CS Machine Learning

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Henry Tirri* Petri Myllymgki

The Good Judgment Project: A large scale test of different methods of combining expert predictions

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Laboratorio di Intelligenza Artificiale e Robotica

Problems of the Arabic OCR: New Attitudes

Artificial Neural Networks written examination

Evolutive Neural Net Fuzzy Filtering: Basic Description

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Switchboard Language Model Improvement with Conversational Data from Gigaword

Assignment 1: Predicting Amazon Review Ratings

Introduction to Simulation

A Neural Network GUI Tested on Text-To-Phoneme Mapping

CS 446: Machine Learning

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

(Sub)Gradient Descent

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

CSL465/603 - Machine Learning

Lecture 1: Basic Concepts of Machine Learning

Laboratorio di Intelligenza Artificiale e Robotica

Pp. 176{182 in Proceedings of The Second International Conference on Knowledge Discovery and Data Mining. Predictive Data Mining with Finite Mixtures

AQUA: An Ontology-Driven Question Answering System

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

INTRODUCTION TO DECISION ANALYSIS (Economics ) Prof. Klaus Nehring Spring Syllabus

Radius STEM Readiness TM

Integrating E-learning Environments with Computational Intelligence Assessment Agents

Planning with External Events

Agent-Based Software Engineering

Reducing Features to Improve Bug Prediction

ICT SUPPORTED ENGINEERING COURSE CASE STUDY AND GUIDELINES

Learning Methods for Fuzzy Systems

Content-based Image Retrieval Using Image Regions as Query Examples

Reinforcement Learning by Comparing Immediate Reward

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Cooperative evolutive concept learning: an empirical study

Seminar - Organic Computing

Welcome to. ECML/PKDD 2004 Community meeting

Improving Simple Bayes. Abstract. The simple Bayesian classier (SBC), sometimes called

An OO Framework for building Intelligence and Learning properties in Software Agents

Softprop: Softmax Neural Network Backpropagation Learning

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Guru: A Computer Tutor that Models Expert Human Tutors

Axiom 2013 Team Description Paper

Discriminative Learning of Beam-Search Heuristics for Planning

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Human Emotion Recognition From Speech

Proceedings of the Federated Conference on Computer Science DOI: /2016F560 and Information Systems pp ACSIS, Vol. 8.

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Probability estimates in a scenario tree

Go fishing! Responsibility judgments when cooperation breaks down

Evolution of Symbolisation in Chimpanzees and Neural Nets

Learning and Transferring Relational Instance-Based Policies

Semi-Supervised Face Detection

Chapter 2 Rule Learning in a Nutshell

Probabilistic Latent Semantic Analysis

Action Models and their Induction

Curriculum Vitae FARES FRAIJ, Ph.D. Lecturer

Managing Experience for Process Improvement in Manufacturing

Efficient Online Summarization of Microblogging Streams

Calibration of Confidence Measures in Speech Recognition

Learning goal-oriented strategies in problem solving

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Classification Using ANN: A Review

The CTQ Flowdown as a Conceptual Model of Project Objectives

Toward Probabilistic Natural Logic for Syllogistic Reasoning

Evaluating and Comparing Classifiers: Review, Some Recommendations and Limitations

A Bayesian Learning Approach to Concept-Based Document Classification

An Investigation into Team-Based Planning

Courses in English. Application Development Technology. Artificial Intelligence. 2017/18 Spring Semester. Database access

***** Article in press in Neural Networks ***** BOTTOM-UP LEARNING OF EXPLICIT KNOWLEDGE USING A BAYESIAN ALGORITHM AND A NEW HEBBIAN LEARNING RULE

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Australian Journal of Basic and Applied Sciences

Introduction to Causal Inference. Problem Set 1. Required Problems

Learning to Rank with Selection Bias in Personal Search

A Comparison of Two Text Representations for Sentiment Analysis

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

Conversational Framework for Web Search and Recommendations

Generation of Attribute Value Taxonomies from Data for Data-Driven Construction of Accurate and Compact Classifiers

Transcription:

TASK QUARTERLY 8 No 2(2004), 1001 1005 A NEW ALGORITHM FOR GENERATION OF DECISION TREES JERZYW.GRZYMAŁA-BUSSE 1,2,ZDZISŁAWS.HIPPE 2, MAKSYMILIANKNAP 2 ANDTERESAMROCZEK 2 1 DepartmentofElectricalEngineeringandComputerScience, Kansas University, Lawrence(KS) USA jerzy@eecs.ku.edu 2 DepartmentofExpertSystemsandArtificialIntelligence, University of Information Technology and Management, Sucharskiego 2, 35-225 Rzeszow, Poland {zhippe, mknap, tmroczek}@wenus.wsiz.rzeszow.pl (Received 30 January 2004) Abstract: A new algorithm for development of quasi-optimal decision trees, based on the Bayes theorem, has been created and tested. The algorithm generates a decision tree on the basis of Bayesian belief networks, created prior to the formation of the decision tree. The efficiency of this new algorithm was compared with three other known algorithms used to develop decision trees. The data set used for the experiments was a set of cases of skin lesions, histopatolgically verified. Keywords: artificial intelligence, supervised machine learning, decision trees, Bayes networks 1. Introduction The main goal of our research was developing a computer-assisted methodology of early and noninvasive diagnosis of one of the most dangerous human diseases, skin cancer[1]. Research described in this paper involved unsupervised machine learning in classification and identification of melanocytic skin lesions. To discover the knowledge hidden in medical datasets a computer program suite was created. Such datasets are frequently uncertain, e.g. conflicting. Until now, four computer programs have been created:affinityseeker R (usingaminimal-distancemethodtofindsimilaritybetween theinvestigatedobjects,inourcase,melanocyticskinlesions)[2],beliefseeker R (generatingstochasticbeliefnetworks)[3],treeseeker R (generatingquasi-optimal decisiontrees)[4],andplaneseeker R (usingoptimizedalgorithmsoflinearmachine learning to identify multicategory objects using a binary recurrent classification engine)[5]. Main features of these information systems, developed at the Kansas University in Lawrence, KS USA are(i) a uniform format of input data, compatible with the format used by the LERS system(which generates learning models in the form of sets of rules)[6], and(ii) the ability to generate twofold learning models: tq208m-c/1001 4 III 2004 BOP s.c., http://www.bop.com.pl

1002 J. W. Grzymała-Busse, Z. S. Hippe, M. Knap and T. Mroczek a certain model(for data sets without conflicting cases) and a possible model(in case of data sets with conflicting cases). In this paper we concentrate on algorithms togeneratedecisiontrees.inourresearch[7]wehaveconcludedthatthistypeof learning model provides the most promising results in diagnosis of melanocytic skin lesions.itwasnecessarytoequipthetreeseeker R systemwithadditional,new algorithms for creating decision trees and checking their effectiveness in comparison with the earlier algorithms implemented by Czerwiński[8] and Quinlan[9]. Our experience in generating belief networks[10] suggests to use these algorithms for selection of attributes, the most significant part of the process of classification of adataset. 2. Algorithms for generation of decision trees In our research we have used the following algorithms to generate decision trees:(i) Czerwiński s algorithm[8],(ii) our own implementation of the classical Quinlan algorithm[9], i.e. C4.5 using information entropy for attribute selection, (iii) the TVR algorithm(creating decision trees from fragments, which are sequences of paths from selected attributes to the decision attribute)[11] and(iv) VDP, a new algorithm searching for the most significant set of attributes, required for the correct classification of the training data. This algorithm has been the main subject of our research. The VDP algorithm is based on generating belief networks with varies Dirichlet s parameter[12]. Let us note that the descriptive attribute which has the greatest marginal influence[13] on the decision attribute is placed in the root of the decision tree. 3. Research methodology Four algorithms for generating decision trees(czerwiński s, Quinlan s, TVR and VDP) were compared using the following assumptions. All four algorithms were usedforthenevidataset,whichispresentedindetailin[14].intheinitialstageof research, 250 cases of melanocytic skin lesions were randomly divided into two groups of167and83cases.then,decisiontreesweregeneratedforthe167casesusingall four algorithms. Quality of these trees was estimated on the basis of classification results of the 83 unseen cases. Obviously, the best algorithm is the algorithm that generates a decision tree with the lowest error rate. 4. Results The generated decision trees are presented in Figures 1 4. Results of classificationofunseenobjectsarepresentedintable1.foranideaoftheaveragenumber of questions see[15]. Table 1. Quality of the tested algorithms Tested algorithm Mean number of questions Number of nodes Error of classification[%] Czerwiński s 2.70 7 2.41 Quinlan s 2.43 7 2.41 TVR 2.00 3 1.20 VDP 2.32 5 2.27 tq208m-c/1002 4 III 2004 BOP s.c., http://www.bop.com.pl

A New Algorithm for Generation of Decision Trees 1003 Figure 1. Decision tree generated by Czerwiński s algorithm Figure 2. Decision tree generated by Quinlan s algorithm Figure 3. Decision tree generated by the TVR algorithm 5. Conclusions Letusnotethattheerrorrateswhileusingallfouralgorithmswerelow comparedwithe.g.asetofrules[1].thetvralgorithmgeneratesadecisiontreewith an unusually low error of classification of unseen objects(1.20%) requires verification. This algorithm pruned the decision tree by eliminating 4 cases from the source data set. However, using belief networks to improve the process of decision tree generation, we obtained surprisingly good results, better than the results obtained by means tq208m-c/1003 4 III 2004 BOP s.c., http://www.bop.com.pl

1004 J. W. Grzymała-Busse, Z. S. Hippe, M. Knap and T. Mroczek Figure 4. Decision tree generated by the VDP algorithm of Czerwiński s algorithm and/or the classic Quinlan algorithm. In our research we observed that all four algorithms selected the TDS attribute for the root of the decisiontree.ineverycasethisattributehadthesamerangeofvalueafterthe discretization process. Similarly, most of the used algorithms selected the C BLUE attribute as the next test in the process of diagnosis of melanocytic skin lesions. The further selected attributes were different. The mean number of questions and the numberofnodeswithtestsshowaverysimilarcharacter.inconclusion,wecansay that the proposed new algorithm to generate quasi-optimal decision trees, applying Bayesian belief networks, yielded promising results. The algorithm requires further verification, especially in relation to decision tables containing attributes with mixed (numeric and symbolic) values. Acknowledgements FinancialsupportofourresearchprojectNo7T11E03021obtainedfromthe State Committee for Scientific Research(Warsaw) is gratefully acknowledged. References [1] Grzymała-Busse J W and Hippe Z S 2000 Advances in Soft Computing(Intelligent Information Systems)(Kłopotek M, Michalewicz M and Wierzchoń S T, Eds.), Physica-Verlag, Heidelberg, pp. 27 34 [2]HippeZSandBłajdoP2002MethodsofArtificialIntelligence(BurczyńskiT,CholewaW and Moczulski W, Eds.), Silesian University of Technology Edit. Office, Gliwice, pp. 181 185 [3]LauriaEJMandTayiGK2003DataMining:OpportunitiesandChallenges(WangJ,Ed.), Idea Group Publishing, Hershey(PA), pp. 260 277 [4]HippeZS,KnapMandPajaW2002MethodsofArtificialIntelligence(BurczyńskiT, Cholewa W and Moczulski W, Eds.), Silesian University of Technology Edit. Office, Gliwice, pp. 177 180 [5] Hippe Z S and Wrzesień M 2002 Methods of Artificial Intelligence(Burczyński T, Cholewa W and Moczulski W, Eds.), Silesian University of Technology Edit. Office, Gliwice, pp. 185 189 [6] Grzymała-Busse J W 1997 Fundameta Informaticae 31 27 [7]Grzymała-BusseJW,HippeZS,KnapMandMroczekT2003InżynieriaWiedzyiSystemy Ekspertowe(Bubnicki Z and Grzech A, Eds.), Wroclaw University of Technology Edit. Office, Wroclaw 1, pp. 239 247(in Polish) [8] Czerwiński Z 1970 Przegląd Statystyczny 1970/2, PWN, Warsaw(in Polish) [9] Quinlan J R 1993 C4.5 Programs for Machine Learning, Morgan Kaufmann, San Mateo(CA) tq208m-c/1004 4 III 2004 BOP s.c., http://www.bop.com.pl

A New Algorithm for Generation of Decision Trees 1005 [10]HippeZSandMroczekT2003ComputerRecognitionSystems(KurzyńskiM,PuchałaEand Woźniak M, Eds.), Wroclaw University of Technology Edit. Office, Wroclaw, pp. 337 342 [11] Hippe Z S and Knap M 2003 Regułowy algorytm generowania drzew decyzji, Internal Report, Kat. Syst. Ekspertowych i Szt. Inteligencji, WSIZ, Rzeszow(in Polish) [12] Jensen F V 2001 Bayesian Networks and Decision Graphs, Springer-Verlag, Heidelberg [13] Heckerman D??? A Tutorial on Learning with Bayesian Networks, Technical Report MSR- TR-95-06 [14]HippeZS1999TASKQuart.4483 [15] Dąbrowski A 1974 O teorii informacji, WSiP, Warsaw(in Polish) tq208m-c/1005 4 III 2004 BOP s.c., http://www.bop.com.pl

1006 TASK QUARTERLY 8 No 2(2004) tq208m-c/1006 4 III 2004 BOP s.c., http://www.bop.com.pl