DATA MINING DECISION TREES ALGORITHMS OPTIMIZATION

Similar documents
Rule Learning With Negation: Issues Regarding Effectiveness

A Case Study: News Classification Based on Term Frequency

Rule Learning with Negation: Issues Regarding Effectiveness

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Lecture 1: Machine Learning Basics

Python Machine Learning

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Applications of data mining algorithms to analysis of medical data

CS Machine Learning

On-Line Data Analytics

learning collegiate assessment]

Learning From the Past with Experiment Databases

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Assignment 1: Predicting Amazon Review Ratings

Australian Journal of Basic and Applied Sciences

A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and

(Sub)Gradient Descent

Human Emotion Recognition From Speech

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Seminar - Organic Computing

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Theory of Probability

Radius STEM Readiness TM

Mathematics subject curriculum

RECRUITMENT AND EXAMINATIONS

Disambiguation of Thai Personal Name from Online News Articles

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

UNIVERSITY OF CALIFORNIA SANTA CRUZ TOWARDS A UNIVERSAL PARAMETRIC PLAYER MODEL

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Mining Association Rules in Student s Assessment Data

Probabilistic Latent Semantic Analysis

Delaware Performance Appraisal System Building greater skills and knowledge for educators

Diploma in Library and Information Science (Part-Time) - SH220

Calibration of Confidence Measures in Speech Recognition

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Massachusetts Department of Elementary and Secondary Education. Title I Comparability

South Carolina English Language Arts

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Mathematics. Mathematics

User education in libraries

San Marino Unified School District Homework Policy

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Some Principles of Automated Natural Language Information Extraction

Modeling function word errors in DNN-HMM based LVCSR systems

Task Types. Duration, Work and Units Prepared by

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Word Segmentation of Off-line Handwritten Documents

MYP personal project guide 2011 overview of objectives

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Science Fair Project Handbook

K-Medoid Algorithm in Clustering Student Scholarship Applicants

Linking Task: Identifying authors and book titles in verbose queries

Pp. 176{182 in Proceedings of The Second International Conference on Knowledge Discovery and Data Mining. Predictive Data Mining with Finite Mixtures

GACE Computer Science Assessment Test at a Glance

IT Students Workshop within Strategic Partnership of Leibniz University and Peter the Great St. Petersburg Polytechnic University

Improving Simple Bayes. Abstract. The simple Bayesian classier (SBC), sometimes called

WHEN THERE IS A mismatch between the acoustic

Modeling function word errors in DNN-HMM based LVCSR systems

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Computerized Adaptive Psychological Testing A Personalisation Perspective

Chapter 2 Rule Learning in a Nutshell

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Indian Institute of Technology, Kanpur

Lecture 1: Basic Concepts of Machine Learning

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Learning Methods in Multilingual Speech Recognition

Designing a case study

Miami-Dade County Public Schools

10.2. Behavior models

PhD in Computer Science. Introduction. Dr. Roberto Rosas Romero Program Coordinator Phone: +52 (222) Ext:

FRAMEWORK FOR IDENTIFYING THE MOST LIKELY SUCCESSFUL UNDERPRIVILEGED TERTIARY STUDY BURSARY APPLICANTS

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

An Effective Framework for Fast Expert Mining in Collaboration Networks: A Group-Oriented and Cost-Based Method

ON BEHAVIORAL PROCESS MODEL SIMILARITY MATCHING A CENTROID-BASED APPROACH

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Laboratorio di Intelligenza Artificiale e Robotica

Analyzing sentiments in tweets for Tesla Model 3 using SAS Enterprise Miner and SAS Sentiment Analysis Studio

Statewide Framework Document for:

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation

SARDNET: A Self-Organizing Feature Map for Sequences

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

A Comparison of Standard and Interval Association Rules

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

CLASSROOM USE AND UTILIZATION by Ira Fink, Ph.D., FAIA

Softprop: Softmax Neural Network Backpropagation Learning

Problems of the Arabic OCR: New Attitudes

Transcription:

UNIVERSITY OF CRAIOVA Faculty of Automation, Computers and Electronics Ph.D. Student Laviniu Aurelian Bădulescu - PH.D. THESIS ABSTRACT - DATA MINING DECISION TREES ALGORITHMS OPTIMIZATION Supervisor: Acad. Prof. Univ. Dr. Ing. Mircea Petrescu 1. TABLE OF CONTENTS (Excerpts) Chapter 1. INTRODUCTION 1.1. Knowledge discovery from databases and data mining 1.2. Classification problem 1.3. Decision trees 1.4. Advantages and disadvantages of decision trees 1.5. Thesis organisation 1.6. Personal contribution Chapter 2. THE EXISTING LITERATURE PRESENTATION ABOUT DECISION TREES 2.1. AID, THAID, CHAID and FIRM algorithms 2.2. CART algorithms 2.3. ID3, C4.5 and other related algorithms 2.4. Lazy decision trees 2.5. Decision trees algorithms from SAS package 2.6. Multidimensional decision trees algorithms 2.7. Oblique multivariate decision trees algorithms 2.8. Omnivariate or mixed decision trees algorithms 2.9. Oblivious decision trees algorithms 2.10. Parallel decision trees algorithms 2.11. Decision trees algorithms from WEKA package 2.12. Decision trees with genetic algorithms 2.13. Decision trees algorithms for data streams 2.14. Orthogonal decision trees algorithms I

2.15. Hybrid decision trees 2.16. Other decision trees algorithms 2.17. Online adaptive decision trees 2.18. Distributed decision trees algorithms Chapter 3. DECISION TREE'S INDUCTION AND PRUNING ALGORITHMS AND SPLITTING CRITERIA USED IN EXPERIMENTS 3.1. Decision tree's induction and pruning algorithms 3.2. Splitting criteria used in experiments 3.2.1. Impurity based criteria used in experiments 3.2.2. Normalized impurity based criteria used in experiments 3.2.3. Binary criteria used in experiments 3.2.4. K2 and Bayes-Dirichlet criteria used in experiments 3.3. Decision trees pruning methods used in experiments 3.3.1. Cost complexity pruning 3.3.2. Reduced error pruning 3.3.3. Minimum error pruning 3.3.4. Pessimistical pruning used in experiments 3.3.5. Error based pruning or confidence level pruning used in experiments 3.3.6. Optimal pruning 3.3.7. Minimum description length pruning 3.4. Decision rules used in experiments 3.5. Handling attributes with missing values in experiments Chapter 4. MATERIALS AND METHODS USED IN EXPERIMENTS 4.1. Databases used in experiments 4.1.1. General information about Abalone database used in experiments 4.1.2. General information about Cylinder Bands database used in experiments 4.1.3. General information about Image Segmentation database used in experiments 4.1.4. General information about Iris database used in experiments 4.1.5. General information about Monk s Problem database used in experiments 4.1.6. General information about Adult database used in experiments 4.1.7. General information about Census Income database used in experiments 4.1.8. General information about Forest Covertype database used in experiments 4.2. The method used in experiments. Software package 4.2.1. Splitting criteria used in experiments 4.2.2. Decision tree induction program used in experiments 4.2.3. Decision tree pruning program used in experiments 4.2.4. Decision tree classification accuracy testing program used in experiments 4.2.4.1. Confusion matrix determination program used in experiments 4.2.5. Decision rules set generation program used in experiments Chapter 5. EXPERIMENTAL RESULTS AND DISCUSSIONS 5.1. Experiments on Abalone database 5.1.1. Experiments with decision trees induction based on miscellaneous measures 5.1.1.1. Experiments with decision rules extraction from unpruned decision trees 5.1.1.2. Experiments on unpruned decision trees execution 5.1.1.2.1. Experiments with decision trees classification accuracy on training dataset 5.1.1.2.2. Experiments with decision trees classification accuracy on test dataset II

5.1.2. Experiments with decision trees pruning with confidence level pruning method 5.1.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.1.2.2. Experiments on confidence level pruned decision trees execution on test dataset 5.1.3. Experiments with decision trees pruning with pessimistic pruning method 5.1.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees 5.1.3.2. Experiments on pessimistic pruned decision trees execution on test dataset 5.1.4. Decision rules number and classification error rate on test dataset for three types of decision trees 5.1.5. Related experiments on Abalone database 5.2. Experiments on Cylinder bands database 5.2.1. Experiments with decision trees induction based on miscellaneous measures 5.2.1.1. Experiments with decision rules extraction from unpruned decision trees 5.2.1.2. Experiments on unpruned decision trees execution on test dataset 5.2.2. Experiments with decision trees pruning with confidence level pruning method 5.2.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.2.2.2. Experiments on confidence level pruned decision trees execution on test dataset 5.2.3. Experiments with decision trees pruning with pessimistic pruning method 5.2.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees 5.2.3.2. Experiments on pessimistic pruned decision trees execution on test dataset 5.2.4. Decision rules number and classification error rate on test dataset for three types of decision trees 5.2.5. Related experiments on Cylinder Bands database 5.3. Experiments on Statlog. Image Segmentation /Satimage database 5.3.1. Experiments with decision trees induction based on miscellaneous measures 5.3.1.1. Experiments with decision rules extraction from unpruned decision trees 5.3.1.2. Experiments on unpruned decision trees execution on test dataset 5.3.2. Experiments with decision trees pruning with confidence level pruning method 5.3.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.3.2.2. Experiments on confidence level pruned decision trees execution on test dataset 5.3.3. Experiments with decision trees pruning with pessimistic pruning method 5.3.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees 5.3.3.2. Experiments on pessimistic pruned decision trees execution on test dataset 5.3.4. Computing confusion matrix based on experimental results 5.3.5. Decision rules number and classification error rate on test dataset for three types of decision trees 5.3.6. Related experiments on Image Segmentation database 5.4. Experiments on Iris database 5.4.1. Experiments with decision trees induction based on miscellaneous measures 5.4.1.2. Experiments with decision rules extraction from unpruned decision trees 5.4.2. Experiments with decision trees pruning with confidence level pruning method 5.4.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.4.3. Experiments with decision trees pruning with pessimistic pruning method 5.4.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees 5.4.4. Decision rules number for three types of decision trees 5.5. Experiments on Monk s problem database 5.5.1. Experiments with decision trees induction based on miscellaneous measures 5.5.1.1. Experiments with decision rules extraction from unpruned decision trees 5.5.1.2. Experiments on unpruned decision trees execution on test dataset 5.5.2. Experiments with decision trees pruning with confidence level pruning method 5.5.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.5.2.2. Experiments on confidence level pruned decision trees execution on test dataset 5.5.3. Experiments with decision trees pruning with pessimistic pruning method 5.5.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees III

5.5.3.2. Experiments on pessimistic pruned decision trees execution on test dataset 5.5.4. Decision rules number and classification error rate on test dataset for three types of decision trees 5.5.5. Related experiments on Monk 1 database 5.6. Experiments on Adult database 5.6.1. Experiments with decision trees induction based on miscellaneous measures 5.6.1.1. Experiments with decision rules extraction from unpruned decision trees 5.6.1.2. Experiments on unpruned decision trees execution on test dataset 5.6.2. Experiments with decision trees pruning with confidence level pruning method 5.6.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.6.2.2. Experiments on confidence level pruned decision trees execution on test dataset 5.6.3. Experiments with decision trees pruning with pessimistic pruning method 5.6.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees 5.6.3.2. Experiments on pessimistic pruned decision trees execution on test dataset 5.6.4. Decision rules number and classification error rate on test dataset for three types of decision trees 5.6.5. Related experiments on Adult database 5.7. Experiments on Census Income database 5.7.1. Experiments with decision trees induction based on miscellaneous measures 5.7.1.1. Experiments with decision rules extraction from unpruned decision trees 5.7.1.2. Experiments on unpruned decision trees execution on test dataset 5.7.2. Experiments with decision trees pruning with confidence level pruning method 5.7.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.7.2.2. Experiments on confidence level pruned decision trees execution on test dataset 5.7.3. Experiments with decision trees pruning with pessimistic pruning method 5.7.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees 5.7.3.2. Experiments on pessimistic pruned decision trees execution on test dataset 5.7.4. Decision rules number and classification error rate on test dataset for three types of decision trees 5.7.5. Related experiments on Census Income database 5.8. Experiments on Forest Covertype database 5.8.1. Experiments with decision trees induction based on miscellaneous measures 5.8.1.1. Experiments with decision rules extraction from unpruned decision trees 5.8.1.2. Experiments on unpruned decision trees execution on test dataset 5.8.2. Experiments with decision trees pruning with confidence level pruning method 5.8.2.1. Experiments with decision rules extraction from confidence level pruned decision trees 5.8.2.2. Experiments on confidence level pruned decision trees execution on test dataset 5.8.3. Experiments with decision trees pruning with pessimistic pruning method 5.8.3.1. Experiments with decision rules extraction from pessimistic pruned decision trees 5.8.3.2. Experiments on pessimistic pruned decision trees execution on test dataset 5.8.4. Decision rules number and classification error rate on test dataset for three types of decision trees 5.8.5. Related experiments on Forest Covertype database 5.8.6. Experiments to improve the classification accuracy on test data 5.8.6.1. Experiments on the reversal training dataset with test dataset Chapter 6. CONCLUSIONS OF THE EXPERIMENTS AND FUTURE DIRECTIONS 6.1. Conclusions of the experiments upon the performances of the miscellaneous measures used for decision trees induction 6.2. Conclusions of the experiments upon the performances of the pruning methods 6.3. Directions for future experiments References IV

2. KEYWORDS: data mining, classification problem, decision trees, training dataset, test dataset, class, class labels, splitting criteria, knowledge discovery, decision rules, unpruned decision tree, pruning methods, pruned decision tree, classification error rate, classification accuracy. 3. SYNTHESIS OF THE PH.D. THESIS MAIN PARTS Ph.D. thesis objectives targeted: Decision Trees (DT) algorithms optimisation, in order to find optimal splitting criteria and optimal pruning methods; developing a unified algorithmically framework for approaching the DT algorithms, the splitting criteria and the pruning methods; developing an experimental framework where, based on proposed software package for induction, pruning and execution of the DT, for computing confusion matrix and associated decision rules, the research to be done. The Ph.D. thesis is divided into six chapters preceded by a table of contents and followed by References. The table of contents is structured in accordance with the domain s themes and it includes the following parts: Introduction (Chapter 1), State of knowledge (Chapter 2), Own contributions (Chapter 3, Chapter 4, Chapter 5), Final conclusions (Chapter 6) and References. In Chapter 1, Introduction, an introduction is realized to establish the theme treated. Thus, it shows the relation between Knowledge Discovery from Databases and Data Mining (DM), the classification problem, basic concepts surrounding DT, as well as advantages and disadvantages of DT algorithms. At the end of the chapter, it presents how the thesis is organized and the original contributions of its author. One of the main personal contributions on a theoretical level is the construction of a unified frame that unitarily treats several dozen splitting criteria and, in the same mathematical framework, the presentation of a few pruning methods. Considering that the main approaches of optimization of DT algorithms must follow two directions: choosing an optimal splitting criterion and the best pruning method, the major personal contribution of the thesis on a practical level is represented by the discovery, based on a large number of personal experiments, of a splitting criterion that systematically shows the best performances, compared to dozens of other criteria considered for experimentation. At the same time, the confidence level pruning DT method was demonstrated to be superior. Thus, new DT algorithms have been conceived, optimized, representing a personal contribution of the thesis. Another key contribution of the PhD thesis was the development of an experimental framework which, based on a software package meant for the induction, pruning and execution of the DT, for generating the confusion matrix and the decision rules associated to DT, made numerous experiments using highly differing databases. The experimental results were compared with many results of other researchers that have experimented on the same databases, but have used different DM algorithms, finding that personal experimental results are just as accurate as the best literature values. Another major personal contribution is the study in Chapter 2, the result of an exhaustive and up-to-date browsing of an enormous volume of works in the DT domain, in conjunction with a sustained effort of synthesizing and systematization of the reading material. This chapter shows a personal synthesis of the main DT algorithms, accomplished exclusively based to bibliography from foreign literature. V

Chapter 2, The existing literature presentation about Decision Trees, makes a personal synthesis of the main DT algorithms, obtained as a result of browsing and systematization of an important bibliography of foreign literature. Thus, there are numerous DT algorithms present: the AID, THAID, CHAID and FIRM algorithms, the CART algorithm, the ID3, C4.5 and other related algorithms, lazy DT, DT algorithms from SAS package, multidimensional DT algorithms, oblique multivariate DT algorithms, omnivariate DT algorithms, oblivious DT algorithms, parallel DT algorithms, DT algorithms from WEKA package, DT with genetic algorithms, DT algorithms for data streams, orthogonal DT algorithms, hybrid DT, online adaptive DT, distributed DT algorithms as well as other DT algorithms. Chapter 3, Decision Tree's induction and pruning algorithms and splitting criteria used in experiments, are an original contribution, proposing a unified algorithmic framework to address the DT algorithms and providing a thorough description of dozens of attribute selection methods for splitting the dataset corresponding to a node and several DT pruning methods (cost complexity pruning, reduced error pruning, minimum error pruning, pessimistical pruning, error based pruning or confidence level pruning, optimal pruning, minimum description length pruning). Chapter 3 presents proposed algorithms for induction and pruning of DT, impurity based criteria, normalized impurity based criteria, binary criteria, K2 and Bayes-Dirichlet criteria later used in experiments. At the same time, it presents the method of transformation of DT in the decision rules set and the way of treating missing values attributes. Chapter 4, Materials and methods used in experiments, presents the 8 databases used in the experiments, whose dimensions are close to 600.000 cases, 55 attributes (continuous and nominal), 7 label classes, numerous cases with missing values and duplicate or contradictory records. Also in this chapter the proposed and utilized in experiments software package is presented: the DT induction program, the DT pruning program, the program for testing DT classification accuracy, the one for calculating the confusion matrix and the program for generating the decision rules associated to DT. Chapter 5, Experimental results and discussions, presents the experiments performed for each of the 8 databases based on a set of identical processing. The performances of each step of processing were shown in tables and charts with comments at every step of the experiments. 1. The first step of the experiments was the DT induction based on the training dataset. In this step there were induced 28 DT, corresponding to the 28 attribute selection measures. Were taken into account the values of parameters of the building process of DT, such as: the number of nodes of the induced DT, the number of attributes necessary for the DT induction, the number of levels of the DT, the DT growth time, the size of the file containing the inducted DT. These values were compared and discussed, statistical indicators were calculated, tables and charts with comments were presented, extracting conclusions on the different behavior of the 28 DT induced with the 28 measures. 2. In the second step of the experiments, these 28 DT, in the unpruned form, were processed in order to extract from them the 28 sets of decision rules, one set for each DT obtained in step 1. Were taken into account the values of parameters of the DT decision rules extraction process, such as: the decision rules number, the time required to build the file containing the decision rules, the time required to read the file containing the DT, the size of the file containing the decision rule set. These values were compared and discussed, statistical indicators were calculated, tables and charts with comments were shown, and extracting conclusions on the different behavior of the 28 DT induced with the 28 attribute selection measures. 3. The third step of the experiments, accomplished on each database, involved the execution of the 28 DT induced in the first step. These experiments verified the classification model represented by the DT based on a test dataset, unknown dataset in step 1, when DT was generated. It is the most important step of the experiments, because now the values of the VI

classification error rate on test data or model accuracy are obtained. For each database 28 values of the classification accuracy of the unpruned DT on test data are obtained, corresponding to each of the 28 measures with which the DT was induced. By comparing these values we can draw conclusions about the performance of every attribute selection measures. 4. The fourth step of the experiments targeted the pruning of the 28 DT obtained in step 1, with the confidence level pruning method. Were taken into account the values of some parameters of the DT pruning process: the confidence level, the number of attributes necessary to build the pruned DT, the number of nodes and the number of levels of the pruned DT, the pruning time of the file containing the unpruned DT, the size of the file containing the pruned DT. These values were compared and discussed, statistical indicators were calculated, tables and charts with comments were shown, and extracting conclusions on the different behavior of the 28 confidence level pruned DT. 5. The fifth step of the experiments involved the extraction of the decision rules from the 28 DT reduced with the confidence level pruning method in step 4. The processing resembles those in step 2. 6. The sixth step involved the execution on test data of the 28 confidence level pruned DT obtained on step 4. The values of the classification error rate on the test data were retained, values which were later compared with values obtained in the execution of the unpruned DT on test data (see step 3) and with values obtained in the execution of the pessimistic pruned DT on test data (see step 7). 7. The seventh step of the experiments involved the pruning of the 28 DT, obtained in step 1, with the pessimistic pruning method. The processing resembles those in step 4. 8. The eighth step of the experiments involved the extraction of the decision rules from the 28 DT pessimistically pruned in step 7. The processing resembles those in step 2. 9. The ninth step involved the execution on the test data of the 28 DT pessimistically pruned in step 7. The processing resembles those in steps 3 and 6. 10. In the tenth step a discussion was realized on the values of the number of decision rules and the classification error rate values on test data for the three types of DT: unpruned, pruned with confidence level method and pruned with pessimistic pruning method, showing which attribute selection measure and which pruning method had the best behavior in the experiments. 11. Finally, in the eleventh step, the results of some representative works from literature were reviewed, presenting performance tests on each database and, in the same time, we made a comparison of the results obtained in those experiments with the results obtained by the 28 criteria and the two pruning methods. For reasons of space, the confusion matrix was calculated for only one database (Image). In Chapter 6, Conclusions of the experiments and future directions are presented. Thus, three techniques are used to compare the classification performances of the 28 induced and pruned DT with two pruning methods on the 8 databases. The first time the comparison is realized based on the values of the arithmetic mean and based on the values of the standard deviation of the classification error rate on test data for all types of DT, all databases and all attribute selection measures. Finds that the best classify DT induced with the rciq (quadratic information gain ratio) measure, with an average of classification error rate on test data of 18.59% (also having the smallest value for standard deviation: 14.11), followed by the DT induced with rel (relief), k2, bd (Bayes-Dirichlet), lmdfa (minimum description length in absolute frequencies) and csh (stochastic complexity) measure with small values for standard deviation, while the smallest average performance is achieved by DT induced with ciqp (balanced quadratic information gain) measure, with an average value of 31.12%. VII

DT induced by rciq measure has the best value of the mean classification error rate and the smallest standard deviation, resulting in the smallest spread of classification error rate values on test data around their mean. This result shows that the performance of DT induced by rciq measure is not affected too much by the features of the database, i.e. one of the fundamental characteristics of the rciq measure is its independence from the domain, a very important feature today, when databases are made up of data, with attributes pertaining to different domains, collected together. The diversity of domains from the composition of the databases is one of the reasons that have increased the need to use tools for automated knowledge discovery from databases and inductive learning algorithms. The results of the experiments show that the smallest mean value (1,147.54) of the number of decision rules is achieved by the DT induced by mapd measure, and the worst mean performance value (1,641.25) is achieved by the DT induced by rcs measure. To note that the DT induced by rciq measure, which has the best behavior regarding classification accuracy on test data, is placed on fourth spot in the hierarchy of the smallest mean values for the number of decision rules. The second comparison method of the classification performance values of the 28 DT was represented by the win/tie/loss technique. Also in this case, the performances of the DT induced by rciq measure are always the best. The DT induced by rciq measure presents the largest number of cases with a better performance value than all the other DT induced by the other measures and at the same time, the smallest number of cases with a lower performance value than the other DT. The next five places are occupied by induced DT with another 5 measures csh, bd, k2, lmdfa and rel, which no longer present the property shown by the rciq measure, of having both a large number of cases when they overcome the performance values of induced DT with all other measures, and a small number of cases when they are overcome by the performance values of the other DT. The third comparison method of the classification performance values of the 28 DT was represented by the geometric mean of the classification error rate ratio. The performances of the DT induced by rciq measure surpass, in all the cases considered in the experiments, the performances of the DT induced with the other 27 measures. Is followed, in order, by the performance values achieved by the DT induced with the rel, gim (modified gini index), bd, k2 lmdfa and csh measures. Thus, we can conclude that, regardless of the comparison criterion considered, optimizing the DT algorithms by using the rciq measure induces the most efficient DT. Even the optimization based on the rel, k2, bd, lmdfa and csh measures produces DT with good performances. Regarding the search for the best pruning method, analysis of the values obtained from previous experiments presented in the thesis shows that the DT pruned with the confidence level pruning method systematically obtains the best values for classification accuracy on test data, simultaneously generating a small number of decision rules. In second place are the performances of the pessimistically pruned DT and in last place the unpruned DT performances. Of note is this differentiation, which always places confidence level pruned DT in first place and the pessimistically pruned DT in second place, it is much clearer in regard to the number of decision rules than the classification accuracy on test data. At the same time, of further note is that the pruning improves the DT performances. The experiments that we wish to carry out in the near future will involve much larger databases, with many attributes and cases, which we wish to verify the conclusions obtained for the databases utilized until now. At the same time, we will include in the set of measures on the basis of which the splitting of a node is realized other the attribute selection criteria and we will utilize other DT pruning methods, outside of the two used in our experiments. At the same time, we will have to take into account the information provided by the confusion matrix. VIII