Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Similar documents
Rule Learning With Negation: Issues Regarding Effectiveness

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Python Machine Learning

Learning From the Past with Experiment Databases

Australian Journal of Basic and Applied Sciences

Rule Learning with Negation: Issues Regarding Effectiveness

A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and

Lecture 1: Machine Learning Basics

CS Machine Learning

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Learning Methods for Fuzzy Systems

Applications of data mining algorithms to analysis of medical data

Learning Methods in Multilingual Speech Recognition

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Data Fusion Through Statistical Matching

BENCHMARK TREND COMPARISON REPORT:

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

Mining Association Rules in Student s Assessment Data

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

MMOG Subscription Business Models: Table of Contents

Specification of the Verity Learning Companion and Self-Assessment Tool

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

A Case Study: News Classification Based on Term Frequency

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Matching Similarity for Keyword-Based Clustering

Word Segmentation of Off-line Handwritten Documents

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

Analyzing the Usage of IT in SMEs

Software Maintenance

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

Generative models and adversarial training

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

(Sub)Gradient Descent

On-Line Data Analytics

Massachusetts Department of Elementary and Secondary Education. Title I Comparability

Implementing a tool to Support KAOS-Beta Process Model Using EPF

Abstractions and the Brain

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

The Use of Statistical, Computational and Modelling Tools in Higher Learning Institutions: A Case Study of the University of Dodoma

National Longitudinal Study of Adolescent Health. Wave III Education Data

Evolutive Neural Net Fuzzy Filtering: Basic Description

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Ordered Incremental Training with Genetic Algorithms

GRADUATE STUDENT HANDBOOK Master of Science Programs in Biostatistics

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Empirical research on implementation of full English teaching mode in the professional courses of the engineering doctoral students

The CTQ Flowdown as a Conceptual Model of Project Objectives

Conference Presentation

Computerized Adaptive Psychological Testing A Personalisation Perspective

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Measurement & Analysis in the Real World

arxiv: v1 [cs.cl] 2 Apr 2017

ACADEMIC AFFAIRS GUIDELINES

Setting Up Tuition Controls, Criteria, Equations, and Waivers

Universidade do Minho Escola de Engenharia

Welcome to. ECML/PKDD 2004 Community meeting

STA 225: Introductory Statistics (CT)

SARDNET: A Self-Organizing Feature Map for Sequences

Probabilistic Latent Semantic Analysis

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Analysis: Evaluation: Knowledge: Comprehension: Synthesis: Application:

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Fragment Analysis and Test Case Generation using F- Measure for Adaptive Random Testing and Partitioned Block based Adaptive Random Testing

Detailed Instructions to Create a Screen Name, Create a Group, and Join a Group

What is related to student retention in STEM for STEM majors? Abstract:

A NEW ALGORITHM FOR GENERATION OF DECISION TREES

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Softprop: Softmax Neural Network Backpropagation Learning

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

National Survey of Student Engagement at UND Highlights for Students. Sue Erickson Carmen Williams Office of Institutional Research April 19, 2012

Detecting English-French Cognates Using Orthographic Edit Distance

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Introduction to Questionnaire Design

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

DegreeWorks Advisor Reference Guide

Mathematics Program Assessment Plan

Assignment 1: Predicting Amazon Review Ratings

Disambiguation of Thai Personal Name from Online News Articles

Model Ensemble for Click Prediction in Bing Search Ads

Dinesh K. Sharma, Ph.D. Department of Management School of Business and Economics Fayetteville State University

Major Milestones, Team Activities, and Individual Deliverables

A Case-Based Approach To Imitation Learning in Robotic Agents

A study of speaker adaptation for DNN-based speech synthesis

PROJECT DESCRIPTION SLAM

The development and implementation of a coaching model for project-based learning

Problems of the Arabic OCR: New Attitudes

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

A Context-Driven Use Case Creation Process for Specifying Automotive Driver Assistance Systems

learning collegiate assessment]

Linking Task: Identifying authors and book titles in verbose queries

Transcription:

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees Mariusz Łapczy ski 1 and Bartłomiej Jefma ski 2 1 The Chair of Market Analysis and Marketing Research, Cracow University of Economics, Cracow, Poland lapczynm@uek.krakow.pl 2 The Chair of Econometrics and Computer Science, Wrocław University of Economics, Wrocław, Poland bartlomiej.jefmanski@ue.wroc.pl Abstract. Building predictive models in customer relationship management refers to each stage in the customer s lifecycle, i.e. the customer acquisition, development and retention. One may notice that the construction of predictive models is more and more frequently accompanied by an attempt to combine analytical tools of the same type or combining various methods. The first approach is named ensemble methods, while the second can be referred to as a hybrid one. Authors decided to combine k-means algorithm with decision trees and examine whether cluster validity measures influence the performance of a model. During experiments 5 different cluster validity indices and 8 datasets were used. The performance of models was evaluated by using popular measures such as: accuracy, precision, recall, G-mean, F-measure and lift in the first and in the second decile. The results are far from causing enthusiasm, however, they are promising in some fields. Keywords: hybrid predictive models, k-means, C&RT, cluster validity measures, performance of models 1 Introduction This article aims at verifying in what manner the measures indicating the optimal number of clusters influence the quality of hybrid predictive models combining the k- means algorithm with classification and regression trees (C&RT). Combining the clustering analysis with decision trees has recently become a popular method of increasing the performance of predictive models. Research studies covering this area pertain to numerous disciplines, such as customer relationship management, web usage mining, medical sciences, petroleum geology, anomalies in computer networks, etc. The inspiration to undertake the subject came from the successful experiment referring to [1] profiling users clicking on the banner ad of a cosmetics company, which was placed on a social networking website which was popular in Poland. 153

The construction of predictive models in customer relationship management refers to each stage in the customer s lifecycle, i.e. the customer acquisition, development and retention. In these areas one frequently applies supervised methods such as decision trees, neural networks, Random Forest, boosted trees, logistic regression, discriminant analysis, etc. Generally, the analyst s target is the construction of such a model that will in the best possible way anticipate the customer s sense of belonging to a particular category of the dependent variable (potential customer, potential churner, etc.). One may notice that the construction of predictive models is more and more frequently accompanied by an attempt to combine analytical tools of the same type and create the so-called ensemble models, also referred to as committees. There are also attempts combining various methods, which are described with the terms hybrid, two stage classification, cascade classification or cross-algorithm ensemble. In numerous cases such combined attempts permitted to achieve a better performance. The authors have decided to conduct an experiment consisting in combining the k- means algorithm with decision trees (C&RT). While creating clusters they implemented 5 different cluster validity measures and observed in what manner the number of clusters influences the performance of the model. The analysis was carried out on 8 data sets collected from publicly accessible repositories. The dependent variable in each dataset possessed two categories, and the set itself as much as possible pertained to the broadly understood marketing activities of a company. In the first section there appears a brief review of the literature in which one combined clustering with decision trees during the construction of predictive models. The second section contains a description of model hybridization, characteristics of cluster validity indices as well as characteristics of the implemented datasets. Section III will present the results of the experiment alongside with the performance evaluation. Section IV contains the summary and proposals regarding the successive experiments in this area. 2 Hybrid Predictive Models Based on Clustering and Decision Trees Literature Review Combining clustering with decision trees for building predictive models has long been of interest to many researchers. It seems that in the field of marketing churn modeling has become popular in recent years. Some authors [2] combined the results obtained from clustering algorithms (k-means, k-medoid, self-organizing maps (SOM), fuzzy c-means and Balanced Iterative Reducing and Clustering using Hierarchies (BIRCH)) with the results obtained from the decision tree (C5.0) with boosting. Their goal was to predict the customer churn. Two methods of hybridization were examined. In the first approach a new variable was added whose categories informed about the cluster membership while building the decision tree. In the second approach different decision trees were separately for each cluster. Chu et al [3] proposed a hybrid model to predict churning in the area of customer relationship management. C5.0 decision trees and Growing Hierarchical Self- 154

Organizing Map (GHSOM) were combined. In the first step the predictive model was constructed on the basis of such independent variables as: defection history, deactivation data, payment history, usage patterns etc. In the second step GHSOM was applied to build four disjoint clusters containing churners. Another slightly differing approach of constructing hybrid models was called the model of two-step classification and was proposed [4] as an alternative approach for churn modeling in the security industry. In the first stage of the procedure selforganizing maps (SOM) were used to divide customers into 9 clusters. In the next step the authors chose the largest segment with the highest churn rate. In the last stage a decision was that provided a high accuracy of classification. The combination of the K-means algorithm and decision trees (ID3) was also used in the classification of anomalies in computer networks [5]. The approach of joining these two machine learning algorithms was called the cascade one. Hybrid models combining the cluster analysis with decision trees are also referred to as integrated ones. An example of such an approach was an attempt at predicting heart diseases [6], in which the dataset was divided into clusters by using k-means algorithm, and afterwards one decision was for each cluster. The authors investigated the impact of different initial centroid selection methods on the performance of decision trees. 3 Description of Hybridization and s 3.1 Hybridization Authors treat building a hybrid model as a sequential combination of unsupervised and supervised models. Another reason for naming this approach "hybrid" is a combination of classical statistical tools (k-means method) with the algorithm derived from data mining (C&RT). In the first stage objects were clustered by using the k-means algorithm. In the second stage C&RT algorithm was applied, treating cluster membership of the objects as a new independent variable. As the experiment involved the application of eight different datasets, the authors made an attempt to unify the procedure. With the lack of knowledge of the research problems, it was decided that the set of variables utilized during the analysis of clusters will refer exclusively to numerical variables. The new categorical variable informing about the class membership was then attached to the remaining categorical variables, and the set completed in such a way constituted the basis for building a decision tree. Therefore, it may be assumed that the cluster analysis played here the role of the method of reducing the number of independent variables, and ultimately was intended to facilitate the interpretation of the model. Data mining, apart from psychology, biology, statistics and machine learning, constitutes one of the most important areas in which the methods of cluster analysis are widely applied. Different variants of the k-means method result from the manner in which the initial positions of centroids are determined, the way of calculating centroids in successive steps of the algorithm, or the implemented measure of distance. In 155

this work authors applied the Hartigan and Wong method [7], available in the R package stats. A characteristic feature of the methods optimizing the initial partition of objects is determining a priori the number of clusters. One of the ways of conduct in this area is establishing this number on the basis of classification quality measures. However, as emphasized by Everit et al. [8], the selection of the optimal number of clusters should result from the synthesis of results obtained with the help of different methods. Such a conduct is justified by e.g. the fact that each of the methods is based on predefined assumptions referring to the structure of classes, which not always must be satisfied. Therefore, in this analysis we applied several measures frequently implemented in empirical research studies and available in the R package clustersim: the Calinski-Harabasz index (CH), the Krzanowski-Lai index (KL), the Davies and Bouldin index (DB), the Hartigan index (H), the Gap Statistic (Gap). Classification and Regression Trees (CART), which was developed by Breiman et al [9], is a recursive partitioning algorithm. It is used to build a classification tree if the dependent variable is nominal, and a regression tree if the dependent variable is continuous. Decision trees usually do not have high predictive power, however, they deliver a set of rules and a graphical model that can be helpful in understanding the problem. The experiment involved the application of the C&RT algorithm with equal a priori probabilities and equal misclassification costs. The minimal number of instances in terminal nodes was established at the level of 10% of the learning sample. 3.2 s The authors did their best to ensure that the datasets applied in the experiment refer to the marketing activity of companies. For this purpose they utilized popular repositories selecting datasets with a binary target variable. The first dataset referred to direct marketing campaigns of a Portuguese banking institution [10]. The dependent variable in the second dataset (German Credit) was related to good or bad credit risks [11]. The third dataset was used in the CoIL 2000 Challenge [12]. It was related to predicting the willingness to purchase a caravan insurance policy. The fourth dataset referred to direct marketing and was used during KDD Cup 1998. The file was hosted on http://kdd.ics.uci.edu by I. Parsa and K. Howes. The fifth dataset included target variable churn and 20 independent variables [13]. The sixth dataset was also related to churn modeling and was used during KDD Cup in 2009 [14]. The seventh dataset (CINA) consisted of census data [15]. The binary dependent variable indicated whether the income exceeds 50,000. The last dataset referred to credit card applications [11] with the binary target variable and a set of 14 independent variables. The characteristics of all datasets, including size, number and kind of independent variables as well as the percentage of category 1 of the dependent variable was illustrated in Table 1. Table 1. Characteristics of datasets applied in experiment Number of cases Number of independent variables Percentage of category 1 of depend- 156

D1 D2 D3 Bank Marketing Data Set Statlog (German Credit) Insurance Company Benchmark 45,211 1,000 5,822 D4 KDD 1998 95,412 Churn 5,000 D6 KDD 2009 50,000 D7 D8 CINA Marketing Data Set Statlog (Australian Credit) 16,033 690 7 numerical 9 categorical 7 numerical 12 categorical 80 numerical or binary 5 categorical 286 numerical 194 categorical and dates 16 numerical 4 categorical including phone number 190 numerical 39 categorical 21 numerical 111 binary 6 numerical 8 categorical ent variable 11.70% 30.00% 5.98% 5.08% 14.14% 7.34% 24.57% 44.49% Each set of observations was divided into the learning sample (70%) and the test sample (30%). In order to make the cluster interpretation simpler the number of variables applied while clustering could not exceed 15. If the dataset consisted of their larger number, the feature selection was undertaken with the help of Random Forest. The variables for which the amount of missing data exceeded 10% as well as the cases for which the missing data exceeded 50% were removed from the sets. Categorical variables with a very large number of categories were grouped with the help of selforganizing maps and introduced into the analysis as an additional independent variable. In cases where data were missing mean or mode were applied instead. The variables referring to ID, phone numbers and dates were excluded from the analysis. 4 Results of experiment It was decided before the initiation of the clustering procedure that the number of clusters cannot be larger than 15. Table 2 illustrates the optimal number of subgroups which was indicated by particular cluster validity measures. Lines (-) mean that in the range from 2 to 15 clusters no optimal number of classes was indicated by the measurement. It seems that the Davies-Bouldin index has a tendency to differentiate the highest number of clusters. On the other hand, the Hartigan index indicated the smallest number of subgroups or could not find an optimal solution at all. Hence, eventually 5 hybrid models were on the basis of eight datasets. Table 2. Number of clusters indicated by particular cluster validity measures Cluster validity measures CH KL DB H Gap D1 6 6 9 2 3 D2 15 11 7-6 157

D3 2 15 2 - - D4 2 6 12 2 2 2 5 12 2 15 D6 4 12 15 4 - D7 2 4 14 2 4 D8 5 8 11-2 The following popular performance measures were utilized for the assessment of models: accuracy, recall, precision, G-mean, F-measure, and lift in the first and in the second decile. The successive tables (3-9) contain the results for eight datasets taken into account in the experiment as well as for 6 models. Five out of six decision tree models were modified by adding new categorical variables, while the sixth model remained unmodified. It was on the basis of the entire set of independent variables (categorical and numerical). The table boxes highlighted with a shade of gray signify that the hybrid model reached a higher value of the quality measurement than the unmodified model. Table 3. Values of accuracy Hybrid Hybrid Hybrid Hybrid Unmodified decision Hybrid H CH KL DB Gap D1 0.602 0.602 0.791 0.704 D2 0.723 0.733 0.733-0.733 0.726 D3 0.796 0.698 0.796 - - 0.506 D4 0.395 0.395 0.395 0.395 0.395 0.395 0.820 0.807 0.851 D6 0.416 0.416 0.606 0.416 0.416 D7 0.881 0.897 0.907 0.881 0.897 0.905 D8 0.878 0.878 0.878-0.878 0.878 If the values of accuracy (Table 3) are to be taken into account, one can clearly see that the best hybrid models were created with the number of clusters indicated by the Davies-Bouldin index (DB). Only one solution (on the basis of dataset ) proved to be worse than the unmodified model. Table 4. Values of recall Hybrid Hybrid Hybrid Hybrid Unmodified decision Hybrid H CH KL DB Gap D1 0.512 0.512 0.437 0.726 D2 0.495 0.680 0.680-0.680 0.454 D3 0.364 0.523 0.364 - - 0.869 D4 0.725 0.725 0.725 0.725 0.725 0.725 0.472 0.561 0.374 D6 0.732 0.732 0.732 0.732-0.732 D7 0.907 0.903 0.901 0.907 0.903 0.885 D8 0.952 0.952 0.952-0.952 0.952 158

In the case of recall values (Table 4) the results for hybrid models were the same as the ones for the unmodified model (datasets: D4, D6, D8), or better (datasets: D2,, D7). The best solutions were provided by the Davies-Bouldin index (DB) and the gap statistic (Gap). Table 5. Values of precision Hybrid Hybrid Hybrid Hybrid Unmodified decision Hybrid H CH KL DB Gap D1 0.147 0.147 0.259 0.240 D2 0.571 0.564 0.564-0.564 0.587 D3 0.119 0.105 0.119 - - 0.098 D4 0.058 0.058 0.058 0.058 0.058 0.058 0.387 0.376 0.465 D6 0.085 0.085 0.0902 0.085-0.085 D7 0.699 0.740 0.764 0.699 0.740 0.766 D8 0.806 0.806 0.806-0.806 0.806 As far as precision is concerned (Table 5), the best results were achieved by using the Davies-Bouldin index (DB). However, more often solutions were identical with the ones in the unmodified model (datasets: D4, D8) or worse (D2,, D7). Table 6. Values of G-mean Hybrid Hybrid Hybrid Hybrid Unmodified decision Hybrid H CH KL DB Gap D1 0.561 0.561 0.605 0.714 D2 0.641 0.718 0.718-0.718 0.622 D3 0.548 0.609 0.548 - - 0.647 D4 0.522 0.522 0.522 0.522 0.522 0.522 0.644 0.689 0.589 D6 0.536 0.536 0.552 0.536-0.536 D7 0.890 0.899 0.904 0.890 0.899 0.898 D8 0.883 0.883 0.883-0.883 0.883 Considering the values of G-mean (Table 6) one may observe a relatively large effectiveness of hybrid models, in particular those based on the Davies-Bouldin index (DB) and the gap statistic (Gap). Only two solutions (datasets: D1, D3) proved to be better in the case of the unmodified decision. Table 7. Values of F-measure Hybrid Hybrid Hybrid Hybrid Unmodified decision Hybrid H CH KL DB Gap D1 0.228 0.228 0.325 0.361 D2 0.530 0.617 0.617-0.617 0.512 D3 0.179 0.175 0.179 - - 0.177 D4 0.108 0.108 0.108 0.108 0.108 0.108 159

0.425 0.450 0.415 D6 0.152 0.152 0.153 0.152-0.152 D7 0.790 0.813 0.827 0.790 0.813 0.821 D8 0.873 0.873 0.873-0.873 0.873 As far as the F-measure is concerned (Table 7) one can again see the advantage of the Davies-Bouldin index (DB). Hybrid models proved to be worse exclusively in the dataset D1, and the same in datasets D4 and D8. Table 8. Values of lift in 1st decile Hybrid Hybrid Hybrid Hybrid Hybrid Unmodified decision CH KL DB H Gap D1 1.24 1.24 2.25 3.59 D2 1.94 1.93 1.97-1.97 2.11 D3 1.95 2.28 1.95 - - 2.35 D4 1.55 1.55 1.55 1.55 1.55 1.55 2.75 2.67 3.30 D6 1.19 1.19 1.26 1.19-1.19 D7 2.83 2.99 3.09 2.83 2.99 3.10 D8 2.12 2.12 2.12-2.12 2.16 Table 9. Values of lift in 2nd decile Hybrid Hybrid Hybrid Hybrid Hybrid Unmodified decision CH KL DB H Gap D1 1.24 1.24 1.31 2.08 D2 1.81 1.93 1.97-1.97 1.86 D3 1.46 1.72 1.46 - - 2.35 D4 1.55 1.55 1.55 1.55 1.55 1.55 1.00 2.67 1.00 D6 1.19 1.19 1.26 1.19-1.19 D7 2.83 2.99 3.09 2.83 2.99 3.10 D8 2.12 2.12 2.12-2.12 2.16 Unfortunately, the authors anticipations regarding the values of the lift measure (Table 8 and Table 9) were not confirmed. Apart from the dataset D6 in the first decile and datasets D2,, D6 in the second decile, the results were the same or even more frequently considerably worse. Therefore, it may be noted that the unmodified decision outperformed hybrid models. Table 10. Presence of variable indicating class membership in Does the new independent variable (membership in clusters) participate in the partition of the tree? Unmodified decision tree Hybrid CH Hybrid KL Hybrid DB Hybrid H Hybrid Gap model (the number of numerical variables in the tree) 160

D1 yes yes yes 2 D2 yes yes no - no 2 D3 yes yes yes - - 4 D4 no no no no no 0 yes yes 4 D6 no no yes no - 0 D7 no yes yes no yes 1 D8 no no yes - no 2 It is worth conducting a verification whether the introduction of new independent variables indicating the class membership alters the interpretation of the model (Table 10). It turns out that if at least one primary split in the unmodified decision tree is numerical, then the class membership variable appears in the hybrid model. In the case of the dataset D6 there arose a situation where the hybrid model DB contained a new variable, whereas in the unmodified model there was no numerical variable. Therefore, it may be concluded that the structure of hybrid models based on the k- means algorithm and C&RT may sometimes enrich the content-related interpretation of the solution. 5 Conclusions The construction of hybrid models based on the k-means algorithm and C&RT decision trees may in some situations improve the performance of predictive models. It appears that cluster validity indices, which determine a different optimal number of clusters, play an important role here. It may be concluded from the conducted experiment that the Davies-Bouldin index and the gap statistic prove to perform the best. Hybrid models supply higher values of accuracy, G-mean, F-measure. In some cases they are better as far as recall and precision are concerned. However, it seems that they do not work when it comes to improving the lift measure, which plays an important role in marketing application. The best results are obtained in the case of hybrid models, in which the number of clusters is relatively high. This constitutes a certain inconvenience as an excessively high number of subgroups complicates their interpretation. No connection was noted between the performance measures of hybrid models and the percentage of class 1 of the dependent variable. Similarly, no dependence was observed between the performance and the ratio of numerical independent variables to categorical independent variables. The authors see the need for the extension of the experiment onto other datasets, the modification of parameters of the decision tree (e.g. a priori probabilities, misclassification costs and minimum number of instances in the terminal node), and experiments with fuzzy clustering methods. 161

References 1. Łapczy ski, M., Surma, J.: Hybrid Predictive Models for Optimizing Marketing Banner Ad Campaign in On-line Social Network. In: Stahlbock, R., Weiss, G.M. (eds.) Proceedings of the 2012 International Conference on Data Mining, CSREA Press, Las Vegas Nevada, USA, 2012, pp.140-146 (2012) 2. Bose, I., Chen, X.: Hybrid Models Using Unsupervised Clustering for Prediction of Customer Churn. Journal of Organizational Computing and Electronic Commerce. vol. 19, no. 2, April-June, pp. 133-151 (2009) 3. Chu, B-H., Tsai, M-S., Ho, Ch-S.: Toward a Hybrid Data Mining Model for Customer Retention. Knowledge-Based Systems. no. 20, pp. 703-718 (2007) 4. Li, Y., Deng, Z., Qian, Q., Xu, R.: Churn Forecast Based on Two-step Classification in Security Industry. Intelligent Information Management. no. 3, pp. 160-165 (2011) 5. Gaddam, S.R., Phoha, V.V., Balagani, K.S.: K-means + ID3: A Novel Method for Supervised Anomaly Detection by Cascading K-means Clustering and ID3 Decision Tree Learning Methods. In: IEEE Transactions on Knowledge and Data Engineering, vol. 19, no. 3, March pp. 345-354 (2007) 6. Shouman, M., Turner, T., Stocker, R.: Integrating Decision Tree and K-Means Clustering with Different Initial Centroid Selection Methods in the Diagnosis of Heart Disease Patients. In: Stahlbock, R., Weiss, G.M. (eds.) Proceedings of the 2012 International al Conference on Data Mining, CSREA Press, Las Vegas Nevada, USA, pp. 24-30 (2012) 7. Hartigan, J.A.: Wong M.A. A K-means Clustering Algorithm. Applied Statistics. vol. 28, no. 1, pp. 100-108 (1979) 8. Everit, B.S., Landau, S., Leese, M., Stahl, D.: Cluster Analysis. 5 th Edition. John Wiley & Sons, Chichester (2011) 9. Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and Regression Trees. Belmont, CA, Wadsworth International Group (1984) 10. Moro, S., Laureano, R., Cortez, P.: Using Data Mining for Bank Direct Marketing: An Application of the CRISP-DM Methodology. In: Novais, P. et al. (eds.), Proceedings of the European Simulation and Modelling Conference - ESM'2011, Guimarães, Portugal, October, pp. 117-121 (2011) 11. Frank, A., Asuncion, A.: UCI Machine Learning Repository [http://archive.ics.uci.edu/ml]. Irvine, CA: University of California, School of Information and Computer Science (2010) 12. van der Putten, P., van Someren, M. (eds): CoIL Challenge 2000: The Insurance Company Case. In: Also a Leiden Institute of Advanced Computer Science Technical Report 2000-09, Sentient Machine Research, Amsterdam, June 22, (2000) 13. Blake, C.L., Merz, C.J.: Churn Data Set, UCI Repository of Machine Learning Databases. http://www.sgi.com/tech/mlc/db, University of California, Department of Information and Computer Science, Irvine, CA (1998) 14. KDD Cup 2009, http://www.kddcup-orange.com 15. Causality Workbench. Challenges in Machine Learning, http://www.causality.inf.ethz.ch/data/cina.html 162