Pp. 176{182 in Proceedings of The Second International Conference on Knowledge Discovery and Data Mining. Predictive Data Mining with Finite Mixtures

Similar documents
Improving Simple Bayes. Abstract. The simple Bayesian classier (SBC), sometimes called

Lecture 1: Machine Learning Basics

Clouds = Heavy Sidewalk = Wet. davinci V2.1 alpha3

Probabilistic Latent Semantic Analysis

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Rule Learning With Negation: Issues Regarding Effectiveness

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Python Machine Learning

Learning Methods for Fuzzy Systems

Learning From the Past with Experiment Databases

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

A Comparison of Standard and Interval Association Rules

Henry Tirri* Petri Myllymgki

Introduction to Simulation

Rule Learning with Negation: Issues Regarding Effectiveness

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Exploration. CS : Deep Reinforcement Learning Sergey Levine

CS Machine Learning

Mining Student Evolution Using Associative Classification and Clustering

A Case Study: News Classification Based on Term Frequency

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Infrastructure Issues Related to Theory of Computing Research. Faith Fich, University of Toronto

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

Mining Association Rules in Student s Assessment Data

The Strong Minimalist Thesis and Bounded Optimality

Transductive Inference for Text Classication using Support Vector. Machines. Thorsten Joachims. Universitat Dortmund, LS VIII

PH.D. IN COMPUTER SCIENCE PROGRAM (POST M.S.)

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Australian Journal of Basic and Applied Sciences

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

A Model of Knower-Level Behavior in Number Concept Development

The Effects of Ability Tracking of Future Primary School Teachers on Student Performance

Accuracy (%) # features

The distribution of school funding and inputs in England:

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Computerized Adaptive Psychological Testing A Personalisation Perspective

Action Models and their Induction

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

Multi-Dimensional, Multi-Level, and Multi-Timepoint Item Response Modeling.

SARDNET: A Self-Organizing Feature Map for Sequences

stateorvalue to each variable in a given set. We use p(x = xjy = y) (or p(xjy) as a shorthand) to denote the probability that X = x given Y = y. We al

Applications of data mining algorithms to analysis of medical data

IT Students Workshop within Strategic Partnership of Leibniz University and Peter the Great St. Petersburg Polytechnic University

Software Maintenance

A Rough Set Approach to Data Mining: Extracting a Logic of Default Rules from Data. Torulf Mollestad

Generative models and adversarial training

Semi-Supervised Face Detection

Toward Probabilistic Natural Logic for Syllogistic Reasoning

Lecture 1: Basic Concepts of Machine Learning

Summarizing Text Documents: Carnegie Mellon University 4616 Henry Street

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

CONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS

Proceedings of the 19th COLING, , 2002.

phone hidden time phone

Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology

CSL465/603 - Machine Learning

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Predicting Students Performance with SimStudent: Learning Cognitive Skills from Observation

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations

STA 225: Introductory Statistics (CT)

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Lecture 10: Reinforcement Learning

10.2. Behavior models

Assignment 1: Predicting Amazon Review Ratings

DIDACTIC MODEL BRIDGING A CONCEPT WITH PHENOMENA

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Speech Recognition at ICSI: Broadcast News and beyond

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Generation of Attribute Value Taxonomies from Data for Data-Driven Construction of Accurate and Compact Classifiers

Graphical Data Displays and Database Queries: Helping Users Select the Right Display for the Task

On-the-Fly Customization of Automated Essay Scoring

Problems of the Arabic OCR: New Attitudes

Reducing Features to Improve Bug Prediction

Integrating E-learning Environments with Computational Intelligence Assessment Agents

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Hierarchical Linear Modeling with Maximum Likelihood, Restricted Maximum Likelihood, and Fully Bayesian Estimation

Rule-based Expert Systems

Why Did My Detector Do That?!

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

Cooperative evolutive concept learning: an empirical study

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Axiom 2013 Team Description Paper

Statewide Framework Document for:

INPE São José dos Campos

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

How do adults reason about their opponent? Typologies of players in a turn-taking game

An Online Handwriting Recognition System For Turkish

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Citrine Informatics. The Latest from Citrine. Citrine Informatics. The data analytics platform for the physical world

2 Mitsuru Ishizuka x1 Keywords Automatic Indexing, PAI, Asserted Keyword, Spreading Activation, Priming Eect Introduction With the increasing number o

Georgetown University at TREC 2017 Dynamic Domain Track

An Introduction to Simio for Beginners

A Note on Structuring Employability Skills for Accounting Students

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Transcription:

Pp. 176{182 in Proceedings of The Second International Conference on Knowledge Discovery and Data Mining (Portland, OR, August 1996). Predictive Data Mining with Finite Mixtures Petri Kontkanen Petri Myllymaki Henry Tirri Complex Systems Computation Group (CoSCo) P.O.Box 26, Department of Computer Science FIN-00014 University of Helsinki, Finland URL: http://www.cs.helsinki.fi/research/cosco/ Email: Firstname.Lastname@cs.Helsinki.FI Abstract In data mining the goal is to develop methods for discovering previously unknown regularities from databases. The resulting models are interpreted and evaluated by domain experts, but some model evaluation criterion is needed also for the model construction process. The optimal choice would be to use the same criterion as the human experts, but this is usually impossible as the experts are not capable of expressing their evaluation criteria formally. On the other hand, it seems reasonable to assume that any model possessing the capability of making good predictions also captures some structure of the reality. For this reason, in predictive data mining the search for good models is guided by the expected predictive error of the models. In this paper we describe the Bayesian approach to predictive data mining in the nite mixture modeling framework. The nite mixture model family is a natural choice for domains where the data exhibits a clustering structure. In many real world domains this seems to be the case, as is demonstrated by our experimental results on a set of public domain databases. Introduction Data mining aims at extracting useful information from databases by discovering previously unknown regularities from data (Fayyad et al. 1996). In the most general context, nding such interesting regularities is a process (often called knowledge discovery in databases) which includes the interpretation of the extracted patterns based on the domain knowledge available. Typically the pattern extraction phase is performed by a structure searching program, and the interpretation phase by a human expert. The various proposed approaches dier in the representation language for the structure to be discovered (association rules (Agrawal et al. 1996), Bayesian networks (Spirtes, Glymour, & Scheines 1993), functional dependencies (Mannila & Raiha 1991), prototypes (Hu & Cercone 1995) etc.), and in the search methodology used for discovering such structures. A large body of the data mining research is exploratory in nature, i.e., search for any kind of structure in the database in order to understand the domain better. Akin to the practice of multivariate exploratory analysis in social sciences (Basilevsky 1994), much of the work in the data mining area relies on a task-specic expert assessment of the model goodness. We depart from this tradition, and assume that the discovery process is performed with the expected prediction capability in mind. Consequently, we are trying to answer the question \Which of the models best explains a given database?" by addressing the (in many practical cases more pertinent) question \Which of the models yields the best predictions for future observations from the same process which generated the given database?" In our work the evaluation criteria in the model construction process is based directly on the expected predictive capability of the models, not on more implicit criteria embedded in the search algorithm. The use of predictiveness as a model selection criteria can be justied by the observation that a model with a good predictive capability must have captured some regularities that also reect properties of the data generating process. We call this approach predictive data mining. Predictive data mining is relevant in a wide variety of application areas from credit card fraud detection and sales support systems to industrial process control. Our current work is motivated by large scale conguration problems (e.g., building large generators) where properties of new congurations can be predicted using the regularities in the existing conguration For estimating the expected predictive performance, there exist theoretical measures (see e.g., (Wallace & Freeman 1987; Rissanen 1989; Raftery 1993)) which oer a solid evaluation criterion for the models, but such measures tend to be hard to compute for highdimensional spaces. In the case of large databases several approximations to these criteria could be used, but many of them are inaccurate with small databases as pointed out in (Kontkanen, Myllymaki, & Tirri 1996a). Alternatively we can choose some prediction problem, and evaluate prediction error empirically by using the available An example of such a prediction task would be to predict an unknown attribute value of a data item, given a set of some 176

other instantiated attributes. It should be observed that we do not assume that the set of predicted attributes are xed in advance during the discovery process prediction can be seen as a pattern completion task, where the errors in incomplete pattern completion can be used as a model measure for the goodness of the model. In this work we adopt the empirical approach and use the crossvalidation method (Stone 1974; Geisser 1975) for model selection on a set of public domain databases. In the work presented below we have adopted the basic concepts from the general framework of exploring computational models of scientic discovery (Shrager & Langley 1990). Given a database, we do not attempt to discover arbitrary structures, but restrict the possible patterns (models) to be members of a predened set, which we call the model space. Examples of such model spaces are the set of all possible association rules with a xed set of attributes, or a set of all nite mixture distributions (Everitt & Hand 1981; Titterington, Smith, & Makov 1985). A choice of a model space necessarily introduces prior knowledge to the search process. We would like the model space to be simple enough to allow tractable search, yet powerful enough to include models with good prediction capabilities. Therefore in the current work we have restricted ourselves to a simple, computationally ecient set of probabilistic models from the family of nite mixtures. Intuitively this choice reects our a priori assumption that the real life data is generated by several distinct processes, which is revealed as a cluster structure in the data. A nite mixture model for a set of random variables is a weighted sum of a relatively small number of independent mixing distributions. The main advantage of using nite mixture models lies in the fact that the computations for probabilistic reasoning can be implemented as a single pass computation (see the next section). Finite mixtures have also a natural means to model multimodal distributions and are universal in the sense that they can approximate any distribution arbitrarily close as long as a sucient number of component densities can be used. Finite mixture models can also be seen to oer a Bayesian solution to the case matching and case adaptation problems in instance-based reasoning (see the discussion in (Tirri, Kontkanen, & Myllymaki 1996)), i.e., they can also be viewed as a theoretically sound representation language for a \prototype" model space. This is interesting from the a priori knowledge acquisition point of view, since in many cases the domain experts seem to be able to express their expert knowledge very easily by using prototypical examples or distributions, which can then be coded as mixing distributions in our nite mixture framework. In order to nd probabilistic models for making good predictions, we follow the Bayesian approach (Gelman et al. 1995; Cheeseman 1995), as it oers a solid theoretical framework for combining both (suitably coded) a priori domain information and information from the sample database in the model construction process. Bayesian approach also makes a clear separation between the search component and the model measure, and allows therefore modular combinations of dierent search algorithms and model evaluation criteria. Our approach is akin to the AutoClass system (Cheeseman et al. 1988), which has been successfully used for data mining problems, such as Land- Sat data clustering (Cheeseman & Stutz 1996). In the case of nite mixtures, the model search problem can be seen as searching for the missing values of the unobserved latent clustering variable in the dataset. The model construction process consists of two phases: model class selection and model class parameter selection. The model class selection can be understood as nding the proper number of mixing distributions, i.e., the number of clusters in the data space, and the model class parameter selection as nding the attribute value probabilities for each mixture component. The model search problem in this framework is only briey outlined in this paper a more detailed exposition can be found in (Kontkanen, Myllymaki, & Tirri 1996b; 1996a). One should observe that theoretically the correct Bayesian approach for obtaining maximal predictive accuracy would be to use the sum of outcomes of all the possible dierent models, weighted by their posterior probability, i.e., in our case a \mixture of all the mixtures". This is clearly not feasible for data mining considerations, since such a model can hardly be given any useful semantic interpretation. We therefore use only a single, maximum a posteriori probability (MAP) model for making predictions. The feasibility of this approach is discussed in (Cheeseman 1995). Bayesian inference by nite mixture models In our predictive data mining framework the problem domain is modeled by m discrete random variables X1; : : : ; X m. A data instantiation ~ d is a vector in which all the variables X i have been assigned a value, ~d = (X1 = x1; : : : ; X m = x m ); where x i 2 fx i1; : : : ; x ini g. Correspondingly we can view the database D as a random sample ( d1; ~ : : : ; d ~ N ), i.e., a set of N i.i.d. (independent and identically distributed) data instantiations, where each d ~ j is sampled from P, the joint distribution of the variables (X1; : : : ; X m ). In our work we assume that the database D is generated by K dierent mechanisms, which all can have their own distributions, and that each data vector originates from exactly one of these mechanisms. Thus the instantiation space is divided into K clusters, each of which consists of the data vectors generated by the corresponding mechanism. From the assumptions above it follows that a natural candidate for a probabilistic 177

model family is the family of nite mixtures (Everitt & Hand 1981; Titterington, Smith, & Makov 1985), where the problem domain probability distribution is approximated as a weighted sum of mixture distributions: (1) KX P ( d) ~ = P (Y = y k )P ( d ~ j Y = y k ) : Here the values of the discrete clustering random variable Y correspond to the separate clusters of the instantiation space, and each mixture distribution P ( ~ d j Y = y k ) models one data generating mechanism. Moreover, we assume that the problem domain data is tightly clustered so that the clusters can actually be regarded as points in the instantiation space, and data vectors belonging to the same cluster represent noisy versions of that (unknown) point. Therefore we can assume that the variables X i inside each cluster are independent by which (1) becomes P ( ~ d) = P (X1 = x1; : : : ; X m = x m ) KX = my P (Y = y k ) P (X i = x i jy = y k ) i=1 In our model both the cluster distribution P (Y ) and the intra-class conditional distributions P (X i jy = y k ) are multinomial. Thus a nite mixture model can be dened by rst xing K, the model class (the number of the mixing distributions), and then by determining the values of the model parameters = (; ); 2, where = (1; : : : ; K ); k = P (Y = y k ), and = (11; : : : ; 1m; : : :; K1; : : : ; Km ); ki = ( ki1; : : :; kini ); where kil = P (X i = x il jy = y k ). Given a nite mixture model that models the cluster structure of the database, predictive inference can be performed in a computationally ecient manner. The Bayesian approach to predictive inference (see e.g., (Bernardo & Smith 1994)) aims at predicting unobserved future quantities by means of already observed quantities. More precisely, let I = fi1; : : : ; i t g be the indices of the instantiated variables, and let X = fx is = x isls ; s = 1; : : : ; tg denote the corresponding assignments. Now we want to determine the distribution P (X i = x il j; X ) =! P K k kil Q t s=1 ki sl s P K k Q t s=1 ki sl s : The conditional predictive distribution of X i can clearly be calculated in time O(Ktn i ), where K is the number of clusters, t the number of instantiated variables and n i the number of values of X i. Observe that K is usually small compared to the sample size N, and : thus the prediction computation can be performed very eciently (Myllymaki & Tirri 1994). The predictive distributions can be used for classication and regression tasks. In classication problems, we have a special class variable X c which is used for classifying data. In more general regression tasks, we have more than one variable for which we want to compute the predictive distribution, given that the values of the other variables are instantiated in advance. As in the conguration problems mentioned earlier, nite mixture models can also be used for nding the most probable value assignment combination for all the uninstantiated variables, given the values of the instantiated variables. These assignment combinations are useful when modeling actual objects such as machines, where probability information is in any case used to select a proper conguration with instantiated values for all the attributes. Learning nite mixture models from data In the previous section we described how the prediction of any variable could be made given a nite mixture model. Here we will briey outline how to learn such models from a given database D. Let D = ( ~ d1; : : : ; ~ d N ) be a database of size N. By learning we mean here the problem of constructing a single nite mixture model M K () which represents the problem domain distribution P as accurately as possible in terms of the prediction capability. This learning process can be divided into two separate phases: in the rst phase we wish to determine the optimal value for K, the number of mixing distributions (the model class), and in the second phase we wish to nd MAP parameter values ^ for the chosen model class. In the Bayesian framework, the optimal number of mixing distributions (clusters) can be determined by evaluating the posterior probability for each model class M K given the data: P (M K jd) / P (DjM K )P (M K ); K = 1; : : :; N; where the normalizing constant P (D) can be omitted since we only need to compare dierent model classes. The number of clusters can safely be assumed to be bounded by N, since otherwise the sample size is clearly too small for the learning problem in question. Assuming equal priors for the model classes, they can be ranked by evaluating the evidence P (DjM K ) (or equivalently the stochastic complexity (Rissanen 1989)) for each model class. This term is dened as a multidimensional integral and it is usually very hard to evaluate, although with certain assumptions, the evidence can in some cases be determined analytically (Heckerman, Geiger, & Chickering 1995; Kontkanen, Myllymaki, & Tirri 1996a). In the experimental results presented in the next section we chose another approach and estimated the prediction error empirically 178

by using the crossvalidation algorithm (Stone 1974; Geisser 1975). After choosing the appropriate model class, there remains the task of nding the actual model (i.e., the model class parameters). In the Bayesian approach, this is usually done by nding the MAP (maximum a posteriori) estimate of parameters by maximizing the posterior density P (jd). We assume that the prior distributions of the parameters are from the family of Dirichlet densities, since it is conjugate (see e.g., (DeGroot 1970)) to the family of multinomials, i.e., the functional form of parameter distribution remains invariant in the prior-to-posterior transformation. Finding the exact MAP estimate of is, however, computationally infeasible task, thus we are forced to use numerical approximation methods. We used here a variant of the Expectation-Maximization (EM) algorithm (Dempster, Laird, & Rubin 1977) for this purpose, since the method is easily applicable in this domain and produces good solutions quite rapidly, as can be seen in (Kontkanen, Myllymaki, & Tirri 1996b). Empirical results The nite mixture based approach for predictive data mining described above has been implemented as part of a more general software environment for probabilistic modeling. To validate the advocated approach we wanted to test the approach with real data sets, preferably with ones that have been used for other experiments also. The main advantage of using natural databases instead of generated ones is that they are produced without any knowledge of the particular procedures that they are tested on. In addition, although there is no way of telling how the results for a real database generalizes to other problems, we at least know that there are some domains where the results have practical relevance. Below we report results from an ongoing extensive experimentation with the Bayesian nite mixture modeling method using publicly available datasets for classi- cation problems. The selection of databases was done on the basis of their reported use, i.e., we have preferred databases that have been used for testing many dierent methods over databases with only isolated results. Many of the databases used are from the StatLog project (Michie, Spiegelhalter, & Taylor 1994). The experimental setups and the best success rates obtained are shown in Table 1. All our results are crossvalidated, and when possible (for the StatLog datasets) we have used the same crossvalidation schemes (the same number of folds) as in (Michie, Spiegelhalter, & Taylor 1994). The results on each of these datasets are shown in Figures 1{3. In each case, the maximum, minimum and the average success rate on 30 independent crossvalidation runs are given. Even from these preliminary empirical results it can clearly be seen that the Bayesian nite mixture ap- 5 5 5 Figure 1: Crossvalidation results with the Australian 8 6 4 2 8 6 4 Figure 2: 5 4 3 2 1 9 8 Crossvalidation results with the Diabetes 7 Figure 3: Crossvalidation results with the German Credit 179

dataset size #attrs #CVfolds #clusters success rate (%) Australian 690 15 10 17 87.2 Diabetes 768 9 12 20 76.8 German credit 1000 21 10 23 74.1 Glass 214 10 7 30 87.4 Heart disease 270 14 9 8 84.8 Hepatitis 150 20 5 9 88.0 Iris 150 5 5 4 97.3 Lymphography 148 19 5 19 86.6 Primary tumor 339 18 10 21 50.4 Table 1: Description of the experiments. 8 6 4 2 8 0.4 6 0.3 4 0.2 Figure 4: Crossvalidation results with the Glass 2 Figure 6: Crossvalidation results with the Hepatitis 5 1 5 5 0.4 0.3 Figure 5: Crossvalidation results with the Heart Disease 0.2 Figure 7: Crossvalidation results with the Iris 180

5 5 5 Figure 8: Crossvalidation results with the Lymphography 0.45 0.4 0.35 0.3 0.25 0.2 Figure 9: Crossvalidation results with the Primary Tumor proach performs well. The general tendency in our experiments is that the success rate rst increases with increasing number of clusters up to a certain knee point, after which the performance stays relatively constant. Unlike in many other similar studies, the overtting phenomenon does not seem to have a signicant impact here. We believe that this can be explained by the special competitive nature of the EM algorithm in our clustering framework: after the knee point, unnecessary clusters tend to gradually disappear during the EM process, so that the method seems to be able to automatically regulate (to some extent) the number of parameters. This interesting empirical observation deserves further investigation in the future. In a elded data mining application, the resulting mixture structure would be given to the expert for evaluation. In our case with the public domain data sets this was not possible (as we did not have the experts available). Instead we investigated the clusters found with the D-SIDE environment, which provides a graphical user interface for displaying predictive distributions and the mixture structure 1. For the German credit data, a credit assessment problem, where the attributes were understandable also to a layman, clear interpretations could be given for 6 of the 9 clusters. For example the most inuential cluster could be described consisting of working single women (age < 30, 4? 7 years in the current job), living in an apartment they owned, and requesting a small loan (less than 1800 DM) for furniture or home appliances with very little savings. Needless to say classication suggestion for applicants matching such a description was positive. It should be emphasized that since the mixture model built allows calculation of any predictive distribution, it can also be used explore dependencies between attributes. For example one can investigate what is the eect of being just recently employed to the purpose of the loan. Although we have reported in Table 1 the model classes (number of clusters) with the best performing single model, sometimes the model class with the highest average success rate could be more useful for data mining purposes. This is due to the fact that because of the small number of repetitions, the variances of the results are quite high, as depicted in the gures. Unfortunately, as these are public domain data, and not industrial applications, we do not have a domain expert available to evaluate our results. However, we are currently also working with several real-world industrial pilot applications. Conclusion We discussed the use of nite mixture models for predictive data mining, where the search for unknown structures is guided by expected predictive performance. For the model selection problem, we adopted the Bayesian approach. In this work the predictive performance was measured empirically by using the crossvalidation method. The applicability of the approach was demonstrated empirically by presenting extensive experimental results on a set of public domain databases. In addition to revealing a possible cluster structure in the data, the resulting mixture structure allows also more general inspection of the attribute dependencies by the computationally very ecient predictive Bayesian inference. Acknowledgments This research has been supported by the Technology Development Center (TEKES). The Primary Tumor, the Breast Cancer and the Lymphography data were obtained from the University Medical Centre, Institute of Oncology, Ljubljana, Yugoslavia. Thanks go to M. Zwitter and M. Soklic for providing the data. 1 A running Java TM prototype of the D-SIDE software is available for experimentation through our WWW homepage at URL \http://www.cs.helsinki.fi/research/cosco/". 181

References Agrawal, R.; Mannila, H.; Srikant, R.; Toivonen, H.; and Verkamo, A. 1996. Fast discovery of association rules. In Fayyad, U.; Piatetsky-Shapiro, G.; Smyth, P.; and Uthurusamy, R., eds., Advances in Knowledge Discovery and Data Mining. MIT Press. Basilevsky, A. 1994. Statistical Factor Analysis and Related Methods. Theory and Applicatioms. New York: John Wiley & Sons. Bernardo, J., and Smith, A. 1994. Bayesian theory. John Wiley. Cheeseman, P., and Stutz, J. 1996. Bayesian classication (AutoClass): Theory and results. In Fayyad, U.; Piatetsky-Shapiro, G.; Smyth, P.; and Uthurusamy, R., eds., Advances in Knowledge Discovery and Data Mining. Menlo Park: AAAI Press. chapter 6. Cheeseman, P.; Kelly, J.; Self, M.; Stutz, J.; Taylor, W.; and Freeman, D. 1988. Autoclass: A Bayesian classication system. In Proceedings of the Fifth International Conference on Machine Learning, 54{64. Cheeseman, P. 1995. On Bayesian model selection. In Wolpert, D., ed., The Mathematics of Generalization, volume XX of SFI Studies in the Sciences of Complexity. Addison-Wesley. 315{330. DeGroot, M. 1970. Optimal statistical decisions. McGraw-Hill. Dempster, A.; Laird, N.; and Rubin, D. 1977. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series B 39(1):1{38. Everitt, B., and Hand, D. 1981. Finite Mixture Distributions. London: Chapman and Hall. Fayyad, U.; Piatetsky-Shapiro, G.; Smyth, P.; and Uthurusamy, R., eds. 1996. Advances in Knowledge Discovery and Data Mining. Cambridge, MA: MIT Press. Geisser, S. 1975. The predictive sample reuse method with applications. Journal of the American Statistical Association 70(350):320{328. Gelman, A.; Carlin, J.; Stern, H.; and Rubin, D. 1995. Bayesian Data Analysis. Chapman & Hall. Heckerman, D.; Geiger, D.; and Chickering, D. 1995. Learning Bayesian networks: The combination of knowledge and statistical data. Machine Learning 20(3):197{243. Hu, X., and Cercone, N. 1995. Rough sets similaritybased learning from databases. In Fayyad, U., and Uthurusamy, R., eds., Proceedings of the First International Conference on Knowledge Discovery & Data Mining, 162{167. Kontkanen, P.; Myllymaki, P.; and Tirri, H. 1996a. Comparing Bayesian model class selection criteria by discrete nite mixtures. In Proceedings of the ISIS (Information, Statistics and Induction in Science) Conference. (To appear.). Kontkanen, P.; Myllymaki, P.; and Tirri, H. 1996b. Constructing Bayesian nite mixture models by the EM algorithm. Technical Report C-1996-9, University of Helsinki, Department of Computer Science. Mannila, H., and Raiha, K.-J. 1991. The design of relational databases. Addison-Wesley. Michie, D.; Spiegelhalter, D.; and Taylor, C., eds. 1994. Machine Learning, Neural and Statistical Classication. London: Ellis Horwood. Myllymaki, P., and Tirri, H. 1994. Massively parallel case-based reasoning with probabilistic similarity metrics. In Wess, S.; Altho, K.-D.; and Richter, M., eds., Topics in Case-Based Reasoning, volume 837 of Lecture Notes in Articial Intelligence. Springer- Verlag. 144{154. Raftery, A. 1993. Approximate Bayes factors and accounting for model uncertainty in generalized linear models. Technical Report 255, Department of Statistics, University of Washington. Rissanen, J. 1989. Stochastic Complexity in Statistical Inquiry. World Scientic Publishing Company. Shrager, J., and Langley, P., eds. 1990. Computational Models of Scientic Discovery and Theory Formation. San Mateo, CA: Morgan Kaufmann Publishers. Spirtes, P.; Glymour, C.; and Scheines, R., eds. 1993. Causation, Prediction and Search. Springer-Verlag. Stone, M. 1974. Cross-validatory choice and assessment of statistical predictions. Journal of the Royal Statistical Society (Series B) 36:111{147. Tirri, H.; Kontkanen, P.; and Myllymaki, P. 1996. Probabilistic instance-based learning. In Saitta, L., ed., Machine Learning: Proceedings of the Thirteenth International Conference (to appear). Morgan Kaufmann Publishers. Titterington, D.; Smith, A.; and Makov, U. 1985. Statistical Analysis of Finite Mixture Distributions. New York: John Wiley & Sons. Wallace, C., and Freeman, P. 1987. Estimation and inference by compact coding. Journal of the Royal Statistical Society 49(3):240{265. 182