Tilburg University. Intelligible neural networks with BP-SOM Weijters, T.; van den Bosch, Antal; van den Herik, Jaap. Published in: NAIC-97

Similar documents
Evolution of Symbolisation in Chimpanzees and Neural Nets

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

SARDNET: A Self-Organizing Feature Map for Sequences

Lecture 1: Machine Learning Basics

Knowledge Transfer in Deep Convolutional Neural Nets

Constructive Induction-based Learning Agents: An Architecture and Preliminary Experiments

Rule Learning With Negation: Issues Regarding Effectiveness

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Artificial Neural Networks written examination

An Empirical and Computational Test of Linguistic Relativity

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Python Machine Learning

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Softprop: Softmax Neural Network Backpropagation Learning

Learning From the Past with Experiment Databases

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

CS Machine Learning

An empirical study of learning speed in backpropagation

Rule Learning with Negation: Issues Regarding Effectiveness

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

A Case Study: News Classification Based on Term Frequency

Reinforcement Learning by Comparing Immediate Reward

Evolutive Neural Net Fuzzy Filtering: Basic Description

Predicting Students Performance with SimStudent: Learning Cognitive Skills from Observation

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Chapter 2 Rule Learning in a Nutshell

Word Segmentation of Off-line Handwritten Documents

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

COMPUTER-AIDED DESIGN TOOLS THAT ADAPT

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

(Sub)Gradient Descent

BUILD-IT: Intuitive plant layout mediated by natural interaction

Speech Recognition at ICSI: Broadcast News and beyond

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

Abstractions and the Brain

INPE São José dos Campos

Lecture 1: Basic Concepts of Machine Learning

Axiom 2013 Team Description Paper

Modeling function word errors in DNN-HMM based LVCSR systems

Generative models and adversarial training

Learning Methods for Fuzzy Systems

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

Seminar - Organic Computing

Beyond the Pipeline: Discrete Optimization in NLP

On the Combined Behavior of Autonomous Resource Management Agents

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

CSL465/603 - Machine Learning

Computerized Adaptive Psychological Testing A Personalisation Perspective

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

The Good Judgment Project: A large scale test of different methods of combining expert predictions

COMPETENCY-BASED STATISTICS COURSES WITH FLEXIBLE LEARNING MATERIALS

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Learning to Schedule Straight-Line Code

How do adults reason about their opponent? Typologies of players in a turn-taking game

Modeling function word errors in DNN-HMM based LVCSR systems

Using focal point learning to improve human machine tacit coordination

Learning Distributed Linguistic Classes

Australian Journal of Basic and Applied Sciences

A Version Space Approach to Learning Context-free Grammars

Applications of memory-based natural language processing

On-Line Data Analytics

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

Measurement & Analysis in the Real World

Henry Tirri* Petri Myllymgki

Running Head: STUDENT CENTRIC INTEGRATED TECHNOLOGY

A Note on Structuring Employability Skills for Accounting Students

Discriminative Learning of Beam-Search Heuristics for Planning

Human Emotion Recognition From Speech

Cooperative evolutive concept learning: an empirical study

Improvements to the Pruning Behavior of DNN Acoustic Models

BENCHMARK TREND COMPARISON REPORT:

Automatic Phonetic Transcription of Words. Based On Sparse Data. Maria Wolters (i) and Antal van den Bosch (ii)

An extended dual search space model of scientific discovery learning

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

TD(λ) and Q-Learning Based Ludo Players

arxiv: v1 [cs.lg] 15 Jun 2015

Using Web Searches on Important Words to Create Background Sets for LSI Classification

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Parallel Evaluation in Stratal OT * Adam Baker University of Arizona

OCR LEVEL 3 CAMBRIDGE TECHNICAL

The recognition, evaluation and accreditation of European Postgraduate Programmes.

An ICT environment to assess and support students mathematical problem-solving performance in non-routine puzzle-like word problems

Mining Student Evolution Using Associative Classification and Clustering

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

WHEN THERE IS A mismatch between the acoustic

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

NCEO Technical Report 27

Evidence for Reliability, Validity and Learning Effectiveness

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Transcription:

Tilburg University Intelligible neural networks with BP-SOM Weijters, T.; van den Bosch, Antal; van den Herik, Jaap Published in: NAIC-97 Publication date: 1997 Link to publication Citation for published version (APA): Weijters, T., van den Bosch, A., & van den Herik, H. J. (1997). Intelligible neural networks with BP-SOM. In K. Van Marcke, & W. Daelemans (Eds.), NAIC-97: Proceedings of the Ninth Dutch Conference on Artificial Intelligence, University of Antwerp, 12-13 november, 1997 (pp. 27-36). Antwerpen: CLIF/Bolesian. General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. - Users may download and print one copy of any publication from the public portal for the purpose of private study or research - You may not further distribute the material or use it for any profit-making activity or commercial gain - You may freely distribute the URL identifying the publication in the public portal Take down policy If you believe that this document breaches copyright, please contact us providing details, and we will remove access to the work immediately and investigate your claim. Download date: 25. feb. 218

Intelligible Neural Networks with BP-SOM Ton Weijters, Antal van den Bosch, H. Jaap van den Herik Department of Computer Science Universiteit Maastricht P.O. Box 616, NL-62 MD Maastricht The Netherlands fweijters,antal,herikg@cs.unimaas.nl Abstract Back-propagation learning (BP) is known for its serious limitations in generalising knowledge from certain types of learning material. BP-SOM is an extension of BP which overcomes some of these limitations. BP-SOM is a combination of a multi-layered feed-forward network (MFN) trained with BP, and Kohonen s self-organising maps (SOMs). In earlier reports, it has been shown that BP-SOM improved the generalisation performance whereas it decreased simultaneously the number of necessary hidden units without loss of generalisation performance. These are only two effects of the use of SOM learning during training of MFNs. In this paper we focus on two additional effects. First, we show that after BP-SOM training, activations of hidden units of MFNs tend to oscillate among a limited number of discrete values. Second, we identify SOM elements as adequate organisers of instances of the task at hand. We visualise both effects, and argue that they lead to intelligible neural networks and can be employed as a basis for automatic rule extraction. 1 Background Back-propagation learning [RHW86] (henceforth referred to as BP) is known for its serious limitations in (i) generalising knowledge from certain types of training material [Nor89,WK91] and (ii) inducing intelligible models [Nor89,SL97], In earlier publications [Wei95,Wei96,WVVP97] experimental results are reported in which the generalisation performances of BP-SOM are compared to two other learning algorithms for MFNs, viz. BP, and BP augmented with weight decay [Hin86]. Henceforth, the latter is referred to as BPWD. The reported experiments show that (i) BP-SOM learning results in MFNs with a better generalisation performance, as compared to MFNs trained with BP and BPWD, and (ii) an increased amount of hidden units can be pruned without loss of generalisation performance. These are only two effects of the use of SOM learning when training BP-SOM networks. In this paper, we investigate two additional effects: (i) hidden-unit activations

MFN SOM class A elements class B elements unlabelled element Figure 1: An example BP-SOM network. tend to end up oscillating between a limited number of discrete values, and (ii) the SOM can be seen as an organiser of the instances of the task at hand, dividing them into a limited number of subsets. Furthermore, we visualise these effects, and show that they lead to better intelligibility of trained MFNs. Moreover, the division of the training and testing material into homogeneous subsets by SOM elements is shown to be a useful step in automatic rule extraction. Section 2 summarises the BP-SOM architecture and learning algorithm. In Section 3, we present and visualise experimental observations of the BP-SOM architecture and learning algorithm. In section 4, we provide our conclusions. 2 BP-SOM Below we give a brief characterisation of the functioning of BP-SOM. For details we refer to [Wei95,Wei96,WVVP97]. The aim of the BP-SOM learning algorithm is to establish a cooperation between BP-learning and SOM-learning in order to find adequate hidden-layer representations for learning classification tasks. To achieve this aim, the traditional MFN architecture [RHW86] is combined with self-organising maps (SOMs) [Koh89]: each hidden layer of the MFN is associated with one SOM (see Figure 1). During training of the weights in the MFN, the corresponding SOM is trained on the hidden-unit activation patterns. After a number of training cycles of BP-SOM learning, each SOM, to a certain extent, develops self-organisation, and translates this self-organisation into classification information; i.e., each SOM element is provided with a class label (one of the output classes of the task). For example, let the BP-SOM network displayed in Figure 1 be trained on a classification task which maps instances to either output class A or B. If a SOM element is the best matching element to 4 hidden-unit activation patterns that map to class A, and to 2 hidden-unit activation patterns that map to class B, the class label of that SOM becomes A, with a reliability of 4=6 = :67. As a result, we can visually distinguish areas in the SOM: areas containing elements labelled with class A and class B, and areas containing unlabelled elements (no winning class could be found). In Figure 1, we see four class labels A, four

class labels B, and one element unlabelled. The self-organisation of the SOM is used as an addition to the standard BP learning rule [RHW86]. Classification and reliability information from the SOMs is included when updating the connection weights of the MFN. The error of a hidden-layer vector is a accumulation of the error computed by the BP learning rule, and a SOM error. The SOM error is the difference between the hidden-unit activation vector and the vector of its best-matching element associated with the same class on the SOM (for more details, cf. [Wei95,Wei96]). The effect of including SOM information in the error signals is that clusters of hidden-unit activation patterns of instances associated with the same class tend to become increasingly similar to each other. As soon as BP-SOM learning has finished, the SOMs are redundant, since they do not contribute to the activation feed-forward within the BP-SOM network. Therefore all SOMs are removed from a trained BP-SOM network, resulting in a stand-alone MFN. 3 Visualising the effects: experiments In this section, we present two experimental observations with the BP-SOM architecture and the BP-SOM learning algorithm. The observations are made by training BP-SOM on two benchmark classification tasks, viz. the parity-12 task, and the monks-1 task. The parity- 12 task is hard to learn for many learning algorithms, since it often leads to overfitting [Sch93,Tho95]. The reason is that the output classification depends on the values of all input features. The monks-1 is a well-known benchmark task for automatic rule extraction [Thr91]. For all experiments reported in this paper we have used a fixed set of parameters for the learning algorithms. The BP learning rate was set to.15 and the momentum to. In all SOMs a decreasing interaction strength from.15 to.5, and a decreasing neighbourhoodupdating context from a square with maximally 25 units to only 1 unit (the winner) was used [Koh89]. BP-SOM was trained for a fixed number of cycles m = 2; class labelling was performed at each 5th cycle (n = 5). Early stopping, a common method to prevent overfitting, was used in all experiments: the performance of a trained network was calculated in percentages of incorrectly-processed test instances at the cycle where the classification error on validation material was minimal [Pre94]. For further details on the experiments, we refer to [Wei96]. 3.1 Parity-12 BP, BPWD, and BP-SOM have been applied to the parity-12 task, i.e., the task to determine whether a bit string of s and 1 s of length 12 contains an even number of 1 s. The training set contained 1, different instances selected at random out of the set of 4,96 possible bit strings. The test set and the validation set contained 1 new instances each. The hidden layer of the MFN in all three algorithms contained 2 hidden units, and the SOM in BP-SOM contained 7 7 elements. The algorithms were run with 1 different random weight

BP: % incorrect BPWD: % incorrect BP-SOM: % incorrect Task Train Test Train Test Train Test parity-12 14.1 18.8 27.4 16.4 21.6 24.2 22.4 18.3 5.9 1.2 6.2 1.7 Table 1: Average generalisation performances of BP, BPWD, and BP-SOM, on the partity-12 task (plus standard deviation, after ; averaged over ten experiments) in terms of incorrectly-processed training and test instances. BP BPWD BP-SOM Figure 2: Graphic representation of a 7 7 SOM: associated with a BPtrained MFN (left). with a BPWD-trained MFN (middle), and with a BP-SOM network (right); all are trained on the parity-12 task. White squares represent class even ; black squares represent class odd. The width of a square represents the reliability of the element; a square of maximal size represents a reliability of 1%. initialisations. Table 1 displays the classification errors on training instances and test instances. The results indicate that BP-SOM performed significantly better than BP and BPWD on test material (with p-values p < :1 and p < :5, respectively), and that BP-SOM was able to avoid overfitting better than BP. Although BPWD was also able to avoid overfitting, it was at the cost of a low performance on both training and test material. To visualise the differences among the representations developed at the hidden layers of the MFNs trained with BP, BPWD, and BP-SOM, respectively, we also trained SOMs with the hidden-layer activities of the trained BP and BPWD networks. Figure 2 visualises the class labelling of the SOMs: the left part is an MFN after BP-training; the middle part visualises the SOM after a BPWD-training of an MFN, and the right part displays the SOM of the BP-SOM network after training on the same material. The SOM of the BP-SOM network is much more organised and clustered than that of the SOMs corresponding with the BP-trained and BPWD-trained MFNs. The reliability values of the elements of all three SOMs are represented by the width of the black and white squares. It can be seen that the overall reliability of the

BP BPWD BP-SOM standard deviation.2 standard deviation.2 standard deviation.2 1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 hidden unit 1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 hidden unit 1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 hidden unit Figure 3: Standard deviations of the activations of the 2 hidden units of a BP-trained MFN (left), a BPWD-trained MFN (middle), and a BP-SOM network (right), trained on the parity-12 task (1, instances). SOM of the BP-SOM network is considerably higher than that of the SOM of the BP-trained and BPWD-trained MFNs. Simplified hidden-unit activations By including the SOM in BP-SOM learning, clusters of hidden-unit activation patterns associated with the same class tend to become more similar [Wei96]. When analysing the hidden-unit activations in BP-SOM networks, we observed a valuable additional effect, viz. that hidden-unit activations culminated in a stable activity with a very low variance or that they resulted in oscillating between a limited number of discrete values. This clearly contrasts with hidden units in MFNs trained with BP, of which all activations usually display high variance. Figure 3 displays the standard deviation of the 2 hidden-unit activations of an MFN trained with BP (left), and MFN trained with BPWD (middle) and a BP-SOM network (right), each of them trained on the parity-12 task (1, instances). The standard deviations are computed after presentation of all training instances without further training. As can be seen from Figure 3, the standard deviations of ten out of twenty units in the BP-SOM network are equal to.1 or lower. Whenever a unit has a stable activation with a low standard deviation for all training instances, it is redundant in the input-output mapping. In that case, the unit can be pruned from the network, and its effect on the following layer (i.e., its mean activation times its weights to units in the following layer) can be included into the weights from the bias unit associated with the same layer, to units of the following layer. This pruning can be performed both during and after training. We trained BP, BPWD, and BP-SOM on the parity-12 task, in an attempt to prune hidden units according to a threshold criterion during training. We introduced a stability threshold parameter s, denoting the standard deviation of the unit s activation below which it is pruned. After a number of pilot experiments with different values for s, we performed experiments with s = :1. All three algorithms were trained on the parity-12 task with ten different random initialisations. We found that BP-SOM was able to prune 12 out of 2 hidden units

hidden unit 1 hidden unit 2 1 1.8.8 activations activations.2.2 1_1 1_3 1_5 1_72_7 5_6 4_7 5_3 7_1 test instances sorted by SOM-labeling 7_2 7_4 1_1 1_3 1_5 1_72_7 5_6 4_7 5_3 7_1 test instances sorted by SOM-labeling 7_2 7_4 hidden unit 3 hidden unit 4 1 1.8.8 activations.2 Figure 4: Activations of the four hidden units of a BP-SOM network 1_1 1_3 1_5 1_72_7 5_6 4_7 5_3 7_1 test instances sorted by SOM-labeling SOM elements, indicated by.2 the coordinates of the elements 7_2 7_4 activations trained on the parity-12 task on 1 test instances. The x- axis orders the test instances according to their clustering on (e.g., 1 1 indicates SOM element (1,1)). 1_1 1_3 1_5 1_72_7 5_6 4_7 5_3 7_1 test instances sorted by SOM-labeling (averaged over 1 experiments), without loss of generalisation performance. With the same setting of s, trained on the same tasks, no hidden units could be pruned with BP, nor with BPWD. Only with s = :1 hidden units could be pruned during BP and BPWD learning; however, this led to seriously worse generalisation performance of these networks. Since BP-SOM learning results in a trained MFN, the increased ability to prune hidden units, as illustrated with the parity-12 task, is an advantage over fixed-size networks. The latter may be chosen as small as the MFN end product, but the risk exists that a learning algorithm, such as BP, is faced with too few degrees of freedom to learn the task. We performed a series of ten experiments with BP, using an MFN with one hidden layer of eight units, and obtained an average classification error on test material of 36.5% ( 12.6): setting the number of hidden units explicitly to eight did not help BP in learning the task. To illustrate the second effect, viz. the oscillating of hidden-unit activations between a limited number of discrete values, one experiment with an MFN trained with BP-SOM on the parity-12 task was chosen. In this experiment, an extremely large number of hidden units was pruned (16 out of 2 hidden units), while the performance of the trained MFN on test material was still acceptable (classification error 4%). Figure 4 displays the activations of the four hidden units of the BP-SOM-trained MFN (displayed on the y-axis), measured for each of the 1 test instances (displayed on the x-axis). The test instances are grouped on the basis of their respective SOM classification: we collected for each labelled SOM element all associated test instances. It can be seen from Figure 4 that the four hidden units end up oscillating between a discrete 7_2 7_4

number of values, depending on the SOM element to which a test instance belongs. This effect is witnessed in all BP-SOM-trained MFNs. The less hidden units are pruned, the less discrete activations are used by the remaining hidden units. The apparent discretisation of activations offers a potentially interesting basis for rule extraction from MFNs: In order to extract rules from (MFNs), it is better that (hidden-unit) activations be grouped into a small number of clusters while at the same time preserving the accuracy of the network [SL97]. Current approaches to rule extraction from MFNs have to rely on a procedure which can only be applied after having trained the MFN [SL97]. In contrast, BP-SOM yields the discretisation automatically during learning; no additional discretisation method need to be performed. Future work should elaborate on exploiting the discretisation of hidden-unit activations. 3.2 SOM-based rule extraction In our experiments, we observed that the reliability of SOM elements of successfully trained BP-SOM networks is often very high. This implies that the SOM can be seen as an organiser of the learning material. It divides the material into a limited number of subsets (bounded by the number of SOM elements). Within subsets, all instances are associated with the same class assuming a reliability of the SOM element of 1.. This automatic division into homogeneous subsets can be a useful step in automatic rule extraction. As an example, we trained BP-SOM networks on the monks-1 task [Pre94], a well-known benchmark problem for rule extraction. The task is to classify an instance (a1, a2, a3, a4, a5, a6) with six attributes. The possible values of the attributes are: a1 2 f1; 2; 3g; a2 2 f1; 2; 3g; a3 2 f1; 2g; a4 2 f1; 2; 3g; a5 2 f1; 2; 3; 4g; a6 2 f1; 2g. An instance is mapped to a class 1 when (a1 = a2) or (a5 = 1), and to class otherwise. The task is considered to be easily learnable. We used MFNs with one hidden layer of 1 units, 5 5-sized SOMs, and the same experimental settings as described earlier. After training (at which we obtained a generalisation accuracy of 1% on the test set), we collected for each labelled SOM element all associated training instances. Table 2 lists all the instances associated with the SOM element at SOM coordinates (1,1) (see also Figure 5). The subset of instances associated with SOM element (1,1) are all classified as 1. Moreover, two attributes display a constant value in the subset, viz. attributes a1 (having value 1) and a2 (having value 1); all other attributes display varying values. This regularity can be exploited to form a rule, which states that if a1 = 1 and a2 = 1, the corresponding class is 1. This rule extraction procedure is formalised as follows. For each SOM element, an IF : : : T HEN rule is extracted; it is composed of a conjunction of all attribute values, having a constant value throughout the instance subset associated (they are concatenated on the left-hand side of the rule), and of the classification of the instances (on the right-hand side). This procedure, when applied to SOM element (1,1), results in the rule IF (a1 = 1) & (a2 = 1) T HEN class= 1. Application of this procedure to all labelled SOM elements results in the rules visualised in Figure 5. We find that class 1 is covered by seven SOM elements, and that class is covered by four SOM elements. Although none of the SOM elements fully represent the underlying concept of the monks-1 problem, each extracted rule covers a part of the problem by focusing only on the relevant attributes a1, a2, or a5. The simple rule extraction procedure given

Attributes a1 a2 a3 a4 a5 a6 Class 1 2 3 1 2 3 1 2 1 2 3 1 2 3 4 1 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 Table 2: List of instances of the monks-1 task associated with SOM element (1,1) of a trained BP-SOM network. above cannot produce rules expressing the dependency of attribute values (e.g., a1 = a2), disjunctions, or nested conditions, nor does it express weighting and ordering of rules. However, we stress the fact that the clustering of subsets of instances on the SOM leads to intelligible trained MFNs and provides an important first step in rule extraction. 4 Conclusions In this paper we focused on two effects displayed by BP-SOM learning. First, we showed that after BP-SOM training, activations of hidden units of MFNs tend to oscillate among a limited number of discrete values. The effect is illustrated on the parity-12 task. The discretisation ability makes BP-SOM-trained MFNs suited for rule extraction, without having to resort to

1 2 3 4 5 1 IF a1=1 & a2=1 class= 1 IF a1<>3 & a2<>2 & a5=1 IF a1<>2 & a2<>3 & a5=1 IF a1=2 & a2=3 & a5=1 2 IF a1<>1 & a2<>1 & a5=1 3 IF a1<>2 & IF a1<>2 & a2<>3 & a2<>3 & a5<>1 a5<>1 class= class= IF a1<>1 & a2<>1 & a5<>1 IF a1<>1 & a2<>1 & a5<>1 4 IF a1=3 & a2=2 & a5=1 5 IF a1<>3 & a2<>2 & a5<>1 class= IF a1=2 & a2=3 & a5<>1 class= IF a1=3 & a2=2 & a5<>1 class= Figure 5: Graphic representation of a 5 5 SOM of a BP-SOM network trained on the monks-1 task. Labelled elements contain the IF : : : T HEN rules directly extracted from the instances matching these elements. Unlabelled elements are displayed as grey squares. post-processing rule-extraction methods. Second, we identified SOM elements as adequate organisers of instances of the task at hand. This effect is visualised by inspecting the organisation of instances in subsets on the elements of the SOM. When trained on the monks-1 task, the SOM of BP-SOM can be exploited straightforwardly as a first step in automatic rule extraction. In sum, both effects lead to intelligible neural networks and can be employed as a basis for automatic rule extraction. From the reported results and previous reports on BP-SOM we conclude that BP-SOM addresses essential issues in neural-network research. By letting supervised and unsupervised learning cooperate, BP-SOM aims at finding a combined solution to problems of overfitting, simplifying MFNs (either by pruning or by simplifying hidden-unit representations), and automatic rule extraction. Acknowledgement We wish to thank Eric Postma for stimulating discussions that provided part of the motivation for this work.

References [Hin86] Hinton, G. E. (1986). Learning distributed representations of concepts. In Proceedings of the Eighth Annual Conference of the Cognitive Science Society, 1 12. Hillsdale, NJ: Erlbaum. [Koh89] Kohonen, T. (1989). Self-organisation and Associative Memory. Berlin: Springer Verlag. [Nor89] Norris, D. (1989). How to build a connectionist idiot (savant). Cognition, 35, 277 291. [Pre94] Prechelt, L. (1994). Proben1: A set of neural network benchmark problems and benchmarking rules. Technical Report 24/94, Fakultät für Informatik, Universität Karlsruhe, Germany. [RHW86] Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986). Learning internal representations by error propagation. In D. E. Rumelhart and J. L. McClelland (Eds.), Parallel Distributed Processing: Explorations in the Microstructure of Cognition, volume 1: Foundations (pp. 318 362). Cambridge, MA: The MIT Press. [Sch93] Schaffer, C. (1993). Overfitting avoidance as bias. Machine Learning, 1, 153 178. [SL97] Setiono, R. and Liu, H. (1997). NeuroLinear: A system for extracting oblique decision rules from neural networks. In M. Van Someren and G. Widmer (Eds.), Proceedings of the Ninth European Conference on Machine Learning. Lecture Notes in Computer Science 1224. Berlin: Springer Verlag, 221 233. [Tho95] Thornton, C. (1995). Measuring the difficulty of specific learning problems. Connection Science, 7, 81 92. [Thr91] Thrun, S. B., Bala, J., Bloedorn, E., Bratko, I., Cestnik, B., Cheng, J., De Jong, K., Dˇzeroski, S., Fahlman, S. E., Fisher, D., Hamann, R., Kaufman, K., Keller, S., Kononenko, I., Kreuziger, J., Michalski, R. S., Mitchell, T., Pachowicz, P., Reich, Y., Vafaie, H., Van de Velde, W., Wenzel, W., Wnek, J., and Zhang, J. (1991). The MONK s Problems: a performance comparison of different learning algorithms. Technical Report CMU-CS-91-197, Carnegie Mellon University. [Wei95] Weijters, A. (1995). The BP-SOM architecture and learning rule. Neural Processing Letters, 2, 13 16. [Wei96] Weijters, A. (1996). BP-SOM: A profitable cooperation. In J.-J. Ch. Meyer and L. C. Van der Gaag (Eds.), Proceedings of the Eighth Dutch Conference on Artificial Intelligence, NAIC 96, 381 391. [WVVP97] Weijters, A., Van den Herik, H. J., Van den Bosch, A., and Postma, E. O. (1997). Avoiding overfitting with BP-SOM. To appear in Proceedings of IJCAI 97. [WK91] Weiss, S. and Kulikowski, C. (1991). Computer Systems that Learn. San Mateo, CA: Morgan Kaufmann.