A New Approach to Three Ensemble Neural Network Rule Extraction Using Recursive-Rule extraction algorithm

Size: px
Start display at page:

Download "A New Approach to Three Ensemble Neural Network Rule Extraction Using Recursive-Rule extraction algorithm"

Transcription

1 Proceedings of International Joint Conference on Neural Networks, Dallas, Texas, USA, August 4-9, 2013 A New Approach to Three Ensemble Neural Network Rule Extraction Using Recursive-Rule extraction algorithm Yoichi Hayashi, Ryusuke Sato and Sushmita Mitra Abstract In this paper, we propose a Three Ensemble neural network rule extraction algorithm. Then we investigate Hayashi s first question, Can the Ensemble-Recursive-Rule extraction (E-Re-RX) algorithm be extended to an ensemble neural network consisting of three or more MLPs and extract comprehensible rules? The E-Re-RX algorithm is an effective rule extraction algorithm for dealing with data sets that mix discrete and continuous attributes. Using the experimental results, we consider the three MLP ensemble Re-RX algorithm from various points of view. Finally, we present provisional positive conclusions. I. INTRODUCTION A. Rule Extraction from Neural Network Ensemble In 1990, Hansen and Salamon [1] showed that the generalization ability of learning systems based on artificial neural networks can be significantly improved through ensembles of artificial neural networks, i.e., training multiple artificial neural networks and combining their predictions via voting. Since combining ensembles network remarkably well, it became a very popular topic in both neural network and machine learning communities [2]. Although many authors have generated comprehensible models from individual networks, much less work has been done to explain neural network ensembles [3]. Bologna proposed the Discretized Interpretable Multi-Layer Perceptron (DIMLP) model with generated rules from neural network ensembles [4]. The DIMLP is a special neural network model for which symbolic rules are generated to clarify the knowledge embedded within connections and activation neurons. Bologna described how to translate symbolic rules into the DIMLP and how to extract rules from one or several combined neural networks. The Rule Extraction From Network Ensemble (REFNE) [2] approach proposed by Zhou et al. is designed to extract symbolic rules from trained neural network ensembles that perform classification tasks. REFNE utilizes trained ensembles to generate a number of instances and then extracts rules from those instances. REFNE can gracefully break the ties made by individual neural networks in prediction [2]. Yoichi Hayashi. Author is with the Department of Computer Science, Meiji University, Tama-ku, Kawasaki , Japan ( hayashiy@cs.meiji.ac.jp ) Ryusuke Sato. Author is with the Department of Computer Science, Meiji University, Tama-ku, Kawasaki , Japan ( sally36505@hotmail.com ) Sushmita Mitra. Author is with the Machine Intelligent Unit, Indian Statistical Institute, Kolkaka , India ( sushmita@isical.ac.in ) Zhou et al. [5] analyzed the relationship between the ensemble and its component neural networks from the context of both regression and classification. Their work revealed that it may be better to ensemble many instead of all available neural networks. This result is interesting because most approaches ensemble all available neural networks for prediction. In 2012, Hara and Hayashi proposed two ensemble neural network rule extraction algorithms. The former is for two-class classification [6]. The latter is for multiple-class classification [7]. Both of these algorithms use standard MLPs and the Re-RX algorithm proposed by Setiono [8]. The recognition accuracy of these algorithms is very high. II. PURPOSE OF THIS PAPER In 2013, Hayashi delivered a survey paper on neural data analysis using the ensemble concept and provided three open questions [9] as future work regarding the Ensemble-Recursive Rule extraction (E-Re-RX) algorithm. In this paper, we investigate the first one, Can the E-Re-RX algorithm be extended to an ensemble neural network consisting of three or more MLPs and extract comprehensible rules? III. STRUCTURE OF THE E-RE-RX ALGORITHM A. Re-RX Algorithm The Re-RX algorithm [8] is designed to generate classification rules from data sets that have both discrete and continuous attributes. The algorithm is recursive in nature and generates hierarchical rules. The rule conditions for discrete attributes are disjointed from those for continuous attributes. The continuous attributes only appear in the conditions of the rules lowest in the hierarchy. The outline of the algorithm is as follows. Algorithm Re-RX(S, D, C) Input: Dataset S having discrete attributes D and continuous attributes C. Output: A set of classification rules. 1. Train and prune a neural network [10] using the dataset S and all its D and C attributes. 2. Let D and C be the sets of discrete and continuous attributes, respectively, still present in the network, and let S be the of data samples correctly classified by the pruned network. 3. If D has associated continuous attributes dataset C, /13/$ IEEE 835

2 generate a hyperplane to split the samples in S according to the values of the continuous attributes C, then stop. Otherwise, by using only the discrete attributes D, generate the set of classification rules R for data set S. For each rule Ri generated: If support(ri)>δ 1 and error(ri)>δ 2, then - Let Si be the set of data samples that satisfy the condition of rule Ri and let Di be the set of discrete attributes that do not appear in the rule condition of Ri. - If D has associated continuous attributes C, generate a hyperplane to split the data in Si according to the values of the continuous attributes Ci, then stop. Otherwise, call Re-RX(Si, Di, Ci). In the above, we define the support of (Ri) to be the proportion of dataset covered by rule Ri and the error of (Ri) to be the proportion of data it incorrectly classifies. B. Three Ensemble-Re-RX (E-Re-RX) Algorithm We proposed the Two Ensemble-Recursive-Rule extraction (E-Re-RX) algorithm [6],[7], [9]. In this algorithm, primary rules are generated, followed by secondary rules to handle only those instances that do not satisfy the primary rules, and then these rules are integrated. We showed that this reduces the complexity of using multiple neural networks. This method achieves an extremely high recognition accuracy, even with multiclass problems. In this paper, we restrict ourselves to backpropagation neural networks (MLPs) with one hidden unit because such networks have been shown to possess an universal approximation property. With the neural network ensemble, it is possible to determine the final output by integrating the outputs of three neural networks. With the E-Re-RX algorithm, it is possible to determine the overall final output by integrating the rules. This rule integration enables the reduction of the number of neural networks and irrelevant rules. The essentials of the Three E-Re-RX algorithm are outlined as follows: Fig. 1. Schematic diagram of Three Ensemble-Re-RX algorithm Three Ensemble-Re-RX algorithm Inputs: Learning data sets LD, LDf, LDff Outputs: Primary rule set, secondary rule set, and tertiary rule set 1) Randomly extract data samples of an arbitrary proportion from learning data set LD, and name the set of extracted data samples LD. 2) Train and prune [10] the first neural network using LD. 3) Apply the Re-RX algorithm to the output of step 2, and output the primary rule set. 4) Based on these primary rules, create the set LDf from data samples that do not satisfy the rules. 5) Randomly extract data samples of an arbitrary proportion from learning data set LDf, and name the set of extracted data samples LDf. 6) Train and prune [10] the two ensemble neural network using LDf. 7) Apply the Re-RX algorithm to the output of step 6, and output the secondary rule set. 8) Integrate the primary rule set and the secondary rule set. 9) Based on the rules integrated in step 8, create the set LDff of data samples that do not satisfy these rules. 10) Randomly extract data samples of an arbitrary proportion from learning data set LDff, and name the set of extracted data LDff. 11) Train and prune [10] the three ensemble neural network using LDff. 12) Apply the Re-RX algorithm to the output of step 11, and output the tertiary rule set. 13) Integrate the primary rule set, the secondary rule set, and the tertiary rule set. The following provides supplemental descriptions to the respective steps above. First, in step 1, LD is created with the number of data of an arbitrary proportion of the learning data set. Data samples for LD are randomly selected. The main purpose for selecting data of an arbitrary proportion is to study the influence that the selected proportion imposes on the three ensemble neural network. Moreover, extraction of dataset at an arbitrary proportion enables learning with a smaller learning dataset than that needed in conventional methods. This methodology provides the benefit of preventing overfitting, because a smaller the number of data generally provides fewer optimum local solutions. Thus, in the event of a local solution, the total number of local solutions is few, and sufficient recognition accuracy can be shown. Data samples that cause overfitting can also be eliminated. Dataset that trigger overfitting are considered disparate among all the dataet. Fewer the number of data in the learning dataset lessen the degree of disparity from the whole perspective and can suppress overfitting. In step 2, both learning and pruning use the LD set created in step 1. The learning process and pruning algorithm adopt the methodology in cited literature [10]. Note that the intermediate unit of the neural network is one, which means that attributes left after pruning are those strongly required for 836

3 classifying the classes. Thus, overall rules are expressed with fewer attributes, which helps to secure recognition. In step 3, rules are extracted according to the Re-RX algorithm in cited literature [8]. The extracted rules are appraised, and rules to be re-extracted are selected. The J4.8[12] algorithm is used to extract rules. The rule set ultimately obtained here is considered the primary rule set. In step 4, LDf is created by using the primary rule set obtained in step 3. LDf dataset are selected from the learning dataset but do not satisfy the primary rule set, and so these are established as a dataset. Datasets satisfying the primary rule set are not considered, because such samples, if used for learning and pruning and rule extracting, would lead to extraction of an identical rule set, which would be totally extraneous. By learning and pruning with dataset that do not satisfy the primary rule set, extraction of the input attribute, output relations and rules is possible for coverage not provided by the primary rule set. In step 5, LDf is created with the number of data of an arbitrary proportion of LDf. In step 6, learning and pruning is done in the same manner as in step 2, which uses LDf. According to the Re-RX algorithm, the obtained output is appraised, and rules to be re-extracted are selected in step 7 in the same manner as in step 3 to yield the secondary rule set. In step 8, the obtained primary and secondary rule sets are integrated. The following rules are defined and followed in integrating the rule sets. 1) When all of the attributes and their values appearing in a rule are exactly identical and the class label is identical, then the rule in the secondary rule set is integrated into the primary rule set. 2) When all of the attributes and their values appearing in a rule are partially picked up in another rule whose attributes and associated values are identical, then the latter rule is integrated into the former encompassing rule, regardless of the class label. 3) Conflicting rules are integrated into the primary rule set. The foregoing integrating rules are described with reference to some examples. First, the rule in 1) is obvious. If attributes and their values appearing in the two rules are entirely identical, and their class labels are the same, the rules are identical. Thus, the secondary rule set is integrated into the primary rule set. For example, let us assume that within the primary rule set the following rule is obtained, R: If D42 = 0, then predict Class 1, and within the secondary rule set, the following rule is obtained, Rf: If D42 = 0, then predict Class 1. In this case, the attributes and their values appearing in the rules are entirely identical, and their class labels are the same. In such a case, the rules are deemed identical and integrated with R. Next, we consider the second rule. For this purpose, the following example provides the state of an extracted rule that satisfies this second rule. We assume in the primary rule set the following rule is obtained: R: If D42 = 1 and D38 = 0 and D43 = 0 and D27 = 0 and D24 = 0 and D45 = 0 and D2 = 0 and D21 = 1, then predict Class 1. In the secondary rule set the following rule is obtained: Rf: If D24 = 0 and D2 = 0 and D45 = 0 and D21 = 1, then predict Class 2. In this case, the attributes appearing in Rf are encompassed by R. Their values are also understood to be identical. In this case, although the class labels for rules R and Rf are different, integration is possible from the viewpoint that Rf is included in R, because these if-then rule expressions were originally prepared from decision trees. A single rule can be considered a branch within a decision tree, which although different, derives the same input and output rules. If the remaining attributes of one rule are considered to impose a serious influence in determining the class label, then the longer rule includes the shorter rule. Thus, integration is thought to be possible under this second rule. Finally, we consider the tertiary rule, which handles conflicting rules between the primary rule set and the secondary rule set. Two situations are envisioned for the rules. First, attributes and their values appearing in the rules are entirely identical, but the class labels are different. Second, attributes appearing in the rules are entirely identical but take on different values, while the class labels are identical. For these conflicting rules, in both cases the subject rule within the secondary rule set is integrated with the subject rule within the primary rule set. The grounds for this are related to the decision-tree generation method of the J4.8 algorithm [12]. Generation of the J4.8 algorithm is based on the data volume and the entropy of the dataset. Branches of the decision tree are neither generated in excess nor under low probability. Thus, in comparing the number of data, the primary rule set always is larger than the secondary rule set, and the decision tree of the former reflects the entire learning set more than that of the latter in terms of branches generated, when the entire data set is considered. For example, let us assume that within the primary rule set the following rule is obtained, R: If D42 = 0, then predict Class 1.0, and within the secondary rule set the following rule is obtained, Rf: If D42 = 0, then predict Class 2. These rules compare rules whose appearing attributes and associated values are identical, but whose class labels are different. In this case Rf of the secondary rule set is integrated with R as the sole rule. In the case above, rule Rf of the secondary rule set is integrated with rule R of the primary rule set; however, the second rule set naturally may have R integrated with Rf, which needs to be monitored. In step 9, the integrated rules obtained in step 8 are used to prepare data set LDff that does not satisfy those rules. In step 10, LDff is created with the number of data of an 837

4 arbitrary proportion of LDff. In step 11, learning and pruning are conducted in the same manner as that in step 2 by using LDff. In step 12, according to the Re-RX algorithm, the obtained output is appraised, and the rules to be re-extracted are selected in the same manner as in step 3 to yield the tertiary rule set. In step 13, the obtained primary, secondary, and tertiary rule sets are integrated. Execution of the above 13 steps completes the E-Re-RX algorithm. IV. EXPERIMENTS AND RESULTS A. Characteristics of Data Sets The data set used for the testing comes from data sets obtained from the UCI Repository [11]. The data set in this study utilized data sets having both discrete value attributes and continuous value attributes, for the following two reasons: Re-RX must be valid for data sets having both discrete value attributes and continuous value attributes, and many data sets of real-world problems consist of discrete value attributes and continuous value attributes. In essence, a method demonstrated as valid against data sets having both discrete value attributes and continuous value attributes are more practical. The following Table 1 shows the number of data sample, their number of discrete value attributes, their number of continuous value attributes, and their number of classes. TABLE 1 DATA SET CHARACTERISTICS Card (0.5) Card (0.7) German Credit Card (0.65) Data Attributes Discrete Attributes Continuous Attributes Classes Card dataset [11] was tested in two ways: proportions of 0.5 and 0.7 when LD was created. The purpose was to compare influences on the recognition accuracy and the number of rules depending on the proportion extracted. The German Credit Card data [11] set is also a data set concerning credit cards. When it was obtained from the UCI Repository, it consisted of so-called clustering of alphabetic classifications, and discrete values. The dataset was used after this clustering was substituted with 0 s and 1 s, and discrete value attributes were normalized and amended to continuous values in the range between 0 and 1. The German Credit Card data set was tested at a proportion set to 0.65 when LD was created. B. Results for German Credit Data The German Credit Card data samples were divided into training data and test data. As the divided proportions, training data was 50% of all 1,000 data (500 data). The remaining 50% (500 data) was test data. Learning and pruning of the first neural network used LD' training data at a proportion of 65% that consisted of 325 data samples. From the results, rules were extracted from a decision tree according to J4.8, and rules were re-extracted according to the Re-RX algorithm. The thresholds of rule re-extraction of the Re-RX algorithm were δ 1 = δ 2 = The following primary rule set was obtained. Primary rule set R1: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=0 and D43=0, then R1a: If C5 <= : Class 1 R1b: If C5 > : Class 2 R2: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=0 and D43=1, then Class 2 R3: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=1, then Class 2 R4: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=1, then Class 1 R5: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=1, then Class 1 R6: If D1=0 and D29=0 and D49=0 and D15=0 and D9=1, then Class 2 R7: If D1=0 and D29=0 and D49=0 and D15=1, then Class 2 R8: If D1=0 and D29=0 and D49=1, then Class 2 R9: If D1=0 and D29=1, then Class 1 R10: If D1=1, then Class 1 The recognition rate of the test dataset by the primary rule set was 71.8%. LDf was created, based on this primary rule set. LDf became a data set composed of 139 data. LDf was randomly extracted at a proportion of 65%, that is, 90 data, as LDf, which was used for learning and pruning of the two ensemble neural networks. Re-extraction of the rules was performed in the same manner as in the first neural network. Consequently, the following secondary rule set was obtained. Secondary rule set Rf1: If D20=0 and D2=0 and D15=0, then Class 2 Rf2: If D20=0 and D2=0 and D15=1, then Class 1 Rf3: If D20=0 and D2=1 and D21=0, then Class 1 Rf4: If D20=0 and D2=1 and D21=1, then Class 2 Rf5: If D20=1 and D1=0, then Class 1 (included in R6) Rf6: If D20=1 and D1=1, then Class 2 838

5 A comparison between the primary rule set and the secondary rule set showed that rule Rf5 was included in R6. Therefore, Rf5 was integrated with R6 at this time. The integrated rule set to the secondary rule set became as follows. Integrated Rules (primary rule set and secondary rule set) R1: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=0 and D43=0, then R1a: If C5 <= : Class 1 R1b: If C5 > : Class 2 R2: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=0 and D43=1, then Class 2 R3: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=1, then Class 2 R4: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=1, then Class 1 R5: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=1, then Class 1 R6: If D1=0 and D29=0 and D49=0 and D15=0 and D9=1, then Class 2 R7: If D1=0 and D29=0 and D49=0 and D15=1, then Class 2 R8: If D1=0 and D29=0 and D49=1, then Class 2 R9: If D1=0 and D29=1, then Class 1 R10: If D1=1, then Class 1 R11: If D20=0 and D2=0 and D15=0, then Class 2 R12: If D20=0 and D2=0 and D15=1, then Class 1 R13: If D20=0 and D2=1 and D21=0, then Class 1 R14: If D20=0 and D2=1 and D21=1, then Class 2 R15: If D20=1 and D1=1, then Class 2 The recognition rate of the test data set to the secondary rule set was 92.8%. LDff was created based on this rule set. LDff became a data set consisting of 27 data. LDff was randomly extracted at a proportion of 65%, that is, 18 data, as LDff, which was used for learning and pruning of the three ensemble neural networks. Re-extraction of the rules was performed in the same manner as the one ensemble neural network. Consequently, the following tertiary rule set was obtained. Integrated rules (primary rule set, secondary rule set, and tertiary rule set) R1: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=0 and D43=0, then R1a: If C5 <= : Class 1 R1b: If C5 > : Class 2 R2: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=0 and D43=1, then Class 2 R3: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=0 and D32=1, then Class 2 R4: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=0 and D51=1, then Class 1 R5: If D1=0 and D29=0 and D49=0 and D15=0 and D9=0 and D20=1, then Class 1 R6: If D1=0 and D29=0 and D49=0 and D15=0 and D9=1, then Class 2 R7: If D1=0 and D29=0 and D49=0 and D15=1, then Class 2 R8: If D1=0 and D29=0 and D49=1, then Class 2 R9: If D1=0 and D29=1, then Class 1 R10: If D1=1, then Class 1 R11: If D20=0 and D2=0 and D15=0, then Class 2 R12: If D20=0 and D2=0 and D15=1, then Class 1 R13: If D20=0 and D2=1 and D21=0, then Class 1 R14: If D20=0 and D2=1 and D21=1, then Class 2 R15: If D20=1 and D1=1, then Class 2 R16: If D4=0 and D18=0 and D21=0, then Class 2 R17: If D4=0 and D18=0 and D21=1, then Class 1 R18: If D4=0 and D18=1, then Class 1 R19: If D4=1, then R19a: If D20 = 0: Class 1 R19b: If D20 = 1: Class 2 The recognition rate of the test data set up to the tertiary rule set was 97.4%. The results thus far are shown in the following table. TABLE 2 GERMAN (0.65) EXPERIMENTAL RESULTS Tertiary rule set Rff1: If D4=0 and D18=0 and D21=0, then Class 2 Rff2: If D4=0 and D18=0 and D21=1, then Class 1 Rff3: If D4=0 and D18=1, then Class 1 Rff4: If D4=1 then R4a: If D20 = 0: Class 1 R4b: If D20 = 1: Class 2 A comparison between the integrated rules (primary rule set and the secondary rule set) and the tertiary rule set revealed no conflicts or encompassing rules. Thus, the following integrated rules were obtained at this time. V. DISCUSSION We examine the obtained results from various view point when three ensemble neural networks were used. The test results for the number of rules, recognition rates, and the number of data are summarized in the tables

6 TABLE 3 COMPARISON OF THE NUMBER OF RULES FOR EACH DATA SET A review of the table 4 shows that rule extractions increased with the increase in ensemble neural networks. For Card, however, rules were not extracted at the time of three ensembles. As neural networks increased in number, the proportion of increase became smaller. TABLE 4 COMPARISON OF RECOGNITION RATES Recognition rates rose with the increase in neural networks. The margin of increase was large from the first to the second, but this is not in proportion to the rate of increase in the number of rules. That is explained in the following table for the comparison of input dataset of the neural networks. TABLE 5 COMPARISON OF THE NUMBER OF INPUT DATA SET The number of data sets decreased in inverse proportion to the increase in neural networks. In the transfer to the next neural network, data that did not satisfy the rules previously extracted in a neural network were selected and used as input. Since the amount of input decreased in considerably high proportion from one ensemble neural network to two ensemble neural networks for both Card and German, it is understood that a lot of data satisfied the extracted rules. The recognition rate increased accordingly. Even if many rules are extracted, the recognition rate does not increase if little data satisfies the rules. It may be said that good rules that can satisfy an enormous amount of data efficiently with a single rule were extracted through this testing. Next, the possibility of rule extraction is examined, which is the purpose of this testing. A review of the results of rule extraction in Table 3 shows that rules were extracted when three neural networks were used for German. For Card, however, rules were not extracted. The comparison of the input counts in Table 5 shows that Card had inputs of only single digits at the stage of the third network, whereas German had two digits remaining. When the test was repeated for not only this table but also various data sets, it was found that all attributes were pruned and rules were not extracted in a neural network unless at least 2 digits of the third neural network input remained during learning and pruning in a neural network. Whether or not the number of inputs of "2 digits" could be the threshold to use three ensemble neural networks could not be concluded unconditionally, since not every data set was tested with every neural network. Nonetheless, the following three points were learned from the testing. 1) When there was little input, learning and pruning of a neural network caused most attributes to be pruned, and rules were not extracted. 2) Because Re-Rx used only recognized data for rule extraction, if there was little input, learning became insufficient, unrecognized data increased, and data used for rule extraction became small. 3) If there was little input, rules were not extracted when there was bias in the data. These outcomes happened many times when inputs were 10 or fewer in the course of repeating the testing. Next, recognition rates are related to the possibility of rule extraction. The recognition rate of Card surpassed 80% at the first neural network stages for both cases, as shown in Table 5. In contrast, for German the rate was a low value of 70%. In the algorithm of E-Re-RX, the inputs of the next neural network were a collection of data that did not satisfy the rules extracted in the previous neural network. Data that did not satisfy a rule became smaller and the input to the next neural network decreased as the recognition rate improved. Since the recognition rate of one neural network was high to some extent from the beginning in the data set of Card, the proportion of input that remained in the two ensemble neural networks was lower for Card than for German. Consequently, a tertiary rule set was not extracted in Card, and in contrast, the recognition rate for German, which did extract a tertiary rule set, ultimately became higher. Thus, for the possibility of rule extraction by recognition rate, it can be said that the data sets with low recognition rates in one neural network tend to have rules easily extracted by a neural network of two ensembles or more. Comparing the complexity of data between Card and German revealed that Card consists of 690 data samples with 51 attributes, and German consists of 1,000 data samples with 63 attributes. When the data was more complicated, the effectiveness in extracting rules by neural networks of two ensembles and three ensembles will be improved. 840

7 VI. CONCLUSION The testing results reported in this paper show that rules were not extracted in three ensemble neural networks for the Card data, whereas rules were extracted for the German data. The complexity of data (the number of data and the number of attributes), recognition rates, and the number of input were the predominant causes. The recognition rate for the primary rule set was not so high when the data was complicated. Consequently, the data which did not satisfy the rules increased, and the possibility of rule extraction increased for the secondary rule set and the tertiary rule set. The E-Re-RX algorithm, described in recent research [6, 7, 9], can achieve high recognition rates when rules are extracted. The current testing was performed by using neural networks limited to three, but higher recognition rates can be expected when four or five can be applied. Conditions for the handled data sets become necessary, however, to apply up to four and five. Data sets that can be classified at the stage of a primary rule set, like the Card data, are not suitable for an ensemble neural network. If data is classified well at an early stage, the effectiveness of many ensemble neural networks put to use is validated. The recognition rates for using two ensembles also achieved considerably high values, as the experimental results show, but the foremost advantage of using three ensemble neural networks was a recognition rate even higher than the high recognition rate of using two. For data sets that are not simple and less likely to see much increase in the recognition rate in one neural network only, the effectiveness in using multiple ensemble neural networks can be expected to increase. Empirically, these tests confirmed that rule extractions were successful in using Three Ensemble-Rule extraction for a mono-class and a complicated data set such as German. However, testing of rule extraction that used three ensemble neural networks on a multi-class data set such as Thyroid [11] was not necessarily successful. Such testing will be a future research topic. Possible development beyond the current testing includes the dynamic varying and optimizing of the proportion of classification between training data and test data. For German data, as an example, the proportion of classification was set at 50%. When training data was classified one time at a proportion of 75% during the course of testing and the testing was attempted, there was too much training. The attributes generally did not get pruned and remained in the first neural network, and an enormous number of primary rules were extracted. Finally, as additional future work, we should investigate the following an open question on the Three Ensemble-Re-RX algorithm: 1) Can the Three Ensemble-Re-RX algorithm be extended to an ensemble neural network consisting of three or more MLPs and extract comprehensible rules for the multi-class classification problem? REFERENCES [1] L.K. Hansen and P. Salamon, Neural network ensembles, IEEE Trans. Pattern Analysis and Machine Intelligence, Vol. 12, No , [2] Z.-H. Zhou, Extracting symbolic rules from trained neural network ensembles, AI Communications, Vol. 16, pp. 3 15, , Vol. 19, pp , [3] G. Bologna, Is it worth generating rules from neural network ensembles?, J. of Applied Logic, Vol. 2, pp , [4] G. Bologna, A model for single and multiple knowledge based networks," Artificial Intelligence in Medicine, Vol. 28, pp , [5] Z.-H. Zhou, Ensemble neural networks: Many could be better than all, Artificial Intelligence, Vol. 137, pp , [6] A. Hara and Y. Hayashi, Ensemble neural network rule extraction using Re-RX algorithm, in Proc. of WCCI (IJCNN) 2012, June 10 15, Brisbane, Australia, pp , [7] A. Hara and Y. Hayashi, A new neural data analysis approach using ensemble neural network rule extraction, In Proc. 22nd Int. Conf. Artificial Neural Netw., (Lecture Notes in Computer Science, Vol. 7552), Lausanne, Switzerland, September, pp , [8] R. Setiono, B. Baesens and C. Mues, Recursive neural network rule extraction for data with mixed attributes, IEEE Trans. Neural Netw., Vol. 19, pp , [9] Y. Hayashi, Neural data analysis: Ensemble neural network rule extraction approach and its theoretical and historical backgrounds, In Proc. 12th Int. Conf. Artificial Intelligence and Soft computing (Lecture Notes in Artificial Intelligence), Keynote Speech, Zakopane, Poland, June 9 13, 2013 (Accepted as an invited paperr). [10] R. Setiono, A penalty-function approach for pruning feedforward neural networks, Neural Comp., Vol. 9, No. 1, pp , 1977 [11] University of California, Irvine Machine Learning Repository, [12] 841

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

Rule Learning With Negation: Issues Regarding Effectiveness

Rule Learning With Negation: Issues Regarding Effectiveness Rule Learning With Negation: Issues Regarding Effectiveness S. Chua, F. Coenen, G. Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX Liverpool, United

More information

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System QuickStroke: An Incremental On-line Chinese Handwriting Recognition System Nada P. Matić John C. Platt Λ Tony Wang y Synaptics, Inc. 2381 Bering Drive San Jose, CA 95131, USA Abstract This paper presents

More information

Rule Learning with Negation: Issues Regarding Effectiveness

Rule Learning with Negation: Issues Regarding Effectiveness Rule Learning with Negation: Issues Regarding Effectiveness Stephanie Chua, Frans Coenen, and Grant Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX

More information

Softprop: Softmax Neural Network Backpropagation Learning

Softprop: Softmax Neural Network Backpropagation Learning Softprop: Softmax Neural Networ Bacpropagation Learning Michael Rimer Computer Science Department Brigham Young University Provo, UT 84602, USA E-mail: mrimer@axon.cs.byu.edu Tony Martinez Computer Science

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics Machine Learning from Garden Path Sentences: The Application of Computational Linguistics http://dx.doi.org/10.3991/ijet.v9i6.4109 J.L. Du 1, P.F. Yu 1 and M.L. Li 2 1 Guangdong University of Foreign Studies,

More information

Human Emotion Recognition From Speech

Human Emotion Recognition From Speech RESEARCH ARTICLE OPEN ACCESS Human Emotion Recognition From Speech Miss. Aparna P. Wanare*, Prof. Shankar N. Dandare *(Department of Electronics & Telecommunication Engineering, Sant Gadge Baba Amravati

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

Learning Methods in Multilingual Speech Recognition

Learning Methods in Multilingual Speech Recognition Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

OCR for Arabic using SIFT Descriptors With Online Failure Prediction OCR for Arabic using SIFT Descriptors With Online Failure Prediction Andrey Stolyarenko, Nachum Dershowitz The Blavatnik School of Computer Science Tel Aviv University Tel Aviv, Israel Email: stloyare@tau.ac.il,

More information

Reducing Features to Improve Bug Prediction

Reducing Features to Improve Bug Prediction Reducing Features to Improve Bug Prediction Shivkumar Shivaji, E. James Whitehead, Jr., Ram Akella University of California Santa Cruz {shiv,ejw,ram}@soe.ucsc.edu Sunghun Kim Hong Kong University of Science

More information

Word Segmentation of Off-line Handwritten Documents

Word Segmentation of Off-line Handwritten Documents Word Segmentation of Off-line Handwritten Documents Chen Huang and Sargur N. Srihari {chuang5, srihari}@cedar.buffalo.edu Center of Excellence for Document Analysis and Recognition (CEDAR), Department

More information

Mining Association Rules in Student s Assessment Data

Mining Association Rules in Student s Assessment Data www.ijcsi.org 211 Mining Association Rules in Student s Assessment Data Dr. Varun Kumar 1, Anupama Chadha 2 1 Department of Computer Science and Engineering, MVN University Palwal, Haryana, India 2 Anupama

More information

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING Gábor Gosztolya 1, Tamás Grósz 1, László Tóth 1, David Imseng 2 1 MTA-SZTE Research Group on Artificial

More information

(Sub)Gradient Descent

(Sub)Gradient Descent (Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include

More information

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees Mariusz Łapczy ski 1 and Bartłomiej Jefma ski 2 1 The Chair of Market Analysis and Marketing Research,

More information

Applications of data mining algorithms to analysis of medical data

Applications of data mining algorithms to analysis of medical data Master Thesis Software Engineering Thesis no: MSE-2007:20 August 2007 Applications of data mining algorithms to analysis of medical data Dariusz Matyja School of Engineering Blekinge Institute of Technology

More information

Chapter 2 Rule Learning in a Nutshell

Chapter 2 Rule Learning in a Nutshell Chapter 2 Rule Learning in a Nutshell This chapter gives a brief overview of inductive rule learning and may therefore serve as a guide through the rest of the book. Later chapters will expand upon the

More information

CSL465/603 - Machine Learning

CSL465/603 - Machine Learning CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am

More information

On-Line Data Analytics

On-Line Data Analytics International Journal of Computer Applications in Engineering Sciences [VOL I, ISSUE III, SEPTEMBER 2011] [ISSN: 2231-4946] On-Line Data Analytics Yugandhar Vemulapalli #, Devarapalli Raghu *, Raja Jacob

More information

Axiom 2013 Team Description Paper

Axiom 2013 Team Description Paper Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

Abstractions and the Brain

Abstractions and the Brain Abstractions and the Brain Brian D. Josephson Department of Physics, University of Cambridge Cavendish Lab. Madingley Road Cambridge, UK. CB3 OHE bdj10@cam.ac.uk http://www.tcm.phy.cam.ac.uk/~bdj10 ABSTRACT

More information

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Twitter Sentiment Classification on Sanders Data using Hybrid Approach IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 4, Ver. I (July Aug. 2015), PP 118-123 www.iosrjournals.org Twitter Sentiment Classification on Sanders

More information

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks Devendra Singh Chaplot, Eunhee Rhim, and Jihie Kim Samsung Electronics Co., Ltd. Seoul, South Korea {dev.chaplot,eunhee.rhim,jihie.kim}@samsung.com

More information

Evolution of Symbolisation in Chimpanzees and Neural Nets

Evolution of Symbolisation in Chimpanzees and Neural Nets Evolution of Symbolisation in Chimpanzees and Neural Nets Angelo Cangelosi Centre for Neural and Adaptive Systems University of Plymouth (UK) a.cangelosi@plymouth.ac.uk Introduction Animal communication

More information

Lecture 10: Reinforcement Learning

Lecture 10: Reinforcement Learning Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation

More information

Classification Using ANN: A Review

Classification Using ANN: A Review International Journal of Computational Intelligence Research ISSN 0973-1873 Volume 13, Number 7 (2017), pp. 1811-1820 Research India Publications http://www.ripublication.com Classification Using ANN:

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

Learning From the Past with Experiment Databases

Learning From the Past with Experiment Databases Learning From the Past with Experiment Databases Joaquin Vanschoren 1, Bernhard Pfahringer 2, and Geoff Holmes 2 1 Computer Science Dept., K.U.Leuven, Leuven, Belgium 2 Computer Science Dept., University

More information

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1 Notes on The Sciences of the Artificial Adapted from a shorter document written for course 17-652 (Deciding What to Design) 1 Ali Almossawi December 29, 2005 1 Introduction The Sciences of the Artificial

More information

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Todd Holloway Two Lecture Series for B551 November 20 & 27, 2007 Indiana University Outline Introduction Bias and

More information

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Ajith Abraham School of Business Systems, Monash University, Clayton, Victoria 3800, Australia. Email: ajith.abraham@ieee.org

More information

Knowledge Transfer in Deep Convolutional Neural Nets

Knowledge Transfer in Deep Convolutional Neural Nets Knowledge Transfer in Deep Convolutional Neural Nets Steven Gutstein, Olac Fuentes and Eric Freudenthal Computer Science Department University of Texas at El Paso El Paso, Texas, 79968, U.S.A. Abstract

More information

Lecture 1: Basic Concepts of Machine Learning

Lecture 1: Basic Concepts of Machine Learning Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010

More information

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany Jana Kitzmann and Dirk Schiereck, Endowed Chair for Banking and Finance, EUROPEAN BUSINESS SCHOOL, International

More information

Calibration of Confidence Measures in Speech Recognition

Calibration of Confidence Measures in Speech Recognition Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE

More information

Laboratorio di Intelligenza Artificiale e Robotica

Laboratorio di Intelligenza Artificiale e Robotica Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning

More information

Assignment 1: Predicting Amazon Review Ratings

Assignment 1: Predicting Amazon Review Ratings Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for

More information

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING Yong Sun, a * Colin Fidge b and Lin Ma a a CRC for Integrated Engineering Asset Management, School of Engineering Systems, Queensland

More information

How to Judge the Quality of an Objective Classroom Test

How to Judge the Quality of an Objective Classroom Test How to Judge the Quality of an Objective Classroom Test Technical Bulletin #6 Evaluation and Examination Service The University of Iowa (319) 335-0356 HOW TO JUDGE THE QUALITY OF AN OBJECTIVE CLASSROOM

More information

Australian Journal of Basic and Applied Sciences

Australian Journal of Basic and Applied Sciences AENSI Journals Australian Journal of Basic and Applied Sciences ISSN:1991-8178 Journal home page: www.ajbasweb.com Feature Selection Technique Using Principal Component Analysis For Improving Fuzzy C-Mean

More information

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina

More information

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE EE-589 Introduction to Neural Assistant Prof. Dr. Turgay IBRIKCI Room # 305 (322) 338 6868 / 139 Wensdays 9:00-12:00 Course Outline The course is divided in two parts: theory and practice. 1. Theory covers

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

INPE São José dos Campos

INPE São José dos Campos INPE-5479 PRE/1778 MONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND COVER CLASSIFICATION IN A NEDRAL NETWORK ENVIRONNENT Maria Suelena S. Barros Valter Rodrigues INPE São José dos Campos 1993 SECRETARIA

More information

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford

More information

Computerized Adaptive Psychological Testing A Personalisation Perspective

Computerized Adaptive Psychological Testing A Personalisation Perspective Psychology and the internet: An European Perspective Computerized Adaptive Psychological Testing A Personalisation Perspective Mykola Pechenizkiy mpechen@cc.jyu.fi Introduction Mixed Model of IRT and ES

More information

Test Effort Estimation Using Neural Network

Test Effort Estimation Using Neural Network J. Software Engineering & Applications, 2010, 3: 331-340 doi:10.4236/jsea.2010.34038 Published Online April 2010 (http://www.scirp.org/journal/jsea) 331 Chintala Abhishek*, Veginati Pavan Kumar, Harish

More information

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence

More information

Ordered Incremental Training with Genetic Algorithms

Ordered Incremental Training with Genetic Algorithms Ordered Incremental Training with Genetic Algorithms Fangming Zhu, Sheng-Uei Guan* Department of Electrical and Computer Engineering, National University of Singapore, 10 Kent Ridge Crescent, Singapore

More information

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF) SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF) Hans Christian 1 ; Mikhael Pramodana Agus 2 ; Derwin Suhartono 3 1,2,3 Computer Science Department,

More information

Disambiguation of Thai Personal Name from Online News Articles

Disambiguation of Thai Personal Name from Online News Articles Disambiguation of Thai Personal Name from Online News Articles Phaisarn Sutheebanjard Graduate School of Information Technology Siam University Bangkok, Thailand mr.phaisarn@gmail.com Abstract Since online

More information

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, 2013 10.12753/2066-026X-13-154 DATA MINING SOLUTIONS FOR DETERMINING STUDENT'S PROFILE Adela BÂRA,

More information

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18 Version Space Javier Béjar cbea LSI - FIB Term 2012/2013 Javier Béjar cbea (LSI - FIB) Version Space Term 2012/2013 1 / 18 Outline 1 Learning logical formulas 2 Version space Introduction Search strategy

More information

AQUA: An Ontology-Driven Question Answering System

AQUA: An Ontology-Driven Question Answering System AQUA: An Ontology-Driven Question Answering System Maria Vargas-Vera, Enrico Motta and John Domingue Knowledge Media Institute (KMI) The Open University, Walton Hall, Milton Keynes, MK7 6AA, United Kingdom.

More information

A NEW ALGORITHM FOR GENERATION OF DECISION TREES

A NEW ALGORITHM FOR GENERATION OF DECISION TREES TASK QUARTERLY 8 No 2(2004), 1001 1005 A NEW ALGORITHM FOR GENERATION OF DECISION TREES JERZYW.GRZYMAŁA-BUSSE 1,2,ZDZISŁAWS.HIPPE 2, MAKSYMILIANKNAP 2 ANDTERESAMROCZEK 2 1 DepartmentofElectricalEngineeringandComputerScience,

More information

Speech Emotion Recognition Using Support Vector Machine

Speech Emotion Recognition Using Support Vector Machine Speech Emotion Recognition Using Support Vector Machine Yixiong Pan, Peipei Shen and Liping Shen Department of Computer Technology Shanghai JiaoTong University, Shanghai, China panyixiong@sjtu.edu.cn,

More information

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Neural Network GUI Tested on Text-To-Phoneme Mapping A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis

More information

Evolutive Neural Net Fuzzy Filtering: Basic Description

Evolutive Neural Net Fuzzy Filtering: Basic Description Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:

More information

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford

More information

Issues in the Mining of Heart Failure Datasets

Issues in the Mining of Heart Failure Datasets International Journal of Automation and Computing 11(2), April 2014, 162-179 DOI: 10.1007/s11633-014-0778-5 Issues in the Mining of Heart Failure Datasets Nongnuch Poolsawad 1 Lisa Moore 1 Chandrasekhar

More information

NCEO Technical Report 27

NCEO Technical Report 27 Home About Publications Special Topics Presentations State Policies Accommodations Bibliography Teleconferences Tools Related Sites Interpreting Trends in the Performance of Special Education Students

More information

Laboratorio di Intelligenza Artificiale e Robotica

Laboratorio di Intelligenza Artificiale e Robotica Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning

More information

Australia s tertiary education sector

Australia s tertiary education sector Australia s tertiary education sector TOM KARMEL NHI NGUYEN NATIONAL CENTRE FOR VOCATIONAL EDUCATION RESEARCH Paper presented to the Centre for the Economics of Education and Training 7 th National Conference

More information

Automating the E-learning Personalization

Automating the E-learning Personalization Automating the E-learning Personalization Fathi Essalmi 1, Leila Jemni Ben Ayed 1, Mohamed Jemni 1, Kinshuk 2, and Sabine Graf 2 1 The Research Laboratory of Technologies of Information and Communication

More information

arxiv: v1 [cs.lg] 15 Jun 2015

arxiv: v1 [cs.lg] 15 Jun 2015 Dual Memory Architectures for Fast Deep Learning of Stream Data via an Online-Incremental-Transfer Strategy arxiv:1506.04477v1 [cs.lg] 15 Jun 2015 Sang-Woo Lee Min-Oh Heo School of Computer Science and

More information

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS Md. Tarek Habib 1, Rahat Hossain Faisal 2, M. Rokonuzzaman 3, Farruk Ahmed 4 1 Department of Computer Science and Engineering, Prime University,

More information

BENCHMARK TREND COMPARISON REPORT:

BENCHMARK TREND COMPARISON REPORT: National Survey of Student Engagement (NSSE) BENCHMARK TREND COMPARISON REPORT: CARNEGIE PEER INSTITUTIONS, 2003-2011 PREPARED BY: ANGEL A. SANCHEZ, DIRECTOR KELLI PAYNE, ADMINISTRATIVE ANALYST/ SPECIALIST

More information

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems Published in the International Journal of Hybrid Intelligent Systems 1(3-4) (2004) 111-126 Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems Ioannis Hatzilygeroudis and Jim Prentzas

More information

Time series prediction

Time series prediction Chapter 13 Time series prediction Amaury Lendasse, Timo Honkela, Federico Pouzols, Antti Sorjamaa, Yoan Miche, Qi Yu, Eric Severin, Mark van Heeswijk, Erkki Oja, Francesco Corona, Elia Liitiäinen, Zhanxing

More information

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH ISSN: 0976-3104 Danti and Bhushan. ARTICLE OPEN ACCESS CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH Ajit Danti 1 and SN Bharath Bhushan 2* 1 Department

More information

Proficiency Illusion

Proficiency Illusion KINGSBURY RESEARCH CENTER Proficiency Illusion Deborah Adkins, MS 1 Partnering to Help All Kids Learn NWEA.org 503.624.1951 121 NW Everett St., Portland, OR 97209 Executive Summary At the heart of the

More information

Data Fusion Models in WSNs: Comparison and Analysis

Data Fusion Models in WSNs: Comparison and Analysis Proceedings of 2014 Zone 1 Conference of the American Society for Engineering Education (ASEE Zone 1) Data Fusion s in WSNs: Comparison and Analysis Marwah M Almasri, and Khaled M Elleithy, Senior Member,

More information

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for Email Marilyn A. Walker Jeanne C. Fromer Shrikanth Narayanan walker@research.att.com jeannie@ai.mit.edu shri@research.att.com

More information

Reinforcement Learning by Comparing Immediate Reward

Reinforcement Learning by Comparing Immediate Reward Reinforcement Learning by Comparing Immediate Reward Punit Pandey DeepshikhaPandey Dr. Shishir Kumar Abstract This paper introduces an approach to Reinforcement Learning Algorithm by comparing their immediate

More information

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Notebook for PAN at CLEF 2013 Andrés Alfonso Caurcel Díaz 1 and José María Gómez Hidalgo 2 1 Universidad

More information

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach Data Integration through Clustering and Finding Statistical Relations - Validation of Approach Marek Jaszuk, Teresa Mroczek, and Barbara Fryc University of Information Technology and Management, ul. Sucharskiego

More information

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Proceedings of 28 ISFA 28 International Symposium on Flexible Automation Atlanta, GA, USA June 23-26, 28 ISFA28U_12 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Amit Gil, Helman Stern, Yael Edan, and

More information

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES Po-Sen Huang, Kshitiz Kumar, Chaojun Liu, Yifan Gong, Li Deng Department of Electrical and Computer Engineering,

More information

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Chinese Language Parsing with Maximum-Entropy-Inspired Parser Chinese Language Parsing with Maximum-Entropy-Inspired Parser Heng Lian Brown University Abstract The Chinese language has many special characteristics that make parsing difficult. The performance of state-of-the-art

More information

Using focal point learning to improve human machine tacit coordination

Using focal point learning to improve human machine tacit coordination DOI 10.1007/s10458-010-9126-5 Using focal point learning to improve human machine tacit coordination InonZuckerman SaritKraus Jeffrey S. Rosenschein The Author(s) 2010 Abstract We consider an automated

More information

TITLE 23: EDUCATION AND CULTURAL RESOURCES SUBTITLE A: EDUCATION CHAPTER I: STATE BOARD OF EDUCATION SUBCHAPTER b: PERSONNEL PART 25 CERTIFICATION

TITLE 23: EDUCATION AND CULTURAL RESOURCES SUBTITLE A: EDUCATION CHAPTER I: STATE BOARD OF EDUCATION SUBCHAPTER b: PERSONNEL PART 25 CERTIFICATION ISBE 23 ILLINOIS ADMINISTRATIVE CODE 25 TITLE 23: EDUCATION AND CULTURAL RESOURCES : EDUCATION CHAPTER I: STATE BOARD OF EDUCATION : PERSONNEL Section 25.10 Accredited Institution PART 25 CERTIFICATION

More information

Indian Institute of Technology, Kanpur

Indian Institute of Technology, Kanpur Indian Institute of Technology, Kanpur Course Project - CS671A POS Tagging of Code Mixed Text Ayushman Sisodiya (12188) {ayushmn@iitk.ac.in} Donthu Vamsi Krishna (15111016) {vamsi@iitk.ac.in} Sandeep Kumar

More information

Miami-Dade County Public Schools

Miami-Dade County Public Schools ENGLISH LANGUAGE LEARNERS AND THEIR ACADEMIC PROGRESS: 2010-2011 Author: Aleksandr Shneyderman, Ed.D. January 2012 Research Services Office of Assessment, Research, and Data Analysis 1450 NE Second Avenue,

More information

SARDNET: A Self-Organizing Feature Map for Sequences

SARDNET: A Self-Organizing Feature Map for Sequences SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu

More information

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 6, Ver. IV (Nov Dec. 2015), PP 01-07 www.iosrjournals.org Longest Common Subsequence: A Method for

More information

Different Requirements Gathering Techniques and Issues. Javaria Mushtaq

Different Requirements Gathering Techniques and Issues. Javaria Mushtaq 835 Different Requirements Gathering Techniques and Issues Javaria Mushtaq Abstract- Project management is now becoming a very important part of our software industries. To handle projects with success

More information

Utilizing Soft System Methodology to Increase Productivity of Shell Fabrication Sushant Sudheer Takekar 1 Dr. D.N. Raut 2

Utilizing Soft System Methodology to Increase Productivity of Shell Fabrication Sushant Sudheer Takekar 1 Dr. D.N. Raut 2 IJSRD - International Journal for Scientific Research & Development Vol. 2, Issue 04, 2014 ISSN (online): 2321-0613 Utilizing Soft System Methodology to Increase Productivity of Shell Fabrication Sushant

More information

Semi-Supervised Face Detection

Semi-Supervised Face Detection Semi-Supervised Face Detection Nicu Sebe, Ira Cohen 2, Thomas S. Huang 3, Theo Gevers Faculty of Science, University of Amsterdam, The Netherlands 2 HP Research Labs, USA 3 Beckman Institute, University

More information

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012 Text-independent Mono and Cross-lingual Speaker Identification with the Constraint of Limited Data Nagaraja B G and H S Jayanna Department of Information Science and Engineering Siddaganga Institute of

More information

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning Hendrik Blockeel and Joaquin Vanschoren Computer Science Dept., K.U.Leuven, Celestijnenlaan 200A, 3001 Leuven, Belgium

More information

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION JOURNAL OF MEDICAL INFORMATICS & TECHNOLOGIES Vol. 11/2007, ISSN 1642-6037 Marek WIŚNIEWSKI *, Wiesława KUNISZYK-JÓŹKOWIAK *, Elżbieta SMOŁKA *, Waldemar SUSZYŃSKI * HMM, recognition, speech, disorders

More information

Cognitive Prior-Knowledge Testing Method for Core Development of Higher Education of Computing in Academia

Cognitive Prior-Knowledge Testing Method for Core Development of Higher Education of Computing in Academia 290 Int'l Conf. Frontiers in Education: CS and CE FECS'15 Cognitive Prior-Knowledge Testing Method for Core Development of Higher Education of Computing in Academia Mohit Satoskar 1 1 Research Associate,

More information

CONSISTENCY OF TRAINING AND THE LEARNING EXPERIENCE

CONSISTENCY OF TRAINING AND THE LEARNING EXPERIENCE CONSISTENCY OF TRAINING AND THE LEARNING EXPERIENCE CONTENTS 3 Introduction 5 The Learner Experience 7 Perceptions of Training Consistency 11 Impact of Consistency on Learners 15 Conclusions 16 Study Demographics

More information

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT PRACTICAL APPLICATIONS OF RANDOM SAMPLING IN ediscovery By Matthew Verga, J.D. INTRODUCTION Anyone who spends ample time working

More information