Semi-Supervised Learning in Diagnosing the Unilateral Loss of Vestibular Functions

Size: px
Start display at page:

Download "Semi-Supervised Learning in Diagnosing the Unilateral Loss of Vestibular Functions"

Transcription

1 Semi-Supervised Learning in Diagnosing the Unilateral Loss of Vestibular Functions Final Report for COMP150-05, 2011 Spring Mengfei Cao Dept. Computer Science, Tufts University 161 College Ave., Medford, MA, U.S. ABSTRACT Although various vestibular tests have been conduct on the subjects in order to diagnose loss of vestibular function, subjects with some specific test results can not be diagnosed. These subjects are refered as in the non-definite group and turn out to be 26.80% in our dataset. This paper presents an approach to recognizing subjects with any possible test results based on a semi-supervised learning strategy. In addition, the redundancy of the various vestibular tests is explored. The approach uses a modified co-training framework; moreover, a special missing values problem among the vestibular tests data, noted as Missing Sets(MS) problem, can be internally solved by the co-training embedding where the whole feature space is split into several sub-test space. The MS problem differs from usual missing values problem in that the missing values tend to distribute intensively along each sub-tests of vestibular tests instead of spreading randomly: either none or all of the test results are missing. According to the nature of the modified cotraining method, the distinguishability of different vestibular tests can be compared and the MS problem can be solved. The work is evaluated on a clinical dataset with the total size of The experiments illustrate that the modified co-training method provide a stable and well-performed learner; also, there exists a great deal of overlapped information among various vestibular tests. Keywords Semi-supervised learning, minority class problem, missing sets problem, vestibular function diagnosis 1. INTRODUCTION Patients who see a specialist in Vestibular or Balance disorders are asked to undergo seven different physical tests to determine whether they have unilateral loss of vestibular Comp-150, Machine Learning in Predictive Medicine, by Prof. Schmid, Prof. Brodley function (balance) or are normal each test produces 7-22 different measures. We are interested in whether machine learning methods can be applied to this data to diagnose patients as normal or abnormal (i.e., unilateral loss). A challenge in this data is that patients often choose not to undergo one or more of the tests seem intimidating, or they may quit half-way through due to excessive dizziness and nausea. Thus patients may be missing one or more sets of features. We call this problem the missing set problem, and in this paper we provide a method for applying machine learning in this scenario. A second challenge of this data is that our data is imbalanced. In the population at patients who visit a Otoneurology department XX% of patients are abnormal. In a dataset collected by Dr. Lewis and associates at Mass Eye and Ear 11% of the cases are abnormal. A difficulty in creating a classifier for a datasets with a skewed class distribution is that sensitivity often suffers and specificity is boosted [3]. Finally, we were provided with both labeled and unlabeled data. What makes this a particularly interesting challenge is that the unlabeled data that is unlabled because the doctors are unsure of the label given the tests. Note that there are other non-vestibular tests that can help them make a differntial diagnosis in these cases, but they are interested in whether these cases can be classified accurately based on just the vestibular data alone. A related a issue of interest is whether any subset of the seven vestibular tests are redundant with respect to prediction performance in this case considerable time and patients discomfort could be ameliorated. In the remainder of this paper we first review the related work in the Otoneurology literature. We next present our approach to the missing set problem, the class imblance problem and we present a method for applying co-training to utilize the unlabeled data and thus we can classify patients that the physicians were unable to. We present results of an empirical evaluation that the co-training strategy that learns from both labeled and unlabeled data can perform at least as well as the one learnt singly on labeled data, but it has the capability to classifiy the unlabeled data ; also the method to deal with missing set problem improves the prediction performance compared to that without splitting. We conclude that the co-trainging semisupervised learning method fits in the vestibular diagnosis situation and the splitting strategy

2 helps solve the missing set problem. 2. CLASSIFYING VESTIBULAR FUNCTION Current, there are not automated methods that can effectively diagnose all patients; specifically, for some patient s test results, past strategies will make conflicting diagnoses. For example, in our data there are 26.80% non-definite patients. However, obtaining information in these cases will help for decisions in medication or surgerical interventions. Previous work on diagnosing vestibular loss use 1) a linearized threshold parameter noted as paresis originated from the electronystagmogram test (ENG test) [?][13] or 2) a linearized threshold parameter noted as time constant originated from the sinusoidal vertical-axis rotation test (SVAR test) [5]. However these two parameters do not cover the whole data space of subjects tests, and further make conflicting predictions on the non-definite group of subjects. (See Figure 1). 3. A SEMI-SUPERVISED LEARNING AP- PROACH Because we are given both labeled and unlabeled data, we can formulate the problem as a semi-supervised learning problem. Namely, given a dataset D and a vector Y = {0, 1} m of labels, where D = n and m < n, each element s D represents a subject s whole vestibular tests values and can be noted as a 97-dimensional vector. Specifically, the values in the vector can be nominal or numerical. In addition, these 97 values of each subject come from the full test battery of seven vestibular tests. As for the label vector Y, we use 0 to indicate normal and 1 abnormal. Without loss of generality, we note that a subject s is normal by Y (s) = 0 and abnormal by Y (s) = 1. All subjects that have a label form a subset L of D and the unlabeled dataset U = D L, where the sizes of L and U are m and q, so we have n = m + q. The purpose of the proposed semisupervised learning algorithm is to take advantage of both L and U, together with Y to train a classifier C that for any subject s in the feature space, C would output a solid label to s. In the vestibular function diagnosis problem, there are two issues to tackle with: the missing set issue and the imbalanced data issue. Next we will elaborate our approaches to tackling this and followed is a semi-supervised framework that embed the whole work. Figure 1: Distribution of the Instances w.r.t. two Parameters Before moving to our approach we first enumerate the seven vestibular tests for which we have data [13]. There seven tests are the standard battery of tests given to patients with balance function issues: 1. POSIT test (positional test) 2. Caloric test 3. Gaze test, 4. Saccade test, 5. SVAR test (Sinusoidal Vertical-Axis Rotation test), 6. VVI test (Visual-vestibular Interaction test) 7. EQUI test( motor control test and sensory orgnization test). 3.1 Missing Set During the full vestibular test battery, a subject can choose not to participate in one or more of the seven vestibular tests. In addition, it happens frequently that during a specific subtest, subjects can quit testing half-time due to dizziness the result is that all of the data for this test are discarded. Thus, we have sets of missing values; either all of the values for a test are present or all of the values are missing. It should be noted that this situation is different from the random missing values. We note this specific missing values problem as the Missing Sets (MS) problem. Classic methods for handling missing data are maximum likelihood estimation and bayesian multiple imputation [11]; these methods are applicable to the data missing at random. However, in our case we have missing sets, thus we explore an alternative method for handling missing sets of values. In particular, we first split the whole feature space into k subspaces, one for each of the k sets of features for both labeled L and unlabeled data U, (for the Vestibular data this results in 2 7 sets). We can then treat the labeled splitted data as k training sets of data. Note that each feature is in one and only one of these k sets; because for some datapoints, one or more subspaces are totally filled with missing values (missing set), thus the sizes of the splitted sets don t have to be equal to the size of the original set. We create k sets of training data as follows: For each instance p in L, we put its q( k) sets of features into k sets; thus we have L 1, L 2,..., L k. In the experimental stage, stratified sampling is applied so as to extract the training data and test data; this sampling routine is applied to k sets, and thus obtain k sets of training data. After we form these k sets, we can then train k classifiers, one for each dataset D i, i = 1, 2,...k.

3 Whenever there comes a new instance, we can combine the prediction of l k classifiers, where l is the number of sets that the instance has the values for, namely in particular, the number of the patients has participated in. The proposed strategy has three advantages: first, it will guarantee that for each learning stage, the learner can take advantage of the current existing data, and avoid the affect of the missing values of other patients; second, it fits perfectly in a semi-supervised learning framework, which combines both the information from labeled subjects and the unlabeled; third, it allows us to consider each subspace separately and thus significant for the problems such as the vestibular project where the relationship between different vestibular tests remains to be explored. Also, this problem has not been discussed before and thus is still an open issue to be optimized. In the Table1), three methods are compared, where the first is to spread the instances with missing values fractional along the branch in the decision tree; the second is to fill the missing values with the avearge values within the same class; the last row represents the proposed approach. The method of filling with average values tends to construct a tree with single node that classifies all the data to be normal and gets extremely low sensitivity ( 0.27% ); compared to the method of spreading the instances with missing values, our approach outperforms in terms of the sensitivity at 95.47%. Although the first method has the highest specificity at 96.80%, however, our approach still keeps specificity 94.92% with nearly 13% gain of the sensitivity. Thus, it might be concluded that our method provides more balanced and still relatively good performance of sensitivity and specificity. 3.2 Class Imbalance The minority class problem is common in terms of the disease statistics. In the clinical vestibular data available, the percentage of the abnormal subjects is as low as 11%. This fact will cause some machine learning techniques to heavily degrade [3]. For instance, when the classifiers are trained using the C4.5 decision tree method on EQUI, GAZE, and POSIT data, all of the decision trees contain only one single node Normal as the root as well as the leaf, and thus will classify all the subjects to be normal. Yet still they keep high accuracies. In order to avoid the effect of the minority class problem and improve the classifiers descriptive capacity, two sampling methods, oversampling and undersampling[9][8][3] are compared with the one without sampling. The oversampling method will over-sample the minority class and results in replications of the minority class data so that during the learning procedure, the minority class data will be weighted more heavily. After duplications, the size of the tree increases but also the noise in the majority class data will be augmented. The undersampling method will sample identical number to minority class of examples from the majority class and form the training data. Thus these two methods will give different results where one biases the sensitivity and the other specificity. According to the results in Figure2, both the sampling techniques bia the sensitivity while degrade in terms of the specificity. Choosing an appropriate threshold for voting among the seven tests, both sample methods can achieve at least one best measurement. In particular, in the case where a false positive is more costy than a false negative, the sampling techniques should be utilized. Since it is hard to quantify a false positive and a false negative, these two together with the method without sampling add up the flexibility for clinicians to choose. Figure 2: Comparison of Different Sampling Methods w.r.t. Different Votes to that Without Splitting 3.3 Co-training Traditional supervised learning strategies only apply L to the learner and thus obtain a classifier that is totally independent from U, which on one hand may not be completely descriptive among the whole feature space, and on the other hand wastes the unlabeled data. What s the most serious shorcoming is that still this classifier can not deal with the non-definite group because no data in this group appears in the training data. Therefore, the semi-supervised framework is more suitable under this situation[2]. Previously the generative models [7] are applied for text classification; the self-training by [14] is a commonly used method for semi-supervised learning; also the method of transductive support vector machine [12] and graph based methods [1] are used in the proposed fields; the method of co-training brought forward by [2] provides a nice intuition for the data that are naturally split to subsets. They work well in the fields such as text classification, image categorization, and object detection; yet the issues such as the local maxima, the high complexity of the models remain to be solved. According to the characteristics of the vestibular data, a method that is able to deal with the minority class issue and the MS problem, and that can easily embed the comparison of distinguishability of different vestibular tests,

4 Table 1: the Comparison Results with std Among Different Methods of Battling Missing Values(MVs) based on J48 Decisiont Tree Algorithm Methods Accuracy Specificity Sensitivity Spread MVs Along Branches 96.80% ± 0.60% 98.61% ± 0.44% 82.53% ± 3.95% Replace Missing Values 88.74% ± 0.09% 99.98% ± 0.05% 0.27% ± 0.56% Proposed Method 94.92% ± 0.74% 94.85% ± 0.91% 95.47% ± 3.34% is required. We proposed an approach with the modified co-training framework, combining both labeled and unlabeled data, embedding C4.5 algorithms [10] [6] in the 7 vestibular tests separately, resulting in a well-performed learner that is applicable to the data with the MS problem the minority issue. In addition, the proposed co-training based method allowed different supervised learning methods embedded to subspaces, and thus more adaptive to any specific data. As declared before, the purpose is to combine both the labeled data L and the unlabeled data U and train a binary classifier that classifies any subject from their vestibular test results. We propose a framework based on co-training. First, we train a set C 1 of classifiers based on the labeled data L 1,..., L k and their label vector Y. Before this stage, the feature space has been split into k subspaces and thus there are k classifiers after this stage. Second, apply these k classifiers to the unlabeled data U 1,..., U k and the labels vectors Y Ui for each U i, where i = 1,..., k, is obtained. Third, get the label vector Y U for U based on the majority votes of Y U1,..., Y Uk, and merge the two label vectors Y and Y U to be Y all. Then, we can train another set C 2 of classifiers based on the data L 1 +U 1,..., L k +U k and their label vector Y all. Finally, organize the classifiers in C 2 and obtain the final classifier C that is constructed over the whole feature space. The way to organize these separate classifiers can simply be voting or more sophisticatedly, by training over the predictions on training data. The three characteristics of this classifier are: first, it is trained from both the labeled and the unlabeled data; secondly, it is applicable to any subjects with any vestibular tests result, instead of partial diagnosis field in the former work that use the two parameters; thirdly, the intermediate result, C 2 can be used to explore the relationship between different vestibular tests. 3.4 Evaluation Methodology The evaluation of the work is based on the comparison of the classifiers. We apply the modified 10-fold stratified cross validation. Namely, after sampling the labeled data into 10 groups, run the learning process for 10 runs; in each run 9 groups serve as the training data and the rest one group is the test data. Nevertheless, it is not easy to tell whether a false positive decision or a false negative decision is more unforgivable, so the better and common way to quantize the performance of a particular diagnosis classifier is to give the accuracy, the sensitivity and the specificity defined as follows. accuracy = #{correctly classified subjects} #{all the subjects} ; Figure 3: the framework of the proposed approach sensitivity = specificity = #{subjects correctly classified as abnormal} #{the abnormal subjects} ; #{subjects correctly classified as normal} #{the normal subjects} ; 3.5 Exploration of the Relationship between Vestibular Tests Based on the work above, the relationship between the seven vestibular tests can be simply explored by processing the separate classifiers originated by these seven tests, namely C 2. In order to capture the redundancy of these vestibular tests, we calculate a redundancy matrix, where each cell of the matrix is the percentage that of the data for which the two particular classifiers make identical decisions. Each classifier is trained on a different vestibular test. Thus in the vestibular project, we finally construct a 7 7 symmetric redundancy matrix, which the values along the diagnal are 1. Although it is true that since all the classifiers are trying to make the correct classification thus ideally each percentage is trivially 100% for each pair of the tests, however, the fact is that they are not perfect in diagnosing all the subjects in the whole feature space. The induced percentage that captures both the number of correct classifications and the number of incorrect classifications, will indicate how much overlapped information the pair of vestibular tests provides. 4. EXPERIMENTS 4.1 Data The experiments are conduct on the clinical data provided by Prof. Lewis. After the preprocessing of the bilateral subjects and other irrelated issues, there are 9102 subjects vestibular tests results, where 6663 of them are labeled as

5 Table 2: the size of the feature vector for each test Test size of features GAZE 20 EQUI 22 SACC 6 SVAR 21 VVI 12 CALORIC 6 POSIT 10 total 97 normal or abnormal and 2439 of them are potentially diagnosed as unilateral loss patients or normal. The 10-fold cross validation is applied and all of the results are presented as the average of the 10 folds experiments. As indicated above, for each subject, their tests results are split into seven parts according to the seven subtest items.(table 2) 4.2 Results & Analyses Comparison of Different Methods to Deal with Missing Values In Table1, three methods are compared. The first classifier is learnt using the default setting of C4.5 decision tree algorithm[10] [4] where the instance with missing values is splitted and spread fractionally along its children in the tree. The second method is the to replace the missing values with the median within the same class. Our method gives the most balanced performance between specificity and sensitivity while the other two both offer a relatively low measurement of sensitivity Comparison of the Classifiers Before and After Semi-supervised Learning Table 3 shows the results of classifiers trained simply on the training data(run1) and the classifiers trained after cotraining(run2). For each of the subtests, the supervised algorithm utilized is the C4.5 from WEKA[10][4]. The overall results come from the voting among the seven tests classifiers with voting threshold 1. The test set consists of the labeled data, which can not be diagnosed with the former strategy. In other words, the test set doesn t include any non-definite data. Thus if the non-definite data during the training stage affect the classifiers negatively, the output classifier will degrade a lot. However, in our results(table 3), the accuracy, specificity and the sensitivity don t change much ( differ within 3% ). What s worth noticing is that the classifers output after the second run are trained on both the original labeled data and the non-definite data; therefore, they can also describe the non-definte data and thus applicable to the non-definite data that can t be dealt with in the former literatures Exploration Table4 gives the redundancy matrix on the unlabeled data. The entry of the matrix is calculated as the percentage the pair of the tests make the identical decisions. It is unknown whether the decisions of the classifiers are correct or not. What s quite surprising is that the classifiers trained on GAZE and EQUI tests share exactly the same decisions on all the unlabeled data. Assuming the experiments are valid and the test data is sufficiently large, we can reach the conclusion that at least one of these two tests is redundant! 5. CONCLUSIONS By using the modified co-training framework, we obtain the classifier whose training is robust to the situations of minority class problem and the MS problem. Also the redundancy of the different vestibular tests can be quantitatively represented and thus provides the clinical scientists with appropriate information about the values and characteristics of these tests. Further work centers on improving the semisupervised framework and how more thoroughly to explore the relationship between the vestibular tests. For the former part, the attempts of organizing the classifications by different tests, such as to stack the trees, and the attempts to use more sophisticated and suitable method to deal with the minority class problem, might be good starts. 6. ACKNOWLEDGMENTS I would like to thank Prof. Schmid, Prof. Lewis and Prof. Brodley for the valuable chance and a lot of advice on the work. I am grateful for Dr. Small and Dr. Navdeep s advice and help! 7. REFERENCES [1] R. W. Baloh and J. M. R. Furman. Modern vestibular function testing. Medical Progress, [2] A. Blum and S. Chawla. Learning from labeled and unlabeled data using graph mincuts. In Proceedings of the Eighteenth International Conference on Machine Learning, ICML 01, pages 19 26, [3] A. Blum and T. Mitchell. Combining labeled and unlabeled data with co-training. In Proceedings of the eleventh annual conference on Computational learning theory, COLT 98, pages , [4] N. V. Chawla. C4.5 and imbalanced data sets: investigating the effect of sampling method, probabilistic estimate, and decision tree structure. In In Proceedings of the ICMLŠ03 Workshop on Class Imbalances, [5] M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I. H. Witten. The weka data mining software: an update. SIGKDD Explor. Newsl., 11, [6] D. Merfeld, R. Lewis, and et al. Potential solutions to several vestibular challenges facing clinicians. Journal of Vestibular Research, [7] T. M. Mitchell. Machine Learning. McGraw-Hill, New York, [8] K. Nigam, A. K. Mccallum, S. Thrun, and T. Mitchell. Text classification from labeled and unlabeled documents using em. In Machine Learning, pages , [9] T. Oates and D. Jensen. The effects of training set size on decision tree complexity. In In Proceedings of The Fourteenth International Conference on Machine Learning, pages Morgan Kaufmann, [10] T. Oates and D. Jensen. Toward a theoretical understanding of why and when decision tree pruning algorithms fail. In In: Proc. 16th National Conference on Arti Intelligence, pages AAAI Press, 1999.

6 Table 3: Comparison of the Classifiers Before and After Semi-supervised Learning GAZE EQUI SACC SVAR VVI CALORIC POSIT Overall Accuracy of Run % 88.72% 88.72% 95.37% 93.28% 96.65% 88.72% 95.23% Accuracy of Run % 88.72% 88.69% 94.66% 93.08% 95.70% 88.72% 93.23% Specificity of Run % % % 97.86% 97.80% 98.90% % 95.19% Specificity of Run % % 99.97% 96.69% 96.75% 98.02% % 92.97% Sensitivity of Run1 0.00% 0.00% 0.00% 75.73% 57.73% 78.93% 0.00% 95.60% Sensitivity of Run2 0.00% 0.00% 0.00% 78.67% 64.27% 77.47% 0.00% 95.33% Table 4: the redundancy matrix on the unlabeled data Test GAZE EQUI SACC SVAR VVI CALORIC POSIT GAZE % % 99.87% 60.36% 67.85% 66.39% % EQUI % % 99.87% 60.36% 67.85% 66.39% % SACC 99.87% 99.87% % 60.49% 67.87% 66.30% 99.87% SVAR 60.36% 60.36% 60.49% % 85.63% 38.81% 60.36% VVI 67.85% 67.85% 67.87% 85.63% % 43.48% 67.85% CALORIC 66.39% 66.39% 66.30% 38.81% 43.48% % 66.39% POSIT % % 99.87% 60.36% 67.85% 66.39% % [11] J. R. Quinlan. C4.5: programs for machine learning. Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, [12] J. L. Schafer and J. W. Graham. Missing data: our view of the state of the art. Psychological Methods, [13] M. Seeger. Learning with labeled and unlabeled data. Technical report, Institute for Adaptive and Neural Computation, University of Edinburgh, [14] C. Wall, R. F. Lewis, and S. D. Rauch. Surgery of the Ear and Temporal Bone, Chapter 5, Evaluation of the Vestibular System. Lippincott Williams Wilkins, [15] D. Yarowsky. Unsupervised word sense disambiguation rivaling supervised methods. In Proceedings of the 33rd Annual Meeting of the Association for Computational Linguistics, pages , 1995.

Rule Learning With Negation: Issues Regarding Effectiveness

Rule Learning With Negation: Issues Regarding Effectiveness Rule Learning With Negation: Issues Regarding Effectiveness S. Chua, F. Coenen, G. Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX Liverpool, United

More information

Rule Learning with Negation: Issues Regarding Effectiveness

Rule Learning with Negation: Issues Regarding Effectiveness Rule Learning with Negation: Issues Regarding Effectiveness Stephanie Chua, Frans Coenen, and Grant Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX

More information

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages Nuanwan Soonthornphisaj 1 and Boonserm Kijsirikul 2 Machine Intelligence and Knowledge Discovery Laboratory Department of Computer

More information

Learning From the Past with Experiment Databases

Learning From the Past with Experiment Databases Learning From the Past with Experiment Databases Joaquin Vanschoren 1, Bernhard Pfahringer 2, and Geoff Holmes 2 1 Computer Science Dept., K.U.Leuven, Leuven, Belgium 2 Computer Science Dept., University

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Todd Holloway Two Lecture Series for B551 November 20 & 27, 2007 Indiana University Outline Introduction Bias and

More information

(Sub)Gradient Descent

(Sub)Gradient Descent (Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include

More information

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System QuickStroke: An Incremental On-line Chinese Handwriting Recognition System Nada P. Matić John C. Platt Λ Tony Wang y Synaptics, Inc. 2381 Bering Drive San Jose, CA 95131, USA Abstract This paper presents

More information

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 98 (2016 ) 368 373 The 6th International Conference on Current and Future Trends of Information and Communication Technologies

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

Assignment 1: Predicting Amazon Review Ratings

Assignment 1: Predicting Amazon Review Ratings Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for

More information

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Twitter Sentiment Classification on Sanders Data using Hybrid Approach IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 4, Ver. I (July Aug. 2015), PP 118-123 www.iosrjournals.org Twitter Sentiment Classification on Sanders

More information

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Good Judgment Project: A large scale test of different methods of combining expert predictions The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania

More information

Semi-Supervised Face Detection

Semi-Supervised Face Detection Semi-Supervised Face Detection Nicu Sebe, Ira Cohen 2, Thomas S. Huang 3, Theo Gevers Faculty of Science, University of Amsterdam, The Netherlands 2 HP Research Labs, USA 3 Beckman Institute, University

More information

A Case Study: News Classification Based on Term Frequency

A Case Study: News Classification Based on Term Frequency A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center

More information

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Notebook for PAN at CLEF 2013 Andrés Alfonso Caurcel Díaz 1 and José María Gómez Hidalgo 2 1 Universidad

More information

Word Segmentation of Off-line Handwritten Documents

Word Segmentation of Off-line Handwritten Documents Word Segmentation of Off-line Handwritten Documents Chen Huang and Sargur N. Srihari {chuang5, srihari}@cedar.buffalo.edu Center of Excellence for Document Analysis and Recognition (CEDAR), Department

More information

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Innov High Educ (2009) 34:93 103 DOI 10.1007/s10755-009-9095-2 Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Phyllis Blumberg Published online: 3 February

More information

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Active Learning. Yingyu Liang Computer Sciences 760 Fall Active Learning Yingyu Liang Computer Sciences 760 Fall 2017 http://pages.cs.wisc.edu/~yliang/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed by Mark Craven,

More information

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS L. Descalço 1, Paula Carvalho 1, J.P. Cruz 1, Paula Oliveira 1, Dina Seabra 2 1 Departamento de Matemática, Universidade de Aveiro (PORTUGAL)

More information

Switchboard Language Model Improvement with Conversational Data from Gigaword

Switchboard Language Model Improvement with Conversational Data from Gigaword Katholieke Universiteit Leuven Faculty of Engineering Master in Artificial Intelligence (MAI) Speech and Language Technology (SLT) Switchboard Language Model Improvement with Conversational Data from Gigaword

More information

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, 2013 10.12753/2066-026X-13-154 DATA MINING SOLUTIONS FOR DETERMINING STUDENT'S PROFILE Adela BÂRA,

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

OCR for Arabic using SIFT Descriptors With Online Failure Prediction OCR for Arabic using SIFT Descriptors With Online Failure Prediction Andrey Stolyarenko, Nachum Dershowitz The Blavatnik School of Computer Science Tel Aviv University Tel Aviv, Israel Email: stloyare@tau.ac.il,

More information

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina

More information

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application International Journal of Medical Science and Clinical Inventions 4(3): 2768-2773, 2017 DOI:10.18535/ijmsci/ v4i3.8 ICV 2015: 52.82 e-issn: 2348-991X, p-issn: 2454-9576 2017, IJMSCI Research Article Comparison

More information

MINUTE TO WIN IT: NAMING THE PRESIDENTS OF THE UNITED STATES

MINUTE TO WIN IT: NAMING THE PRESIDENTS OF THE UNITED STATES MINUTE TO WIN IT: NAMING THE PRESIDENTS OF THE UNITED STATES THE PRESIDENTS OF THE UNITED STATES Project: Focus on the Presidents of the United States Objective: See how many Presidents of the United States

More information

On-Line Data Analytics

On-Line Data Analytics International Journal of Computer Applications in Engineering Sciences [VOL I, ISSUE III, SEPTEMBER 2011] [ISSN: 2231-4946] On-Line Data Analytics Yugandhar Vemulapalli #, Devarapalli Raghu *, Raja Jacob

More information

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Using Web Searches on Important Words to Create Background Sets for LSI Classification Using Web Searches on Important Words to Create Background Sets for LSI Classification Sarah Zelikovitz and Marina Kogan College of Staten Island of CUNY 2800 Victory Blvd Staten Island, NY 11314 Abstract

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1 Notes on The Sciences of the Artificial Adapted from a shorter document written for course 17-652 (Deciding What to Design) 1 Ali Almossawi December 29, 2005 1 Introduction The Sciences of the Artificial

More information

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE Edexcel GCSE Statistics 1389 Paper 1H June 2007 Mark Scheme Edexcel GCSE Statistics 1389 NOTES ON MARKING PRINCIPLES 1 Types of mark M marks: method marks A marks: accuracy marks B marks: unconditional

More information

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT PRACTICAL APPLICATIONS OF RANDOM SAMPLING IN ediscovery By Matthew Verga, J.D. INTRODUCTION Anyone who spends ample time working

More information

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model Xinying Song, Xiaodong He, Jianfeng Gao, Li Deng Microsoft Research, One Microsoft Way, Redmond, WA 98052, U.S.A.

More information

Lecture 1: Basic Concepts of Machine Learning

Lecture 1: Basic Concepts of Machine Learning Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010

More information

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17. Semi-supervised methods of text processing, and an application to medical concept extraction Yacine Jernite Text-as-Data series September 17. 2015 What do we want from text? 1. Extract information 2. Link

More information

Mining Student Evolution Using Associative Classification and Clustering

Mining Student Evolution Using Associative Classification and Clustering Mining Student Evolution Using Associative Classification and Clustering 19 Mining Student Evolution Using Associative Classification and Clustering Kifaya S. Qaddoum, Faculty of Information, Technology

More information

Probability estimates in a scenario tree

Probability estimates in a scenario tree 101 Chapter 11 Probability estimates in a scenario tree An expert is a person who has made all the mistakes that can be made in a very narrow field. Niels Bohr (1885 1962) Scenario trees require many numbers.

More information

Applications of data mining algorithms to analysis of medical data

Applications of data mining algorithms to analysis of medical data Master Thesis Software Engineering Thesis no: MSE-2007:20 August 2007 Applications of data mining algorithms to analysis of medical data Dariusz Matyja School of Engineering Blekinge Institute of Technology

More information

CSL465/603 - Machine Learning

CSL465/603 - Machine Learning CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am

More information

Indian Institute of Technology, Kanpur

Indian Institute of Technology, Kanpur Indian Institute of Technology, Kanpur Course Project - CS671A POS Tagging of Code Mixed Text Ayushman Sisodiya (12188) {ayushmn@iitk.ac.in} Donthu Vamsi Krishna (15111016) {vamsi@iitk.ac.in} Sandeep Kumar

More information

The Strong Minimalist Thesis and Bounded Optimality

The Strong Minimalist Thesis and Bounded Optimality The Strong Minimalist Thesis and Bounded Optimality DRAFT-IN-PROGRESS; SEND COMMENTS TO RICKL@UMICH.EDU Richard L. Lewis Department of Psychology University of Michigan 27 March 2010 1 Purpose of this

More information

CS 446: Machine Learning

CS 446: Machine Learning CS 446: Machine Learning Introduction to LBJava: a Learning Based Programming Language Writing classifiers Christos Christodoulopoulos Parisa Kordjamshidi Motivation 2 Motivation You still have not learnt

More information

Individual Component Checklist L I S T E N I N G. for use with ONE task ENGLISH VERSION

Individual Component Checklist L I S T E N I N G. for use with ONE task ENGLISH VERSION L I S T E N I N G Individual Component Checklist for use with ONE task ENGLISH VERSION INTRODUCTION This checklist has been designed for use as a practical tool for describing ONE TASK in a test of listening.

More information

Linking Task: Identifying authors and book titles in verbose queries

Linking Task: Identifying authors and book titles in verbose queries Linking Task: Identifying authors and book titles in verbose queries Anaïs Ollagnier, Sébastien Fournier, and Patrice Bellot Aix-Marseille University, CNRS, ENSAM, University of Toulon, LSIS UMR 7296,

More information

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models Netpix: A Method of Feature Selection Leading to Accurate Sentiment-Based Classification Models 1 Netpix: A Method of Feature Selection Leading to Accurate Sentiment-Based Classification Models James B.

More information

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za

More information

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE EE-589 Introduction to Neural Assistant Prof. Dr. Turgay IBRIKCI Room # 305 (322) 338 6868 / 139 Wensdays 9:00-12:00 Course Outline The course is divided in two parts: theory and practice. 1. Theory covers

More information

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS 1 CALIFORNIA CONTENT STANDARDS: Chapter 1 ALGEBRA AND WHOLE NUMBERS Algebra and Functions 1.4 Students use algebraic

More information

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study Purdue Data Summit 2017 Communication of Big Data Analytics New SAT Predictive Validity Case Study Paul M. Johnson, Ed.D. Associate Vice President for Enrollment Management, Research & Enrollment Information

More information

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning Hendrik Blockeel and Joaquin Vanschoren Computer Science Dept., K.U.Leuven, Celestijnenlaan 200A, 3001 Leuven, Belgium

More information

Truth Inference in Crowdsourcing: Is the Problem Solved?

Truth Inference in Crowdsourcing: Is the Problem Solved? Truth Inference in Crowdsourcing: Is the Problem Solved? Yudian Zheng, Guoliang Li #, Yuanbing Li #, Caihua Shan, Reynold Cheng # Department of Computer Science, Tsinghua University Department of Computer

More information

A NEW ALGORITHM FOR GENERATION OF DECISION TREES

A NEW ALGORITHM FOR GENERATION OF DECISION TREES TASK QUARTERLY 8 No 2(2004), 1001 1005 A NEW ALGORITHM FOR GENERATION OF DECISION TREES JERZYW.GRZYMAŁA-BUSSE 1,2,ZDZISŁAWS.HIPPE 2, MAKSYMILIANKNAP 2 ANDTERESAMROCZEK 2 1 DepartmentofElectricalEngineeringandComputerScience,

More information

University of Groningen. Systemen, planning, netwerken Bosman, Aart

University of Groningen. Systemen, planning, netwerken Bosman, Aart University of Groningen Systemen, planning, netwerken Bosman, Aart IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document

More information

Using focal point learning to improve human machine tacit coordination

Using focal point learning to improve human machine tacit coordination DOI 10.1007/s10458-010-9126-5 Using focal point learning to improve human machine tacit coordination InonZuckerman SaritKraus Jeffrey S. Rosenschein The Author(s) 2010 Abstract We consider an automated

More information

arxiv: v2 [cs.cv] 30 Mar 2017

arxiv: v2 [cs.cv] 30 Mar 2017 Domain Adaptation for Visual Applications: A Comprehensive Survey Gabriela Csurka arxiv:1702.05374v2 [cs.cv] 30 Mar 2017 Abstract The aim of this paper 1 is to give an overview of domain adaptation and

More information

Learning Methods in Multilingual Speech Recognition

Learning Methods in Multilingual Speech Recognition Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex

More information

Strategy for teaching communication skills in dentistry

Strategy for teaching communication skills in dentistry Strategy for teaching communication in dentistry SADJ July 2010, Vol 65 No 6 p260 - p265 Prof. JG White: Head: Department of Dental Management Sciences, School of Dentistry, University of Pretoria, E-mail:

More information

Axiom 2013 Team Description Paper

Axiom 2013 Team Description Paper Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF Read Online and Download Ebook ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF Click link bellow and free register to download

More information

Reducing Features to Improve Bug Prediction

Reducing Features to Improve Bug Prediction Reducing Features to Improve Bug Prediction Shivkumar Shivaji, E. James Whitehead, Jr., Ram Akella University of California Santa Cruz {shiv,ejw,ram}@soe.ucsc.edu Sunghun Kim Hong Kong University of Science

More information

Lecture 2: Quantifiers and Approximation

Lecture 2: Quantifiers and Approximation Lecture 2: Quantifiers and Approximation Case study: Most vs More than half Jakub Szymanik Outline Number Sense Approximate Number Sense Approximating most Superlative Meaning of most What About Counting?

More information

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA Alta de Waal, Jacobus Venter and Etienne Barnard Abstract Most actionable evidence is identified during the analysis phase of digital forensic investigations.

More information

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks Devendra Singh Chaplot, Eunhee Rhim, and Jihie Kim Samsung Electronics Co., Ltd. Seoul, South Korea {dev.chaplot,eunhee.rhim,jihie.kim}@samsung.com

More information

MYCIN. The MYCIN Task

MYCIN. The MYCIN Task MYCIN Developed at Stanford University in 1972 Regarded as the first true expert system Assists physicians in the treatment of blood infections Many revisions and extensions over the years The MYCIN Task

More information

Developing True/False Test Sheet Generating System with Diagnosing Basic Cognitive Ability

Developing True/False Test Sheet Generating System with Diagnosing Basic Cognitive Ability Developing True/False Test Sheet Generating System with Diagnosing Basic Cognitive Ability Shih-Bin Chen Dept. of Information and Computer Engineering, Chung-Yuan Christian University Chung-Li, Taiwan

More information

SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT

SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT By: Dr. MAHMOUD M. GHANDOUR QATAR UNIVERSITY Improving human resources is the responsibility of the educational system in many societies. The outputs

More information

Discriminative Learning of Beam-Search Heuristics for Planning

Discriminative Learning of Beam-Search Heuristics for Planning Discriminative Learning of Beam-Search Heuristics for Planning Yuehua Xu School of EECS Oregon State University Corvallis,OR 97331 xuyu@eecs.oregonstate.edu Alan Fern School of EECS Oregon State University

More information

Generative models and adversarial training

Generative models and adversarial training Day 4 Lecture 1 Generative models and adversarial training Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University What is a generative model?

More information

Dimensions of Classroom Behavior Measured by Two Systems of Interaction Analysis

Dimensions of Classroom Behavior Measured by Two Systems of Interaction Analysis Dimensions of Classroom Behavior Measured by Two Systems of Interaction Analysis the most important and exciting recent development in the study of teaching has been the appearance of sev eral new instruments

More information

The stages of event extraction

The stages of event extraction The stages of event extraction David Ahn Intelligent Systems Lab Amsterdam University of Amsterdam ahn@science.uva.nl Abstract Event detection and recognition is a complex task consisting of multiple sub-tasks

More information

A survey of multi-view machine learning

A survey of multi-view machine learning Noname manuscript No. (will be inserted by the editor) A survey of multi-view machine learning Shiliang Sun Received: date / Accepted: date Abstract Multi-view learning or learning with multiple distinct

More information

Using dialogue context to improve parsing performance in dialogue systems

Using dialogue context to improve parsing performance in dialogue systems Using dialogue context to improve parsing performance in dialogue systems Ivan Meza-Ruiz and Oliver Lemon School of Informatics, Edinburgh University 2 Buccleuch Place, Edinburgh I.V.Meza-Ruiz@sms.ed.ac.uk,

More information

Model Ensemble for Click Prediction in Bing Search Ads

Model Ensemble for Click Prediction in Bing Search Ads Model Ensemble for Click Prediction in Bing Search Ads Xiaoliang Ling Microsoft Bing xiaoling@microsoft.com Hucheng Zhou Microsoft Research huzho@microsoft.com Weiwei Deng Microsoft Bing dedeng@microsoft.com

More information

A heuristic framework for pivot-based bilingual dictionary induction

A heuristic framework for pivot-based bilingual dictionary induction 2013 International Conference on Culture and Computing A heuristic framework for pivot-based bilingual dictionary induction Mairidan Wushouer, Toru Ishida, Donghui Lin Department of Social Informatics,

More information

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach To cite this

More information

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance Cristina Conati, Kurt VanLehn Intelligent Systems Program University of Pittsburgh Pittsburgh, PA,

More information

Beyond the Pipeline: Discrete Optimization in NLP

Beyond the Pipeline: Discrete Optimization in NLP Beyond the Pipeline: Discrete Optimization in NLP Tomasz Marciniak and Michael Strube EML Research ggmbh Schloss-Wolfsbrunnenweg 33 69118 Heidelberg, Germany http://www.eml-research.de/nlp Abstract We

More information

Clouds = Heavy Sidewalk = Wet. davinci V2.1 alpha3

Clouds = Heavy Sidewalk = Wet. davinci V2.1 alpha3 Identifying and Handling Structural Incompleteness for Validation of Probabilistic Knowledge-Bases Eugene Santos Jr. Dept. of Comp. Sci. & Eng. University of Connecticut Storrs, CT 06269-3155 eugene@cse.uconn.edu

More information

Australian Journal of Basic and Applied Sciences

Australian Journal of Basic and Applied Sciences AENSI Journals Australian Journal of Basic and Applied Sciences ISSN:1991-8178 Journal home page: www.ajbasweb.com Feature Selection Technique Using Principal Component Analysis For Improving Fuzzy C-Mean

More information

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview Algebra 1, Quarter 3, Unit 3.1 Line of Best Fit Overview Number of instructional days 6 (1 day assessment) (1 day = 45 minutes) Content to be learned Analyze scatter plots and construct the line of best

More information

Attributed Social Network Embedding

Attributed Social Network Embedding JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, MAY 2017 1 Attributed Social Network Embedding arxiv:1705.04969v1 [cs.si] 14 May 2017 Lizi Liao, Xiangnan He, Hanwang Zhang, and Tat-Seng Chua Abstract Embedding

More information

Combining Proactive and Reactive Predictions for Data Streams

Combining Proactive and Reactive Predictions for Data Streams Combining Proactive and Reactive Predictions for Data Streams Ying Yang School of Computer Science and Software Engineering, Monash University Melbourne, VIC 38, Australia yyang@csse.monash.edu.au Xindong

More information

Probability and Statistics Curriculum Pacing Guide

Probability and Statistics Curriculum Pacing Guide Unit 1 Terms PS.SPMJ.3 PS.SPMJ.5 Plan and conduct a survey to answer a statistical question. Recognize how the plan addresses sampling technique, randomization, measurement of experimental error and methods

More information

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Neural Network GUI Tested on Text-To-Phoneme Mapping A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis

More information

Create Quiz Questions

Create Quiz Questions You can create quiz questions within Moodle. Questions are created from the Question bank screen. You will also be able to categorize questions and add them to the quiz body. You can crate multiple-choice,

More information

Speech Emotion Recognition Using Support Vector Machine

Speech Emotion Recognition Using Support Vector Machine Speech Emotion Recognition Using Support Vector Machine Yixiong Pan, Peipei Shen and Liping Shen Department of Computer Technology Shanghai JiaoTong University, Shanghai, China panyixiong@sjtu.edu.cn,

More information

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming Data Mining VI 205 Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming C. Romero, S. Ventura, C. Hervás & P. González Universidad de Córdoba, Campus Universitario de

More information

Knowledge Transfer in Deep Convolutional Neural Nets

Knowledge Transfer in Deep Convolutional Neural Nets Knowledge Transfer in Deep Convolutional Neural Nets Steven Gutstein, Olac Fuentes and Eric Freudenthal Computer Science Department University of Texas at El Paso El Paso, Texas, 79968, U.S.A. Abstract

More information

Seminar - Organic Computing

Seminar - Organic Computing Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts

More information

Why Did My Detector Do That?!

Why Did My Detector Do That?! Why Did My Detector Do That?! Predicting Keystroke-Dynamics Error Rates Kevin Killourhy and Roy Maxion Dependable Systems Laboratory Computer Science Department Carnegie Mellon University 5000 Forbes Ave,

More information

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education GCSE Mathematics B (Linear) Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education Mark Scheme for November 2014 Oxford Cambridge and RSA Examinations OCR (Oxford Cambridge

More information

Exposé for a Master s Thesis

Exposé for a Master s Thesis Exposé for a Master s Thesis Stefan Selent January 21, 2017 Working Title: TF Relation Mining: An Active Learning Approach Introduction The amount of scientific literature is ever increasing. Especially

More information

Calibration of Confidence Measures in Speech Recognition

Calibration of Confidence Measures in Speech Recognition Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE

More information

On the Combined Behavior of Autonomous Resource Management Agents

On the Combined Behavior of Autonomous Resource Management Agents On the Combined Behavior of Autonomous Resource Management Agents Siri Fagernes 1 and Alva L. Couch 2 1 Faculty of Engineering Oslo University College Oslo, Norway siri.fagernes@iu.hio.no 2 Computer Science

More information