# Admission Prediction System Using Machine Learning

Save this PDF as:

Size: px
Start display at page:

## Transcription

2 the integer format, a similar issue with dates. They were in different formats which had to be handled while preprocessing Inconsistent Data Containing discrepancies (a lack of compatibility or similarity between two or more facts). Frequency of this kind of data was very high in almost all the fields where one fact was represented in multiple ways using abbreviation, code names, symbols etc. For example, university name: University of Texas, Dallas was represented in other ways like University of Texas at Dallas, UTD, UT Dallas, etc. Computer Science was represented as CS, Comp Sci, Computer Sci, CSc etc Data Quality Certain fields lack attribute values, certain attributes of interest, and contain only aggregate data. Some of the field values of the decision-making parameters were missing. Because of this some of the data had to be added. Another issue was that of aggregate data. Like the 3-different decision parameters Quantitative, Verbal and AWA (Analytical Writing Analysis) was represented as one entity under the GRE tag. Hence this composite field had to be segregated to get 3 different parameters Performance Performance (Deteriorate without pre-processing) containing errors and outliers. Since the data was inaccurate, it was not possible to achieve the expected accuracy without removing errors and outliers. This was one of the major aspects to consider to obtain efficient results Data Skewness Skewness is a measure of symmetry, or more precisely, the lack of symmetry. A distribution, or data set, is symmetric if it looks the same to the left and right of the center point. [1] Karl Pearson coefficient of Skewness Sk = 3(mean - median) / Standard Deviation. = 3(X Me) / S 2. The skewness of a random variable X is denoted or skew(x). It is defined as: where and are the mean and standard deviation of X. [1] Skewness shows the inclination of the whole data set with respect to the normal distribution. In this dataset, majority of the data was of Accepted Results for a given University. Due to this the distribution was balanced to equalize the Accept and Reject fields. 3. Data Preprocessing Following is the flowchart of the whole process. Figure 1: Data preprocessing steps Data cleaning is performed on raw data by performing type checking and normalization. Above Data Issues are handled step by step to make sure data is consistent and compatible with the Machine Learning Algorithm Noisy Data is handled by filtering out the unstructured text followed by changing all the values of those in proper format. Unformatted Text: Deciding the proper format of all the fields and changing all the unformatted values into an appropriate format. Inconsistent Data: If some data was found to be erroneous, all other values in the respective column were considered to evaluate the mean, which was then entered in place of the erroneous data. Quality data: This was done by segregating GRE field into the 3 sub-category parameters: Quantitative, Verbal and AWA (Analytical Writing Analysis), since all these 3 sub fields are independently considered in a set of decision making parameters. Technical Fixes: This involves handling outliers and error data and is performed solely improve the accuracy of the model. Following outliers were removed. Those records in which students who got less grades and those test results were accepted by the university. Those records in which students who got high grades and those test results were rejected. Such kind of data create ambiguity in analysis and result. Data Skewness has been handled by adding appropriate number of reject columns and balance it with the accepted records to get proper distribution of both accept and reject records.

3 After performing all these processes on the data, the dataset is finally consistent. This dataset can be used to perform the required experiments. This is followed by various tabulation and plotting schemes which can be used to obtain proper formatted information. The University Dataset for determining decisions consists of 1686 rows with 18 columns. Student Dataset for determining student probability to get admits consists of 10 datasets each containing 50 to 200 records of data. Result, GRE, AWA, TOEFL and Percentage are the columns, based on which the Student Selection model is designed. There are 3 methods to handle missing data: Listwise Deletion: Delete all data from any participant with missing values. If your sample is large enough, then you likely can drop data without substantial loss of statistical power. Be sure that the values are missing at random and that you are not inadvertently removing a class of participants. [2] Since our dataset was not large enough and the missing values consists of decision making parameters, deletion method was not an option. Recover the Values: You can sometimes contact the participants and ask them to fill out the missing values. For in-person studies, we ve found having an additional check for missing values before the participant leaves helps. [2] This method was practically not possible as the dataset did not have any references or ways to contact those participants. Educated Guessing: It sounds arbitrary and isn t your preferred course of action, but you can often infer a missing value. [2] This is something which can help fix the missing value problem. But rather than going for an arbitrary guess, we chose the mean as the substitution method for missing values. This ensured that the guessed values are not outliers but, fit well within the domain. all categorical values must be converted into proper numeric form. e.g. Results. Feature Scaling was done on all the columns except the Results field as it only contains Accept or Reject values. Normalization was performed on required fields so that various columns could be compared at the same base. Following are some of the Original Dataset Representation which can help to understand the nature of it rather than going through the whole excel sheet data which is time consuming Total Figure 2: Distribution of major The above graph is the representation of various majors and their distribution frequency. On X-axis is the major and on Y-axis is its corresponding number of records. Here majority of the data records are of Computer Science and hence that is taken into consideration for both the models. Due to limited amount of data available for other majors, it is very difficult to maintain good accuracy. Others contain different majors other than the once mentioned here. CE CS EE Others SE (blank) We ignored the record where percentage was not present. Here Listwise Deletion method is used. The number of missing values in percentage were very few compared to the total number of records in the whole dataset. This method is comparatively feasible and appropriate. Changing categorical data to numeric value. All operations and functions were done on numeric value, so

4 Model Development 4.1 Preliminaries: Accept Reject (blank) Machine learning classification technique is a supervised learning that is designed to infer class labels from a welllabeled trained set having input features associated with the class labels. [3] After cleaning the data as mentioned in the prior section, our two models can be designed as Total University Selection Model Classification problem with apriori probability output. Student Selection Model Classification using supervised learning. Figure 3: After pre-processing distribution of result Since there was an imbalance in the number of Accept and Reject records, it was modified and new data was added to balance this issue. After doing this, since the model required more information of the Reject fields than the Accept fields, dataset was modified again by increasing the number of Reject fields by around 100 more than that of Accept. For the University Selection Model, we use the Naïve Bayes classifier. Naive Bayes classifiers are a family of simple probabilistic classifiers based on applying Bayes' theorem with strong (naive) independence assumptions between the features. [4] Naive Bayes has been a popular (baseline) method for text categorization, i.e. the problem of judging documents, belonging to one category or the other (such as spam or legitimate, sports or politics, etc.) with word frequencies as the features. With appropriate preprocessing, Naive Bayes classifiers are highly scalable, requiring a number of parameters linear in the number of variables (features/predictors) in a learning problem. [4] Maximum-likelihood training can be done by evaluating a closed-form expression like posterior = prior likelihood evidence Equation 1: Naive Bayes [4] Which takes linear time, rather than by expensive iterative approximation as used for many other types of classifiers such as decision trees and SVM. Figure 4: Frequency distribution of University Frequency distribution of the dataset grouped by Individual University is shown above. On X-axis, is the list of different universities. On Y-axis, is the number of records available for each university. Here the data of University of Texas, Dallas has the highest number of records. As number of records for other university is very less, we are limiting the scope of the Student Selection model to this university. The same process can be done on other universities to obtain similar results. In Student selection model, 10 datasets of specific universities were created to obtain the probability of a student against each of these universities. For the second model, Student Selection we worked with a variety of models, namely Naïve Bayes, SVM, Decision Tree and Random Forest. Support vector machines are supervised learning models with associated learning algorithms that analyze data used for classification and regression analysis. SVM training algorithm builds a model that assigns new examples to one category or the other, making it a nonprobabilistic binary linear classifier. An SVM model is a representation of the examples as points in space, mapped so that the examples of the separate categories are divided by a clear gap that is as wide as possible. New examples are then mapped into that same space and predicted to belong to a category based on which side of

5 classifier, as mentioned before estimates the classification on the basis of the probability. This fits right into our requirement. We started the pre-processing by extracting top 10 (in terms of a number of records) university data from the original dataset D into 10 separate datasets. Each dataset di is used to train a model Mi. Figure 5: University Selection System The flow chart above gives a basic idea regarding the functioning of the system. the gap they fall. In addition to performing linear classification, SVMs can efficiently perform a non-linear classification using what is called the kernel trick, implicitly mapping their inputs into high-dimensional feature spaces. [5] The decision tree algorithm is a machine learning classification mechanism, where patterns of input features are analyzed to create a predictive model. A decision tree consists of non-leaf nodes representing tests of features, branches between nodes representing the outcomes of the tests, and leaf nodes holding the class labels. [3] Constructing the most optimal and accurate decision tree is usually NP-hard on a given training set [3]. To construct a decision tree model, most of the practical algorithms use a greedy approach using heuristics such as information gain. Using these algorithms, the training data is recursively partitioned into smaller subsets. When partitioning the dataset, the feature with the highest splitting criterion such as information gain is chosen as the splitting feature. This feature minimizes the information needed to classify the data in the resulting partitions and reflects the least randomness in these partitions. The Random forests method consists of multiple decision trees that are constructed by randomly chosen features with a predefined number of features. The random features classify a label by voting, a plurality decision from individual decision trees. Because of the law of large numbers, the Random forests method is less prone to generalization error (overfit) as randomness are added with more trees. In addition, the generalization error converges to a limited value. It is due to this property of Random Forests, we achieved the accuracy of 90% After all the models are generated, any new students information is evaluated against all the models and their corresponding prediction for acceptance Pi is collected into a pool of predictions. This pool is then sorted in descending order to provide the top 5 probable universities. Given below is one such example. Table 1: Probability pool University Probability MTU_pred clemson_pred NE_Boston_pred ASU_pred IITchicago_pred RIT_pred UTD_pred UTA_pred UNC_pred U_southern_cal_pred Table 2: Sample student data GRE AWA TOEFL IELTS Percentage N/A 85 As per the output, the student in Table 2 has the highest probability of getting into Michigan Technological University with the probability of Followed by Clemson University with 0.90 probability. Using this output the student can decide which universities to apply for. 4.2 University Selection Since the main aim of the model is to find the probability of admission of a student given his scores and other attributes, we choose Naïve Bayes Classifier. This

6 N M R A Figure 8: Student selection system N: Represents the new applicants applying to the university Figure 6: Unsorted Probability Output M: Different Models as mentioned R: Class Reject A: Class Accept We saw the highest error rate in SVM Kernel of 13.06%. Followed by SVM Linear of 12.56%. Naïve Bayes produced an error rate of 12.06%. We received the best results in Decision Tree and Random Forests. Given below is the decision tree modeled as per our dataset. Figure 7: Sorted probability output 4.3 Student Selection The main aim of this system is to classify new applications based on previous years data of the students who got admits or rejects in a particular university. Due to the constraints of the data size, we choose to build this system for University of Texas, Dallas. Given below are the steps for developing this system: Figure 9: Decision tree Past Years Data Pre-Processing Techniques Machine Learning Models Predictions Here 1 represents Accept and 0 represents Reject. After pre-processing the data as mentioned in the earlier section, we train different supervised classification models to classify applications into Accept or Reject. The different models used are Naïve Bayes, SVM Linear & Kernel, Decision Tree, and Random Forest. Each of these models was used to test the set of new applicants along with result to derive the accuracy. As per the tree, the most important criteria checked is Analytical Writing Assessment from GRE. If any student scores above a certain threshold, he/she is accepted. If not, the GRE is checked and the candidate is accepted if he/she scores above a set threshold. The third important criterion is Undergrad Percentage. Any candidate with a percentage higher than 89% is accepted. The least weighted criterion is TOEFL, according to which if any student is below the median then they ll be rejected. We are able to achieve an accuracy of 89% using Decision Tree.

8 References [1] "Skewness," [Online]. Available: ess%20and%20kurtosis.pdf. [Accessed ]. [2] J. Sauro, "MeasuringU: 7 Ways to Handle Missing Data," MeasuringU, [Online]. Available: [Accessed ]. [3] J. R. Quinlan, "Induction of Decision Trees," Mach Learn, [4] Wikipedia, "Naive Bayes Classifier," Wikipedia, [Online]. Available: ier. [Accessed ]. [5] Wikipedia, "Support Vector Machine," Wikipedia, [Online]. Available: hine. [Accessed ]. [6] L. B. a. A. Cutler, "Random forests - classification description," Salford Systems, [Online]. Available: orests/cc_home.htm#ooberr. [Accessed ]. [7] P. C. a. A. Silva, "USING DATA MINING TO PREDICT SECONDARY SCHOOL STUDENT PERFORMANCE". [8] H. W. a. Z. Y. Rensong Dong, "The module of prediction of College Entrance Examination aspiration". [9] D. T. E. S. L. R. a. A. P. William Eberle, "Using Machine Learning and Predictive Modeling to Assess Admission Policies and Standards". [10] H. M. Havan Agrawal, "Student Performance Prediction using Machine Learning," IEEE.

### A Few Useful Things to Know about Machine Learning. Pedro Domingos Department of Computer Science and Engineering University of Washington" 2012"

A Few Useful Things to Know about Machine Learning Pedro Domingos Department of Computer Science and Engineering University of Washington 2012 A Few Useful Things to Know about Machine Learning Machine

### Cross-Domain Video Concept Detection Using Adaptive SVMs

Cross-Domain Video Concept Detection Using Adaptive SVMs AUTHORS: JUN YANG, RONG YAN, ALEXANDER G. HAUPTMANN PRESENTATION: JESSE DAVIS CS 3710 VISUAL RECOGNITION Problem-Idea-Challenges Address accuracy

### Decision Tree for Playing Tennis

Decision Tree Decision Tree for Playing Tennis (outlook=sunny, wind=strong, humidity=normal,? ) DT for prediction C-section risks Characteristics of Decision Trees Decision trees have many appealing properties

### Analysis of Different Classifiers for Medical Dataset using Various Measures

Analysis of Different for Medical Dataset using Various Measures Payal Dhakate ME Student, Pune, India. K. Rajeswari Associate Professor Pune,India Deepa Abin Assistant Professor, Pune, India ABSTRACT

### Analytical Study of Some Selected Classification Algorithms in WEKA Using Real Crime Data

Analytical Study of Some Selected Classification Algorithms in WEKA Using Real Crime Data Obuandike Georgina N. Department of Mathematical Sciences and IT Federal University Dutsinma Katsina state, Nigeria

### Adaptive Testing Without IRT in the Presence of Multidimensionality

RESEARCH REPORT April 2002 RR-02-09 Adaptive Testing Without IRT in the Presence of Multidimensionality Duanli Yan Charles Lewis Martha Stocking Statistics & Research Division Princeton, NJ 08541 Adaptive

### Session 1: Gesture Recognition & Machine Learning Fundamentals

IAP Gesture Recognition Workshop Session 1: Gesture Recognition & Machine Learning Fundamentals Nicholas Gillian Responsive Environments, MIT Media Lab Tuesday 8th January, 2013 My Research My Research

### Introduction to Classification, aka Machine Learning

Introduction to Classification, aka Machine Learning Classification: Definition Given a collection of examples (training set ) Each example is represented by a set of features, sometimes called attributes

### Negative News No More: Classifying News Article Headlines

Negative News No More: Classifying News Article Headlines Karianne Bergen and Leilani Gilpin kbergen@stanford.edu lgilpin@stanford.edu December 14, 2012 1 Introduction The goal of this project is to develop

### TOWARDS DATA-DRIVEN AUTONOMICS IN DATA CENTERS

TOWARDS DATA-DRIVEN AUTONOMICS IN DATA CENTERS ALINA SIRBU, OZALP BABAOGLU SUMMARIZED BY ARDA GUMUSALAN MOTIVATION 2 MOTIVATION Human-interaction-dependent data centers are not sustainable for future data

### Performance Analysis of Various Data Mining Techniques on Banknote Authentication

International Journal of Engineering Science Invention ISSN (Online): 2319 6734, ISSN (Print): 2319 6726 Volume 5 Issue 2 February 2016 PP.62-71 Performance Analysis of Various Data Mining Techniques on

### 18 LEARNING FROM EXAMPLES

18 LEARNING FROM EXAMPLES An intelligent agent may have to learn, for instance, the following components: A direct mapping from conditions on the current state to actions A means to infer relevant properties

### Assignment 1: Predicting Amazon Review Ratings

Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for

### Bird Species Identification from an Image

Bird Species Identification from an Image Aditya Bhandari, 1 Ameya Joshi, 2 Rohit Patki 3 1 Department of Computer Science, Stanford University 2 Department of Electrical Engineering, Stanford University

### Modelling Student Knowledge as a Latent Variable in Intelligent Tutoring Systems: A Comparison of Multiple Approaches

Modelling Student Knowledge as a Latent Variable in Intelligent Tutoring Systems: A Comparison of Multiple Approaches Qandeel Tariq, Alex Kolchinski, Richard Davis December 6, 206 Introduction This paper

### Evaluation and Comparison of Performance of different Classifiers

Evaluation and Comparison of Performance of different Classifiers Bhavana Kumari 1, Vishal Shrivastava 2 ACE&IT, Jaipur Abstract:- Many companies like insurance, credit card, bank, retail industry require

### Supervised learning can be done by choosing the hypothesis that is most probable given the data: = arg max ) = arg max

The learning problem is called realizable if the hypothesis space contains the true function; otherwise it is unrealizable On the other hand, in the name of better generalization ability it may be sensible

### Introduction to Classification

Introduction to Classification Classification: Definition Given a collection of examples (training set ) Each example is represented by a set of features, sometimes called attributes Each example is to

### Lecture 1: Machine Learning Basics

1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

### CS545 Machine Learning

Machine learning and related fields CS545 Machine Learning Course Introduction Machine learning: the construction and study of systems that learn from data. Pattern recognition: the same field, different

### Stay Alert!: Creating a Classifier to Predict Driver Alertness in Real-time

Stay Alert!: Creating a Classifier to Predict Driver Alertness in Real-time Aditya Sarkar, Julien Kawawa-Beaudan, Quentin Perrot Friday, December 11, 2014 1 Problem Definition Driving while drowsy inevitably

### P(A, B) = P(A B) = P(A) + P(B) - P(A B)

AND Probability P(A, B) = P(A B) = P(A) + P(B) - P(A B) P(A B) = P(A) + P(B) - P(A B) Area = Probability of Event AND Probability P(A, B) = P(A B) = P(A) + P(B) - P(A B) If, and only if, A and B are independent,

### Dudon Wai Georgia Institute of Technology CS 7641: Machine Learning Atlanta, GA

Adult Income and Letter Recognition - Supervised Learning Report An objective look at classifier performance for predicting adult income and Letter Recognition Dudon Wai Georgia Institute of Technology

### Cost-Sensitive Learning and the Class Imbalance Problem

To appear in Encyclopedia of Machine Learning. C. Sammut (Ed.). Springer. 2008 Cost-Sensitive Learning and the Class Imbalance Problem Charles X. Ling, Victor S. Sheng The University of Western Ontario,

### Unsupervised Learning

17s1: COMP9417 Machine Learning and Data Mining Unsupervised Learning May 2, 2017 Acknowledgement: Material derived from slides for the book Machine Learning, Tom M. Mitchell, McGraw-Hill, 1997 http://www-2.cs.cmu.edu/~tom/mlbook.html

### Classification of News Articles Using Named Entities with Named Entity Recognition by Neural Network

Classification of News Articles Using Named Entities with Named Entity Recognition by Neural Network Nick Latourette and Hugh Cunningham 1. Introduction Our paper investigates the use of named entities

### Linear Models Continued: Perceptron & Logistic Regression

Linear Models Continued: Perceptron & Logistic Regression CMSC 723 / LING 723 / INST 725 Marine Carpuat Slides credit: Graham Neubig, Jacob Eisenstein Linear Models for Classification Feature function

### 36-350: Data Mining. Fall Lectures: Monday, Wednesday and Friday, 10:30 11:20, Porter Hall 226B

36-350: Data Mining Fall 2009 Instructor: Cosma Shalizi, Statistics Dept., Baker Hall 229C, cshalizi@stat.cmu.edu Teaching Assistant: Joseph Richards, jwrichar@stat.cmu.edu Lectures: Monday, Wednesday

### Concession Curve Analysis for Inspire Negotiations

Concession Curve Analysis for Inspire Negotiations Vivi Nastase SITE University of Ottawa, Ottawa, ON vnastase@site.uottawa.ca Gregory Kersten John Molson School of Business Concordia University, Montreal,

### Classification with Deep Belief Networks. HussamHebbo Jae Won Kim

Classification with Deep Belief Networks HussamHebbo Jae Won Kim Table of Contents Introduction... 3 Neural Networks... 3 Perceptron... 3 Backpropagation... 4 Deep Belief Networks (RBM, Sigmoid Belief

### Learning dispatching rules via an association rule mining approach. Dongwook Kim. A thesis submitted to the graduate faculty

Learning dispatching rules via an association rule mining approach by Dongwook Kim A thesis submitted to the graduate faculty in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE

### CS Machine Learning

CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

### White Paper. Using Sentiment Analysis for Gaining Actionable Insights

corevalue.net info@corevalue.net White Paper Using Sentiment Analysis for Gaining Actionable Insights Sentiment analysis is a growing business trend that allows companies to better understand their brand,

### Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Todd Holloway Two Lecture Series for B551 November 20 & 27, 2007 Indiana University Outline Introduction Bias and

### Note that although this feature is not available in IRTPRO 2.1 or IRTPRO 3, it has been implemented in IRTPRO 4.

TABLE OF CONTENTS 1 Fixed theta estimation... 2 2 Posterior weights... 2 3 Drift analysis... 2 4 Equivalent groups equating... 3 5 Nonequivalent groups equating... 3 6 Vertical equating... 4 7 Group-wise

### A COMPARATIVE ANALYSIS OF META AND TREE CLASSIFICATION ALGORITHMS USING WEKA

A COMPARATIVE ANALYSIS OF META AND TREE CLASSIFICATION ALGORITHMS USING WEKA T.Sathya Devi 1, Dr.K.Meenakshi Sundaram 2, (Sathya.kgm24@gmail.com 1, lecturekms@yahoo.com 2 ) 1 (M.Phil Scholar, Department

### UNIT 1: COLLECTING DATA

Core provides a curriculum focused on understanding key data analysis and probabilistic concepts, calculations, and relevance to real-world applications. Through a "Discovery-Confirmation-Practice"-based

### Scheduling Tasks under Constraints CS229 Final Project

Scheduling Tasks under Constraints CS229 Final Project Mike Yu myu3@stanford.edu Dennis Xu dennisx@stanford.edu Kevin Moody kmoody@stanford.edu Abstract The project is based on the principle of unconventional

### CPSC 340: Machine Learning and Data Mining. Course Review/Preview Fall 2015

CPSC 340: Machine Learning and Data Mining Course Review/Preview Fall 2015 Admin Assignment 6 due now. We will have office hours as usual next week. Final exam details: December 15: 8:30-11 (WESB 100).

### WEKA tutorial exercises

WEKA tutorial exercises These tutorial exercises introduce WEKA and ask you to try out several machine learning, visualization, and preprocessing methods using a wide variety of datasets: Learners: decision

### Machine Learning in Patent Analytics:: Binary Classification for Prioritizing Search Results

Machine Learning in Patent Analytics:: Binary Classification for Prioritizing Search Results Anthony Trippe Managing Director, Patinformatics, LLC Patent Information Fair & Conference November 10, 2017

### Classifying Breast Cancer By Using Decision Tree Algorithms

Classifying Breast Cancer By Using Decision Tree Algorithms Nusaibah AL-SALIHY, Turgay IBRIKCI (Presenter) Cukurova University, TURKEY What Is A Decision Tree? Why A Decision Tree? Why Decision TreeClassification?

### Word Sense Disambiguation with Semi-Supervised Learning

Word Sense Disambiguation with Semi-Supervised Learning Thanh Phong Pham 1 and Hwee Tou Ng 1,2 and Wee Sun Lee 1,2 1 Department of Computer Science 2 Singapore-MIT Alliance National University of Singapore

### Advanced Probabilistic Binary Decision Tree Using SVM for large class problem

Advanced Probabilistic Binary Decision Tree Using for large class problem Anita Meshram 1 Roopam Gupta 2 and Sanjeev Sharma 3 1 School of Information Technology, UTD, RGPV, Bhopal, M.P., India. 2 Information

### CSC 4510/9010: Applied Machine Learning Rule Inference

CSC 4510/9010: Applied Machine Learning Rule Inference Dr. Paula Matuszek Paula.Matuszek@villanova.edu Paula.Matuszek@gmail.com (610) 647-9789 CSC 4510.9010 Spring 2015. Paula Matuszek 1 Red Tape Going

### Assignment 6 (Sol.) Introduction to Machine Learning Prof. B. Ravindran

Assignment 6 (Sol.) Introduction to Machine Learning Prof. B. Ravindran 1. Assume that you are given a data set and a neural network model trained on the data set. You are asked to build a decision tree

### Python Machine Learning

Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

### A study of the NIPS feature selection challenge

A study of the NIPS feature selection challenge Nicholas Johnson November 29, 2009 Abstract The 2003 Nips Feature extraction challenge was dominated by Bayesian approaches developed by the team of Radford

### A Procedure for Classifying New Respondents into Existing Segments Using Maximum Difference Scaling

A Procedure for Classifying New Respondents into Existing Segments Using Maximum Difference Scaling Background Bryan Orme and Rich Johnson, Sawtooth Software March, 2009 (with minor clarifications September

### Probability and Statistics Curriculum Pacing Guide

Unit 1 Terms PS.SPMJ.3 PS.SPMJ.5 Plan and conduct a survey to answer a statistical question. Recognize how the plan addresses sampling technique, randomization, measurement of experimental error and methods

### Predicting Academic Success from Student Enrolment Data using Decision Tree Technique

Predicting Academic Success from Student Enrolment Data using Decision Tree Technique M Narayana Swamy Department of Computer Applications, Presidency College Bangalore,India M. Hanumanthappa Department

### I400 Health Informatics Data Mining Instructions (KP Project)

I400 Health Informatics Data Mining Instructions (KP Project) Casey Bennett Spring 2014 Indiana University 1) Import: First, we need to import the data into Knime. add CSV Reader Node (under IO>>Read)

### Improving Machine Learning Through Oracle Learning

Brigham Young University BYU ScholarsArchive All Theses and Dissertations 2007-03-12 Improving Machine Learning Through Oracle Learning Joshua Ephraim Menke Brigham Young University - Provo Follow this

### Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina

### Machine Learning for NLP

Natural Language Processing SoSe 2014 Machine Learning for NLP Dr. Mariana Neves April 30th, 2014 (based on the slides of Dr. Saeedeh Momtazi) Introduction Field of study that gives computers the ability

### Big Data Classification using Evolutionary Techniques: A Survey

Big Data Classification using Evolutionary Techniques: A Survey Neha Khan nehakhan.sami@gmail.com Mohd Shahid Husain mshahidhusain@ieee.org Mohd Rizwan Beg rizwanbeg@gmail.com Abstract Data over the internet

### Classification of Arrhythmia Using Machine Learning Techniques

Classification of Arrhythmia Using Machine Learning Techniques THARA SOMAN PATRICK O. BOBBIE School of Computing and Software Engineering Southern Polytechnic State University (SPSU) 1 S. Marietta Parkway,

### Feedback Prediction for Blogs

Feedback Prediction for Blogs Krisztian Buza Budapest University of Technology and Economics Department of Computer Science and Information Theory buza@cs.bme.hu Abstract. The last decade lead to an unbelievable

### EMPIRICAL ANALYSIS OF CLASSIFIERS AND FEATURE SELECTION TECHNIQUES ON MOBILE PHONE DATA ACTIVITIES

EMPIRICAL ANALYSIS OF CLASSIFIERS AND FEATURE SELECTION TECHNIQUES ON MOBILE PHONE DATA ACTIVITIES Fandi Husen Harmaini and M. Mahmuddin School of Computing, Universiti Utara Malaysia, Sintok Kedah, Malaysia

### 2: Exploratory data Analysis using SPSS

: Exploratory data Analysis using SPSS The first stage in any data analysis is to explore the data collected. Usually we are interested in looking at descriptive statistics such as means, modes, medians,

### Available online:

VOL4 NO. 1 March 2015 - ISSN 2233 1859 Southeast Europe Journal of Soft Computing Available online: www.scjournal.ius.edu.ba A study in Authorship Attribution: The Federalist Papers Nesibe Merve Demir

### Prediction of Useful Reviews on Yelp Dataset

Prediction of Useful Reviews on Yelp Dataset Final Report Yanrong Li, Yuhao Liu, Richard Chiou, Pradeep Kalipatnapu Problem Statement and Background Online reviews play a very important role in information

### Lecture 1: Introduc4on

CSC2515 Spring 2014 Introduc4on to Machine Learning Lecture 1: Introduc4on All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/csc2515_winter15.html

### Computer Vision for Card Games

Computer Vision for Card Games Matias Castillo matiasct@stanford.edu Benjamin Goeing bgoeing@stanford.edu Jesper Westell jesperw@stanford.edu Abstract For this project, we designed a computer vision program

### Artificial Neural Networks

Artificial Neural Networks Outline Introduction to Neural Network Introduction to Artificial Neural Network Properties of Artificial Neural Network Applications of Artificial Neural Network Demo Neural

### Decision Boundary. Hemant Ishwaran and J. Sunil Rao

32 Decision Trees, Advanced Techniques in Constructing define impurity using the log-rank test. As in CART, growing a tree by reducing impurity ensures that terminal nodes are populated by individuals

### Quality Tools. BPF2123 Quality Management System

Quality Tools BPF2123 Quality Management System Chapter Outline Check Sheets Process Flow Diagram Cause-and-Effect Diagram Pareto Diagram Histogram Scatter Diagrams Matrix Analysis Check Sheets A check

### Kobe University Repository : Kernel

Title Author(s) Kobe University Repository : Kernel A Multitask Learning Model for Online Pattern Recognition Ozawa, Seiichi / Roy, Asim / Roussinov, Dmitri Citation IEEE Transactions on Neural Neworks,

### Dimensionality Reduction for Active Learning with Nearest Neighbour Classifier in Text Categorisation Problems

Dimensionality Reduction for Active Learning with Nearest Neighbour Classifier in Text Categorisation Problems Michael Davy Artificial Intelligence Group, Department of Computer Science, Trinity College

### A Quantitative Study of Small Disjuncts in Classifier Learning

Submitted 1/7/02 A Quantitative Study of Small Disjuncts in Classifier Learning Gary M. Weiss AT&T Labs 30 Knightsbridge Road, Room 31-E53 Piscataway, NJ 08854 USA Keywords: classifier learning, small

### Don t Get Kicked - Machine Learning Predictions for Car Buying

STANFORD UNIVERSITY, CS229 - MACHINE LEARNING Don t Get Kicked - Machine Learning Predictions for Car Buying Albert Ho, Robert Romano, Xin Alice Wu December 14, 2012 1 Introduction When you go to an auto

### Linear Regression: Predicting House Prices

Linear Regression: Predicting House Prices I am big fan of Kalid Azad writings. He has a knack of explaining hard mathematical concepts like Calculus in simple words and helps the readers to get the intuition

### A Cartesian Ensemble of Feature Subspace Classifiers for Music Categorization

A Cartesian Ensemble of Feature Subspace Classifiers for Music Categorization Thomas Lidy Rudolf Mayer Andreas Rauber 1 Pedro J. Ponce de León Antonio Pertusa Jose M. Iñesta 2 1 2 Information & Software

### Detection of Insults in Social Commentary

Detection of Insults in Social Commentary CS 229: Machine Learning Kevin Heh December 13, 2013 1. Introduction The abundance of public discussion spaces on the Internet has in many ways changed how we

### CS540 Machine learning Lecture 1 Introduction

CS540 Machine learning Lecture 1 Introduction Administrivia Overview Supervised learning Unsupervised learning Other kinds of learning Outline Administrivia Class web page www.cs.ubc.ca/~murphyk/teaching/cs540-fall08

### An Educational Data Mining System for Advising Higher Education Students

An Educational Data Mining System for Advising Higher Education Students Heba Mohammed Nagy, Walid Mohamed Aly, Osama Fathy Hegazy Abstract Educational data mining is a specific data mining field applied

### 100 CHAPTER 4. MBA STUDENT SECTIONING

Summary Maastricht University is offering a MBA program for people that have a bachelor degree and at least 5 years of working experience. Within the MBA program, students work in groups of 5 during a

### Software Defect Data and Predictability for Testing Schedules

Software Defect Data and Predictability for Testing Schedules Rattikorn Hewett & Aniruddha Kulkarni Dept. of Comp. Sc., Texas Tech University rattikorn.hewett@ttu.edu aniruddha.kulkarni@ttu.edu Catherine

### Automatic Text Summarization for Annotating Images

Automatic Text Summarization for Annotating Images Gediminas Bertasius November 24, 2013 1 Introduction With an explosion of image data on the web, automatic image annotation has become an important area

### 15 : Case Study: Topic Models

10-708: Probabilistic Graphical Models, Spring 2015 15 : Case Study: Topic Models Lecturer: Eric P. Xing Scribes: Xinyu Miao,Yun Ni 1 Task Humans cannot afford to deal with a huge number of text documents

### The Study and Analysis of Classification Algorithm for Animal Kingdom Dataset

www.seipub.org/ie Information Engineering Volume 2 Issue 1, March 2013 The Study and Analysis of Classification Algorithm for Animal Kingdom Dataset E. Bhuvaneswari *1, V. R. Sarma Dhulipala 2 Assistant

### Enriching the Crosslingual Link Structure of Wikipedia - A Classification-Based Approach -

Enriching the Crosslingual Link Structure of Wikipedia - A Classification-Based Approach - Philipp Sorg and Philipp Cimiano Institute AIFB, University of Karlsruhe, D-76128 Karlsruhe, Germany {sorg,cimiano}@aifb.uni-karlsruhe.de

### Switchboard Language Model Improvement with Conversational Data from Gigaword

Katholieke Universiteit Leuven Faculty of Engineering Master in Artificial Intelligence (MAI) Speech and Language Technology (SLT) Switchboard Language Model Improvement with Conversational Data from Gigaword

### Twitter Sentiment Classification on Sanders Data using Hybrid Approach

IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 4, Ver. I (July Aug. 2015), PP 118-123 www.iosrjournals.org Twitter Sentiment Classification on Sanders

### MAT 12O ELEMENTARY STATISTICS I

LAGUARDIA COMMUNITY COLLEGE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF MATHEMATICS, ENGINEERING, AND COMPUTER SCIENCE MAT 12O ELEMENTARY STATISTICS I 3 Lecture Hours, 1 Lab Hour, 3 Credits Pre-Requisite:

### Word Sense Determination from Wikipedia. Data Using a Neural Net

1 Word Sense Determination from Wikipedia Data Using a Neural Net CS 297 Report Presented to Dr. Chris Pollett Department of Computer Science San Jose State University By Qiao Liu May 2017 Word Sense Determination

### Chapter 2: Descriptive and Graphical Statistics

Chapter 2: Descriptive and Graphical Statistics Section 2.1: Location Measures Cathy Poliak, Ph.D. cathy@math.uh.edu Office: Fleming 11c Department of Mathematics University of Houston Lecture 5 - Math

### Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

### SHORT ANSWER. Write the word or phrase that best completes each statement or answers the question.

Math 1342 Review Ch. 1-3 Name 1) Statistics is the science of conducting studies to 1) MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. 2) Which of

### Decision Tree For Playing Tennis

Decision Tree For Playing Tennis ROOT NODE BRANCH INTERNAL NODE LEAF NODE Disjunction of conjunctions Another Perspective of a Decision Tree Model Age 60 40 20 NoDefault NoDefault + + NoDefault Default

### Business Statistics: A First Course, First Canadian Edition Plus MyStatLab

Norean Sharpe, Georgetown University Richard De Veaux, Williams College Paul Velleman, Cornell University Jonathan Berkowitz, University of British Columbia Providing Real Business Context! 2015 October

### SELECTIVE VOTING GETTING MORE FOR LESS IN SENSOR FUSION

International Journal of Pattern Recognition and Artificial Intelligence Vol. 20, No. 3 (2006) 329 350 c World Scientific Publishing Company SELECTIVE VOTING GETTING MORE FOR LESS IN SENSOR FUSION LIOR

### Statistics 571 Statistical Methods for Bioscience I

Statistics 571 Statistical Methods for Bioscience I Lecture 1: Cecile Ane Lecture 2: Nicholas Keuler Department of Statistics University of Wisconsin Madison Fall 2009 Outline 1 Course Information 2 Introduction

### Homework III Using Logistic Regression for Spam Filtering

Homework III Using Logistic Regression for Spam Filtering Introduction to Machine Learning - CMPS 242 By Bruno Astuto Arouche Nunes February 14 th 2008 1. Introduction In this work we study batch learning

### Analysis of Clustering and Classification Methods for Actionable Knowledge

Available online at www.sciencedirect.com ScienceDirect Materials Today: Proceedings XX (2016) XXX XXX www.materialstoday.com/proceedings PMME 2016 Analysis of Clustering and Classification Methods for

### Part-of-Speech Tagging & Sequence Labeling. Hongning Wang

Part-of-Speech Tagging & Sequence Labeling Hongning Wang CS@UVa What is POS tagging Tag Set NNP: proper noun CD: numeral JJ: adjective POS Tagger Raw Text Pierre Vinken, 61 years old, will join the board

### FOCAL POINTS AND TEKS COMPARISON TABLE SET GRADES 6-8

FOCAL POINTS AND TEKS COMPARISON TABLE SET GRADES 6-8 Texas Response to Curriculum Focal Points Highlights Grade 6 The curriculum focal point and its description are presented on the summary page for each

### A Modesto City School Joseph A. Gregori High School 3701 Pirrone Road, Modesto, CA (209) FAX (209)

A Modesto City School Joseph A. Gregori High School 3701 Pirrone Road, Modesto, CA 95356 (09) 550-340 FAX (09) 550-3433 May 4, 016 AP Statistics Parent(s): I am very excited to have your student in AP