A Novel Approach for Professor Appraisal System In Educational Data Mining Using WEKA

Similar documents
Mining Association Rules in Student s Assessment Data

Test How To. Creating a New Test

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

CS Machine Learning

SECTION 12 E-Learning (CBT) Delivery Module

PowerTeacher Gradebook User Guide PowerSchool Student Information System

Creating a Test in Eduphoria! Aware

DegreeWorks Advisor Reference Guide

Creating an Online Test. **This document was revised for the use of Plano ISD teachers and staff.

Houghton Mifflin Online Assessment System Walkthrough Guide

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Create Quiz Questions

Outreach Connect User Manual

Preferences...3 Basic Calculator...5 Math/Graphing Tools...5 Help...6 Run System Check...6 Sign Out...8

Lectora a Complete elearning Solution

On-Line Data Analytics

Australian Journal of Basic and Applied Sciences

MOODLE 2.0 GLOSSARY TUTORIALS

Using Blackboard.com Software to Reach Beyond the Classroom: Intermediate

CHANCERY SMS 5.0 STUDENT SCHEDULING

Field Experience Management 2011 Training Guides

Urban Analysis Exercise: GIS, Residential Development and Service Availability in Hillsborough County, Florida

Appendix L: Online Testing Highlights and Script

Word Segmentation of Off-line Handwritten Documents

ecampus Basics Overview

A Case Study: News Classification Based on Term Frequency

Using SAM Central With iread

STUDENT MOODLE ORIENTATION

Rule Learning With Negation: Issues Regarding Effectiveness

Managing the Student View of the Grade Center

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Learning From the Past with Experiment Databases

An Introductory Blackboard (elearn) Guide For Parents

Beginning Blackboard. Getting Started. The Control Panel. 1. Accessing Blackboard:

Applications of data mining algorithms to analysis of medical data

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Parent s Guide to the Student/Parent Portal

Intel-powered Classmate PC. SMART Response* Training Foils. Version 2.0

Once your credentials are accepted, you should get a pop-window (make sure that your browser is set to allow popups) that looks like this:

i>clicker Setup Training Documentation This document explains the process of integrating your i>clicker software with your Moodle course.

Specification of the Verity Learning Companion and Self-Assessment Tool

Preparing for the School Census Autumn 2017 Return preparation guide. English Primary, Nursery and Special Phase Schools Applicable to 7.

Python Machine Learning

GACE Computer Science Assessment Test at a Glance

INSTRUCTOR USER MANUAL/HELP SECTION

TeacherPlus Gradebook HTML5 Guide LEARN OUR SOFTWARE STEP BY STEP

Your School and You. Guide for Administrators

TIPS PORTAL TRAINING DOCUMENTATION

TotalLMS. Getting Started with SumTotal: Learner Mode

BLACKBOARD TRAINING PHASE 2 CREATE ASSESSMENT. Essential Tool Part 1 Rubrics, page 3-4. Assignment Tool Part 2 Assignments, page 5-10

ACADEMIC TECHNOLOGY SUPPORT

Connect Microbiology. Training Guide

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Moodle 2 Assignments. LATTC Faculty Technology Training Tutorial

Rule Learning with Negation: Issues Regarding Effectiveness

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Lecture 1: Basic Concepts of Machine Learning

Introduction to Moodle

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Detailed Instructions to Create a Screen Name, Create a Group, and Join a Group

K 1 2 K 1 2. Iron Mountain Public Schools Standards (modified METS) Checklist by Grade Level Page 1 of 11

Storytelling Made Simple

Moodle Student User Guide

Office of Planning and Budgets. Provost Market for Fiscal Year Resource Guide

DO NOT DISCARD: TEACHER MANUAL

Mining Student Evolution Using Associative Classification and Clustering

School Year 2017/18. DDS MySped Application SPECIAL EDUCATION. Training Guide

MyUni - Turnitin Assignments

Schoology Getting Started Guide for Teachers

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

The following information has been adapted from A guide to using AntConc.

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

Automating the E-learning Personalization

Introduction to the Revised Mathematics TEKS (2012) Module 1

16.1 Lesson: Putting it into practice - isikhnas

Filing RTI Application by your own

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

We re Listening Results Dashboard How To Guide

Degree Audit Self-Service For Students 1

Student User s Guide to the Project Integration Management Simulation. Based on the PMBOK Guide - 5 th edition

Adult Degree Program. MyWPclasses (Moodle) Guide

How to set up gradebook categories in Moodle 2.

Customized Question Handling in Data Removal Using CPHC

Skyward Gradebook Online Assignments

Automating Outcome Based Assessment

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

EMPOWER Self-Service Portal Student User Manual

IVY TECH COMMUNITY COLLEGE

Dyslexia and Dyscalculia Screeners Digital. Guidance and Information for Teachers

SCT Banner Financial Aid Needs Analysis Training Workbook January 2005 Release 7

SCT Banner Student Fee Assessment Training Workbook October 2005 Release 7.2

Setting Up Tuition Controls, Criteria, Equations, and Waivers

Getting Started Guide

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

PEIMS Submission 1 list

Odyssey Writer Online Writing Tool for Students

Blackboard Communication Tools

CSL465/603 - Machine Learning

Transcription:

A Novel Approach for Professor Appraisal System In Educational Data Mining Using WEKA 1 Thupakula Bhaskar (Asst.Professor), 2 G.Ramakrishna (Asst.Professor) 1 Department of Computer Engineering, 2 Department of Information Technology Sanjivani College of Engineering, Savitribai Phule Pune University Kopargaon (T), Ahmad Nagar (D), Maharashtra, India Abstract- Data mining, the concept of unseen predictive information from big databases is a powerful novel technology with great potential used in various commercial uses including banking, retail industry, e-commerce, telecommunication industry, DNA analysis remote sensing, bioinformatics etc. Education is a required element for the progress of nation. Mining in educational environment is called Educational Data Mining. Educational data mining is concerned with developing new methods to discover knowledge from educational database. In order to analyze opinion of students about their teachers in Professor Appraisal System, this paper surveys an application of data mining in Professor Appraisal System & also present result analysis using WEKA tool. There are varieties of popular data mining task within the educational data mining e.g. classification, clustering, outlier detection, association rule, prediction etc. How each of data mining tasks can be applied to education system is explained. In this paper we analyze the performance of final Professor Appraisal of a semester of a computer engineering department, Sanjivani College of engineering & is presented the result which it is achieved using WEKA tool. We have verified hidden patterns of Professor Appraisal by students and is predicted that which Professor will be invited to faculty classes and which Professor will be refusing and department heads due to Appraisal reasons will ask explanations with these Professor. Keywords- Classification, Clustering, Association rule, Data mining, Appraisal, WEKA. I.INTRODUCTION Data mining has involved a great deal of responsiveness in the information industry and in society as a whole in recent years, due to the wide availability of huge amounts of data and the forthcoming need for turning such data into useful information and knowledge. The information and knowledge gained can be used for applications ranging from market analysis, fraud detection, and customer retention, to production control and science exploration [1]. Manual data analysis has been around for some time now, but it creates a bottleneck for large data analysis. The transition won't occur automatically; in this case, there is a need for data mining. [2]. Mining applied in education was published in 1995 by Sanjeev and Zytkow. Researchers gathered the knowledge discovery as terms like P pattern for data in the range R from university database [3]. Vranić and Skoćır was examined how to improve some aspects of educational quality with data mining algorithms and techniques by taking a specific course students as target audience in academic environments [4].In this paper I have collected information and results of a appraisal about 30 professors in Sanjivani College of Engineering, Department of Computer Engineering on professor's performances in classroom then with data mining algorithms such Association Rule and decision trees (j48), it is proceeded to analyze and predict acceptation of a professor for continuing the teaching in that subject.there are new rules and relations between selected parameters such as Teaching, Professor Degree, Preparation, Communication, Class Control, Teaching experience, Approved Staff to next semesters on professor appraisal system that is interested for Heads of Departments of Institution. II. Methodology In this research study, We have followed a popular data mining methodology called Cross Industry Standard Process for Data Mining (CRISP-DM), which is a six-step process [5]: Problem explanation: Comprises understanding development goals with business perspective. Understanding the data: Includes identifying the sources of data. Formulating the data: Includes pre-processing, cleaning, and transforming the relevant data into a form that can be used by data mining algorithms. Creating the models: Includes developing a wide range of models using comparable analytical techniques. Assessing the models: Includes evaluating and assessing the validity and the utility of the models against each other and against the goals of the study. Using the model: Includes in such activities as deploying the models for use in decision making processes. 75

B. Background In this research I have used WEKA and Data mining. The following subsections contain a summary of these topics. a.weka WEKA is a group of machine learning algorithms for data mining tasks. The WEKA workbench contains a collection of visualization tools and algorithms for data analysis and Predictive modeling, together with graphical user interfaces for easy access to this functionality [6]. Fig.1.A graphical illustration of the methodology employed in this study A. Data In this study 34 records were used which is taken from feedback_2013_14_sem_1 of Department of computer engineering, Sanjivani College of Engineering. Dataset have professors' information such as Teaching, Preparation, Communication, Class Control along with this I have included Professor Degree, Professor Experience, Approved Staff. Table 1. The list of independent variables used in this study Variable Name Data Description Type Teaching Text Teaching Score Professor_Degree Text Professors Degree Preparation Text Preparation Score Communication Text Communication Score Class_Control Text Class Control Score Teaching_experience Text Teaching experience of Professor Approved_Staff Text Approved Professor or not Table 2. The list of independent variables and values used in this study Variable Name Data Values Type Teaching Text {Excellent,Good,Satisfactory,Poor} Professor_Degree Text {BE,ME,PHD} Preparation Text {Excellent,Good,Satisfactory,Poor} Communication Text {Excellent,Good,Satisfactory,Poor} Class_Control Text {Excellent,Good,Satisfactory,Poor} Teaching_experience Text {TRUE,FALSE} Approved_Staff Text {Yes,No} Teaching score of professors which are studying in Sanjivani College of Engineering, Computer Engineering Department Faculty are represented by the word system. Score ranges of these words are shown in Table 3. Table 3. The output variable (Evaluation score) used in the study Raw Score Score < 60 60<=Score < 75 75<=Score< 85 85<=Score<= 100 Nominal Representation Poor Satisfactory Good Excellent Table 4. The output variable (Teaching experience) used in the study b. Data Mining Data mining is the method of defining interesting knowledge from big amount of data stored in database, data warehouse or other information sources. It includes various tasks such as classification, clustering, association rule etc. c. Association Rule Association rules are used to show the relationship between data items. Association rule generation consists of two separate steps: First, minimum support is applied to find all frequent item sets in a database. Second, these frequent item sets and the minimum confidence constraint are used to form rules [6]. Support & confidence are the normal method used to measure the quality of association rule. Association rule can be used in educational data mining and professor s appraisal system for analyzing the learning data. d. Classification Classification is a data mining task that maps the data into predefined groups & classes. It is also called as supervised learning.it consists of two steps: 1. Model construction: It consists of set of predetermined classes. Each sample is assumed to belong to a predefined class. The set of sample used for model construction is training set. The model is represented as classification rules, decision trees, or mathematical formulae. 2. Model usage: This model is used for classifying upcoming or unidentified objects. The known label of test sample is compared with the classified result from the model. Accuracy rate is the percentage of test set samples that are correctly classified by the model. Test set is independent of training set, otherwise over-fitting will occur [6]. e.clustering Clustering is finding groups of objects such that the objects in one group will be similar to one another and different from the objects in another group. Clustering can be considered the most important unsupervised learning technique. In educational data mining and professor s appraisal system, clustering has been used to group the professors according to their behavior e.g. clustering can be used to distinguish active professor from non-active professor according to their performance in activities. III. ARCHTECTURE OF PROPOSED SYSTEM Raw Years of Teaching Years < 3 Years >= 3 Nominal Representation False TRUE In this paper, it is done a feedback of 2013_14 academic year and semester-1 survey from 319 students then it is prepared results of this survey for 30 professors. 76

a.the Explorer Interface of WEKA Initially preprocess will have been selected. This is the tab you select when you want to tell WEKA where to find the data set that you want to use. WEKA processes data sets that are in its own ARFF format. Conveniently, the download will have set up a folder within the WEKA-3.7 folder called data. b. ARFF format files You do not need to know about ARFF format unless you wish to convert data from other formats. However, it is useful to see the information that such files provide to WEKA. The remainder of the file lists the actual examples, in comma separated format; the attribute values appear in the order in which they are declared above. c. Opening a data set. In the Explorer window, click on Open file and then use the browser to navigate to the data folder within the WEKA-3.7 folder. Select the file called Professor_Appraisal.arff. In this case, the normal usage is to learn to predict the Approved_Staff attribute from four others providing information about the Professor evaluation. The Explorer window should now look like this: @relation Professor_Appraisal.symbolic @attribute Teaching {Excellent, Good, Satisfactory, Poor} @attribute Professor_Degree {BE, ME, PHD} @attribute Preparation {Excellent, Good, Satisfactory, Poor} @attribute Communication {Excellent, Good, Satisfactory, Poor} @attribute Class_Control {Excellent, Good, Satisfactory, Poor} @attribute Teaching_Experience {TRUE, FALSE} @attribute Approved_Staff {Yes, No} @data Excellent, ME, Good, Good, Excellent, TRUE, No Excellent, ME, Good, Excellent, Good, TRUE, Yes Satisfactory, ME, Satisfactory, Good, Satisfactory, TRUE, Yes Satisfactory, ME, Satisfactory, Good, Satisfactory, TRUE, Yes Good, ME, Excellent, Good, Excellent, FALSE, No Excellent, PHD, Excellent, Excellent, Excellent, FALSE, No Good, ME, Good, Good, Satisfactory, TRUE, Yes Good, ME, Good, Good, Good, TRUE, Yes Good, ME, Good, Excellent, Excellent, FALSE, No Good, BE, Good, Satisfactory, Satisfactory, FALSE, No Good, ME, Excellent, Excellent, Excellent, FALSE, No Excellent, BE, Good, Excellent, Excellent, TRUE, No Excellent, BE, Excellent, Excellent, Good, TRUE, No Good, ME, Good, Satisfactory, Good, TRUE, Yes Excellent, ME, Excellent, Excellent, Excellent, FALSE, No Good, BE, Good, Good, Good, FALSE, No Good, BE, Good, Good, Good, TRUE, No Good, ME, Good, Good, Poor, FALSE, No Excellent, ME, Excellent, Good, Excellent, TRUE, Yes Excellent, ME, Good, Satisfactory, Good, TRUE, Yes Good, BE, Good, Excellent, Good, FALSE, No Good, ME, Good, Satisfactory, Satisfactory, FALSE, No Fig.2. ARFF file format for dataset in this paper. It consists of three parts. The @relation line gives the dataset a name for use within WEKA. The @attribute lines declare the attributes of the examples in the data set. Each line specifies an attribute s name and the values it may take. In this paper the attributes have nominal values so these are listed explicitly. In other cases attributes might take numbers as values and in such cases this would be indicated as in the following example: @attribute Teacher_Degree numeric Fig.3. Opening Professor appraisal data set in WEKA Most of the information it displays is self-explanatory: it is a data set containing 34 examples (instances) each of which has 7 attributes. The Approved_Staff attribute has been suggested as the class attribute (i.e. the one that will be predicted from the others). Most of the right hand of the window gives you information about the attributes. Initially, it will give you information about the first attribute ( Teaching ). This shows that it has 4 possible values tells you how many there are of each value. The bar chart in the lower right shows how the values of the suggested class variable are distributed across the possible values of the Teaching. If you click on Professor_Degree in the panel on the left, the information about the Teaching attribute will be replaced by the corresponding information about the Teacher Degree attribute. d. Choosing a classifier Next it is necessary to select a machine learning procedure to apply to this data. The task is classification so click on the classify tab near the top of the Explorer window. By default, a classifier called ZeroR has been selected. A different classifier is desired so click on the Choose button. A hierarchical popup menu appears. Click to expand Trees, which appears at the end of this menu, then select J48 which is the decision tree program. The Explorer window now looks like this indicating that J48 has been chosen. 77

Fig.4. Decision tree with J48 tree in WEKA. The other information alongside J48 indicates the parameters that have been chosen for the program. This paper will ignore these. e. Choosing the experimental procedures The panel headed Test options allows the user to choose the experimental procedure. This paper shall has more to say about this later in the course. For the present exercise click on Use training set. (This will simply build a tree using all the examples in the data set). The small panel half way down the left hand side indicates which attribute will be used as the classification attribute. It will currently be set to Approved_Staff. (Note that this is what actually determines the classification attribute the class attribute on the pre-process screen is simply to allow you to see how a variable appears to depend on the values of other attributes). f.running the decision tree program Now, simply click the start button and the program will run. The results will appear in the scrollable panel on the right of the Explorer window. Normally these will be of great interest but for present purposes all this paper needs to notice is that the resulting tree classified all 14 training examples correctly. The tree constructed is presented in indented format, a common method for large trees: Fig.5. Decision tree with J48 tree in WEKA. The panel on the lower left headed Result list (right-click for options) provides access to more information about the results. Right clicking will produce a menu from which Visualize Tree can be selected. This will display the decision tree in a more attractive format: Fig.6. Visualize tree with J48 tree in WEKA. Note that this form of display is really only suitable for small trees. Comparing the two forms should make it clear how the indented format works. IV. CONCLUSION At Professor s appraisal, teaching score of students is very important factor that many colleges/universities gather this information on performance of professor. New rules by using data mining and J48 tree as a decision tree in this paper are results that heads of institutes could use these rules in future decisions to submit new professors and continue with elected old professors. For example is discovered these rules as you see in Fig.6: 1- IF(Teaching=GOOD)THEN (Approved_Staff is Yes means next semester who continue his/her teaching) 2- IF(Teaching=Excellent)AND (Teaching_experience=FALSE means is low)then(approved_staff is Yes means next semester who continue his/her teaching) And etc. Correctness of this rules depending variety of datasets and statistical instances can vary. But data mining tools such as WEKA as is showed in this paper can conclude variety results that help head of the departments in Engineering Colleges. These results will be used by HOD s in decisionmaking. V. REFERENCES 1. Jiawei Han. and Micheline Kamber,Jian Pei "Data Mining: Concepts and Techniques", 3 rd edition. 2. Sunita B Aher, Mr. LOBO L.M.R.J. Data Mining in Educational System using WEKA, IJCA,2011 3. A. P. Sanjeev ve J. M, Zytkow. Discovering Enrollment Knowledge in University Databases, 1th Conference on KDD (Montreal.20-21 August 1995). 4. M.Vranić, D. Pintar, Z.Skoćır, The Use of Data Mining in Education Environment, ConTEL 2007 (Zagrep 13-15 June 2007), 243. 5. C. Shearer, The CRISP-DM model: The new blueprint for data mining Journal of Data Warehousing, (2000). 5: 13-22. 6. International Educational Data Mining, available at http://www.educationaldatamining.org/ 78

AUTHORS PROFILE T.Bhaskar received a M.Tech(CSE) from JNTU Hyderabad..He is currently working as Asst. Professor in Computer Engg, Sanjivani College of Engineering, Kopargaon, Maharashtra India. His research interest includes data mining, network security. G.Ramakrishna is currently working as Asst. Professor in IT Dept, Sanjivani College of Engineering, Kopargaon, Maharashtra India. His research interest includes data mining,bigdata analytics. 79