Case Study on Classification of Glass using Neural Network Tool in MATLAB

Similar documents
Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

A Case Study: News Classification Based on Term Frequency

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Learning Methods for Fuzzy Systems

INPE São José dos Campos

A student diagnosing and evaluation system for laboratory-based academic exercises

Mining Association Rules in Student s Assessment Data

AP Chemistry

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Early Model of Student's Graduation Prediction Based on Neural Network

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Issues in the Mining of Heart Failure Datasets

Evolutive Neural Net Fuzzy Filtering: Basic Description

Speaker Identification by Comparison of Smart Methods. Abstract

The Use of Statistical, Computational and Modelling Tools in Higher Learning Institutions: A Case Study of the University of Dodoma

Assignment 1: Predicting Amazon Review Ratings

Applications of data mining algorithms to analysis of medical data

Test Effort Estimation Using Neural Network

Soft Computing based Learning for Cognitive Radio

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Australian Journal of Basic and Applied Sciences

Human Emotion Recognition From Speech

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

SARDNET: A Self-Organizing Feature Map for Sequences

CS Machine Learning

Software Maintenance

arxiv: v1 [cs.lg] 15 Jun 2015

Word Segmentation of Off-line Handwritten Documents

Lecture 1: Machine Learning Basics

Rule Learning With Negation: Issues Regarding Effectiveness

Knowledge Transfer in Deep Convolutional Neural Nets

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Reinforcement Learning by Comparing Immediate Reward

On-Line Data Analytics

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Python Machine Learning

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach

Evolution of Symbolisation in Chimpanzees and Neural Nets

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Artificial Neural Networks written examination

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Classification Using ANN: A Review

Rule Learning with Negation: Issues Regarding Effectiveness

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Applying Learn Team Coaching to an Introductory Programming Course

Dinesh K. Sharma, Ph.D. Department of Management School of Business and Economics Fayetteville State University

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

Speech Emotion Recognition Using Support Vector Machine

Modeling function word errors in DNN-HMM based LVCSR systems

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Comparison of EM and Two-Step Cluster Method for Mixed Data: An Application

CSL465/603 - Machine Learning

Automating the E-learning Personalization

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

Probabilistic Latent Semantic Analysis

Kamaldeep Kaur University School of Information Technology GGS Indraprastha University Delhi

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Softprop: Softmax Neural Network Backpropagation Learning

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

An Effective Framework for Fast Expert Mining in Collaboration Networks: A Group-Oriented and Cost-Based Method

Data Fusion Through Statistical Matching

K-Medoid Algorithm in Clustering Student Scholarship Applicants

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Time series prediction

Research computing Results

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

A Pipelined Approach for Iterative Software Process Model

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Computational Data Analysis Techniques In Economics And Finance

Modeling function word errors in DNN-HMM based LVCSR systems

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Constructing a support system for self-learning playing the piano at the beginning stage

OPAC and User Perception in Law University Libraries in the Karnataka: A Study

A study of speaker adaptation for DNN-based speech synthesis

Artificial Neural Networks

Linking Task: Identifying authors and book titles in verbose queries

Calibration of Confidence Measures in Speech Recognition

Seminar - Organic Computing

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

arxiv: v1 [cs.cl] 2 Apr 2017

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Using the Artificial Neural Networks for Identification Unknown Person

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

(Sub)Gradient Descent

Multisensor Data Fusion: From Algorithms And Architectural Design To Applications (Devices, Circuits, And Systems)

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Learning From the Past with Experiment Databases

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Transcription:

Case Study on Classification of Glass using Neural Network Tool in MATLAB Devika Chhachhiya Research Scholar Dept. of CS& IT THE IIS University, Amita Sharma Asst. Professor Dept. of CS& IT THE IIS University, Manish Gupta Project Director(OSD) Dept. of Science& Tech. Govt. of Rajasthan ABSTRACT: This paper encompasses application of the Neural Network Tool (NN Tool) in the glass classification problem and also discusses the correlation of the different activation functions with the Mean Square Error (MSE). This paper works on the glass data classification and finds the impact of different Activation functions on the error obtained while training and testing of the neural network model created by the NN Tool provided by the MATLAB Toolbox. Experiment was conducted on the MATLAB (NN tool) with glass data it has been observed that LOGSIG function gives the minimum MSE and gives more accurate results in comparison to the other activation functions provided by MATLAB (NN Tool). This paper highlights the relation of the nature of the dataset and the activation functions on the error obtained from the training of neural network model. In future by observing the limitations and effect of the different parameters such as number of hidden layers, activation functions, nature of the data, adjustments of weights, size of data and many more on the network modeling we will be able to understand and develop an improved algorithm and data mining tool for neural network classification technique with more accurate results. Keywords Neural Networks, Data mining, Activation Function, MATLAB. 1. INTRODUCTION Neural network is getting popular now-a-days due to its capability of generalization and resistance to the noisy and erroneous data. A lot of research is undergoing to improve the efficiency and accuracy of the modeling and training of the neural network architecture. Although lot of research work has been done but still there is no profound theory about the exact number of size of the data, architecture of the network and the best suitable algorithm for the NN modeling because selection of these parameters depends on the nature of the data. In this paper we will study glass classification problem and find the values of different parameters suitable for this problem such that we obtain better and accurate results. In this paper, section 2 describes neural network and tools used for data mining. Section 3 and section 4 discusses the case study and experimental model. Section 5 and section 6 illustrates the result and conclusion. 2. NEURAL NETWORK Neural Network is the biological structure inspired by the working of human nervous system. Neural Network is being applied widely on different application areas due to its learning ability i.e. capability to extract rules and learn from the data and create a network model which can be used for classification, pattern recognition and forecasting on the data. Most promising feature of the Neural network that other classification techniques do not possess is that it helps to simulate the network and create model which can be used further and applied on the new data which was not previously exposed to the network. MATLAB is a data mining tool which provides Neural Network Toolbox for Neural Network modeling. Neural Network Toolbox consist of NN tool which helps to create a neural network model to train and test the data for classification of the data, find hidden patterns, clustering and future forecasting. There are several other neural network tools like SPSS but MATLAB NN Tool is popular because of large number of functionality support. 11

3. GLASS DATASET: CASE STUDY We have used glass dataset for classification because it can prove to be very useful in crime investigation, behavior of glass material etc. The small pieces of glasses obtained from the crime sites can be very crucial evidence for the investigation purpose. We can create a neural network model and train it on glass data which can further help us to classify the unknown type of glass which was previously not exposed to the model. There are 108 instances in the glass dataset with 10 input attributes and 1 output attribute (Type of glass). Attribute Information: (Input attributes) 1) Id number 2) RI: refractive index 3) Na: Sodium 4) Mg: Magnesium 5) Al: Aluminum 6) Si: Silicon 7) K: Potassium 8) Ca: Calcium 9) Ba: Barium 10) Fe: Iron Type of glass: (output attributes) 1. Building_windows_float_processed (20) 2. Building_windows_non_float_processed (20) 3. Vehicle_windows_float_processed (17) 4. Vehicle_windows_non_float_processed (0) 5. Containers (13) 6. Tableware (9) 7. Headlamps (29) Details of the input and output attributes are explained above. Type of glass will depend on the values of the input attributes and by these values network will be trained and tested to classify that the record will lie in which Type of Glass. In Table 1, sample dataset has been displayed. 4. EXPERIMENTATION Experiment was performed on the NN Toolbox provided by the MATLAB with the glass dataset as explained above. In this experiment train data set taken is 70% of the total data set and test dataset is 30% of the total data set. We have created the neural network model by the help of the train data set and varied the Activation functions associated with the layers of the network and analyzed its effect on the classification of the type of glass and MSE obtained in the results. Fig 1: and Fig 2: are the screen shots taken from the MATLAB NN Tool. Fig 1: shows the neural network model created for glass data classification. This figure describes the architecture of the model created and how input parameters are connected to hidden layers and further hidden layers are connected to output layer. Fig 2: shows the detail of the algorithm applied, type of error selected for training the network i.e. MSE (Mean Square Error) and we can also see the performance chart and the regression chart from it. This figure also gives us value of the error obtained while training the network Table 1. Sample data of the actual glass dataset RI Na Mg AL Si K ca Ba Fe GLASS 1.5210 13.64 4.49 1.1 71.78 0.06 8.75 0 0 1 1.5176 13.89 3.6 1.36 72.73 0.48 7.83 0 0 1 1.5159 13.02 3.56 1.54 73.11 0.72 7.9 0 0 2 1.5159 13.02 3.58 1.51 73.12 0.69 7.96 0 0 2 1.5161 13.33 3.53 1.34 72.67 0.56 8.33 0 0 3 1.5167 13.24 3.57 1.38 72.7 0.56 8.44 0 0.1 3 1.5151 14.01 2.68 3.5 69.89 1.68 5.87 2.2 0 5 1.5191 12.73 1.85 1.86 72.69 0.6 10.1 0 0 5 1.5111 17.38 0 0.34 75.41 0 6.65 0 0 6 1.5113 13.69 3.2 1.81 72.81 1.76 5.43 1.19 0 7 Fig 1: Neural Network Model created by the MATLAB. 12

Table 2. Relationships between activation function and training of network with respect to MSE No of times training Activation MSE(Mean performed On the Function Square Error) network model ( first time) LOGSIG 0.0213 ( second time) LOGSIG 0.00817 ( first time) PURELIN 0.365 ( second time) PURELIN 1.369 ( first time) TANSIG 0.783 ( second time) TANSIG 0.089 Table 3. Comparison between actual output and output obtained by neural network model for activation function LOGSIG Target data(type of glass) Output obtained by MATLAB for LOGSIG function (Type of Glass) 1 1.0773 1 1.0759 2 1.9898 Fig 2: Neural Network model training by the MATLAB 5. RESULT ANALYSIS Experimental results are shown in Table 2 and Table 3. From Table 2. It has been observed that: 5.1) The impact of the different activation functions on the error (MSE) while training the neural network model. 5.2) It also shows that error (MSE) decreases by the number of times we train the network. 5.3) From the table given below it has been observed that the LOGSIG function on the second training time has obtained the least MSE i.e. 0.00817 of all the activation functions. Table 3. Shows the result obtained by applying the test dataset on the trained model created by the NN Tool in MATLAB for LOGSIG activation function. Target data column consists of the actual result of the glass type which should be obtained by the trained network model for test data and output obtained is the column of the result obtained by the trained network model for activation function LOGSIG. 2 2.8754 3 1.8978 3 3.3664 5 7 5 6.9006 6 6.0912 6 3.4982 7 6.9485 7 6.9559 The graphs obtained between MSE and no. of epochs for all the three activation functions LOGSIG, PURELIN AND TANSIG are described in Fig 3, Fig 4 and Fig 5 respectively as shown below. 13

6. DISCUSSION AND CONCLUSION By the above results obtained we have observed that 6.1. From the three activation functions LOGSIG, PURELIN AND TANSIG provided by NN Tool in MATLAB LOGSIG have produced the least MSE i.e. 0.00871 which is very less error as compared to other functions i.e. LOGSIG function is most suitable activation function for glass dataset 6.2. It has been found that as the number of times we train our network model MSE decreases as observed from the results obtained from Table 2. but overtraining may also lead to increase in error. 6.3. Table 3. shows that there are some misclassifications. For example although the actual data has to lie glass type 5 by the trained model shows the record to lie in glass type 7 which is an important point to discuss. It may also happen that nature of the input attributes of that record is such that it must actually lie in class 7 as calculated by neural network. In future we will study different parameters and find their effect on the results for glass data set. By studying all the aspects we can design a better algorithm and tool for neural network. Fig 4: Performance Chart obtained by the MATLAB for PURELIN function Fig 3: Performance Chart obtained by the MATLAB for LOGSIG function Fig 5: Performance Chart obtained by the MATLAB for TANSIG function 14

8. REFERENCES [1] Moreira, M., andfiesler, E., Oct 1995 Neural Networks with adaptive Learning Rate and Momentum Terms, IDAIP Technical Report. [2] Li, X.P., and Gao, J., Pad modeling by using artificial neural network", Progress inelectromagnetic Research, PIER 74, PP 167-180,2007. [3] Singh, Y., and Chauhan,A. S., NEURAL NETWORKS IN DATA MINING, Journal oftheoretical and Applied Information Technology 2005-2009 JATIT. [4] Suresh, G.V., Suresh Babu, K.,Karunakar, K., And Vijaya Kumar, N., A Novel Based Approach For Uncertain Data Classification Using Probabilistic Neural Networks, Journal of theoretical andapplied Information Technology, Vol 27, No. 2, 31May, PP 77-84, 2011. [5] Amarendra, K., Lakshmi, K.V., and Ramani, K.V., Research on Data Mining Using Neural Networks, International Journal of Computer Science & Informatics (IJCSI), Vol.- II, Issue-1, 2011. [11] Guide, Mathwork, 2013. [6] Arulmozhi, V., Classification task by using MATLAB Neural Network Tool Box A Beginner s View, International Journal of Wisdom Based Computing,Vol. 1 (2), PP 59-60, August 2011 [7] Faissal Mili and Manel Hamdi, A hybrid Evolutionary Functional Link Artificial Neural Network for Data mining and Classification, International Journal of Advanced Computer Science and Applications (IJACSA), Vol. 3, No.8,PP 89-95, 2012. [8] Arulmozhi, V., Classification task by using MATLABNeural Network Tool Box A Beginner s View, International Journal of Wisdom Based Computing,Vol. 1 (2), PP 59-60, August 2011. [9] Hudson, Mark., Martin, Beale., Hagan, T.,Demuth, Howard B., Neural Network Toolbox 7 User s Guide. [10] M. H. Beale, M. T. Hag, H. B. Demuth, Neural Network Toolbox User s 15