Functional Analysis of Artificial Neural Network for Dataset Classification

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Python Machine Learning

INPE São José dos Campos

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Lecture 1: Machine Learning Basics

Artificial Neural Networks written examination

Test Effort Estimation Using Neural Network

Evolutive Neural Net Fuzzy Filtering: Basic Description

Learning Methods for Fuzzy Systems

CS Machine Learning

Softprop: Softmax Neural Network Backpropagation Learning

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Human Emotion Recognition From Speech

A Neural Network GUI Tested on Text-To-Phoneme Mapping

(Sub)Gradient Descent

Classification Using ANN: A Review

Knowledge Transfer in Deep Convolutional Neural Nets

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Soft Computing based Learning for Cognitive Radio

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Axiom 2013 Team Description Paper

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Assignment 1: Predicting Amazon Review Ratings

On the Formation of Phoneme Categories in DNN Acoustic Models

Rule Learning With Negation: Issues Regarding Effectiveness

Calibration of Confidence Measures in Speech Recognition

Modeling function word errors in DNN-HMM based LVCSR systems

Speaker Identification by Comparison of Smart Methods. Abstract

A study of speaker adaptation for DNN-based speech synthesis

Generative models and adversarial training

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

SARDNET: A Self-Organizing Feature Map for Sequences

CSL465/603 - Machine Learning

arxiv: v1 [cs.lg] 15 Jun 2015

WHEN THERE IS A mismatch between the acoustic

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Rule Learning with Negation: Issues Regarding Effectiveness

A SURVEY OF FUZZY COGNITIVE MAP LEARNING METHODS

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Learning to Schedule Straight-Line Code

Seminar - Organic Computing

Data Fusion Through Statistical Matching

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

Kamaldeep Kaur University School of Information Technology GGS Indraprastha University Delhi

Reinforcement Learning by Comparing Immediate Reward

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

CSC200: Lecture 4. Allan Borodin

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Australian Journal of Basic and Applied Sciences

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

An empirical study of learning speed in backpropagation

A student diagnosing and evaluation system for laboratory-based academic exercises

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Attributed Social Network Embedding

Probability and Statistics Curriculum Pacing Guide

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

EGRHS Course Fair. Science & Math AP & IB Courses

Modeling function word errors in DNN-HMM based LVCSR systems

On-Line Data Analytics

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

A Reinforcement Learning Variant for Control Scheduling

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Time series prediction

Evolution of Symbolisation in Chimpanzees and Neural Nets

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Knowledge-Based - Systems

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Using focal point learning to improve human machine tacit coordination

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Laboratorio di Intelligenza Artificiale e Robotica

Speech Emotion Recognition Using Support Vector Machine

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Mathematics subject curriculum

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

SCORING KEY AND RATING GUIDE

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations

Laboratorio di Intelligenza Artificiale e Robotica

Linking Task: Identifying authors and book titles in verbose queries

Learning Methods in Multilingual Speech Recognition

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Early Model of Student's Graduation Prediction Based on Neural Network

Mining Association Rules in Student s Assessment Data

Reducing Features to Improve Bug Prediction

Circuit Simulators: A Revolutionary E-Learning Platform

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach

Word Segmentation of Off-line Handwritten Documents

A Case Study: News Classification Based on Term Frequency

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

ABSTRACT. A major goal of human genetics is the discovery and validation of genetic polymorphisms

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

Transcription:

Functional Analysis of Artificial Neural Network for Dataset Classification 49 Functional Analysis of Artificial Neural Network for Dataset Classification Rojalina Priyadarshini Deptt. Of IT C.V.Raman College Of Engg, Bhubaneswar r.priya77@yahoo.co.in Nillamadhab Dash 2 Deptt. Of IT C.V.Raman College Of Engg, Bhubaneswar nilamadhab4@ gmail.com Tripti Swarnkar, 3 Deptt. Of CA, SOA University Bhubaneswar. tripti_sarap@yahoo.com Rachita Misra, Deptt. Of IT C.V.Raman College Of Engg, Bhubaneswar rachita3@yahoo.com 4 Abstract- Classification is one of the most active research and application areas of artificial neural networks (ANN). One of the difficulties in using ANN is to find the most suitable combination of training, learning and transfer function for classification of data sets with increasing number of features and classified sets. In this paper we have studied the effect of different combinations of functions while using artificial neural network as a classifier and analyzed the suitability of these functions for different kinds of datasets. The appropriateness of the proposed work has been determined on the basis of mean square error, rate of convergence, and accuracy of the classified dataset. Our inferences are based on the simulation results over the datasets used.. Keywords- Artificial Neural Network, Back propagation Algorithm, Classification, Transfer function I.INTRODUCTION Artificial neural networks (ANN) are used in many applications to solve some real world problems. These have attracted the attention of many researchers in different areas, such as neuroscience, mathematics, physics, electrical and computer engineering, and psychology [2]. In commercial purposes, ANNs can be applied to predict the profit, market movements, and price levels based on the market s historical dataset. In medical applications, doctors can evaluate the situation of many patients depending on the historical dataset of other patients who had the same illness. In industry, engineers can apply ANNs to solve many engineering problems such as classifications, prediction, pattern recognition, and non-linear problems [3] where the issues are very difficult or might be impossible to solve through normal mathematical processes. ANNs were used to predict the rain attenuation on an Earth-space path, to predict the water quality index (WQI), and to signal predictions in a nuclear power plant. They were also used in face recognition. In medical applications, ANNs were utilized in detecting different diseases like Thalassaemia classification [8], Breast Cells Classification [9] and Diabetic Retinopathy Classification [6], in computational biology these are used in protein structure prediction, DNA motif prediction etc. Generally, such systems consist of a large number of simple neuron processing units performing computation by a dense mesh of nodes and connections. II. LEARNING STRATEGY ANNs gather their knowledge by detecting the patterns and relationships in data and learn (or are trained) through experience. An ANN is created from hundreds of single units, artificial neurons or processing elements (PE), connected with coefficients (weights), which constitute their architecture in a layered form. The power of neural computations comes from connecting neurons in a network. Each processing element has weighted inputs, transfer function and one output. The working mode of a neural network is determined by the transfer functions of its neurons, by the learning rule, and by the architecture itself. The weighted sum of the inputs forms the activation of the neuron. The activation signal is passed through transfer function to produce a single output of the neuron. Transfer function introduces nonlinearity to the network. During training, the weights are optimized until the error in Special Issue of IJCCT Vol. Issue 2, 3, 4; 2 for International Conference [ACCTA-2], 3-5 August 2

5 Functional Analysis of Artificial Neural Network for Dataset Classification predictions is minimized and the network reaches the prespecified level of accuracy or the convergence is reached. Once the network is trained and tested, it can take new unknown input to predict the output. Learning and generalization are perhaps the most important topics in neural network research [5]. Learning is the ability to approximate the behavior adaptively from the training data while generalization is the ability to predict the training data. Generalization is a more desirable and critical feature because the most common use of a classifier is to make good prediction on new or unknown objects. A number of practical network design issues related to learning and generalization include network size, sample size, model selection, and feature selection, transfer function to be chosen, learning rate parameter etc. In our study we have taken Multidimensional data sets which can represent many real world problems mainly taken from medical background. The classification and clustering of these data sets are meaningful. To test our approach, we have taken different data class sets from repository of data sets. We have divided the data sets into two parts, i.e. training set and testing set which is not used in the training process, and is used to test and then we have simulated our results with these datasets. Almost 2/3 rd of the total dataset has taken as training set and /3 rd of the rest has taken as test set. This is done through the analysis of the accuracy achieved through testing against these data sets. Then we have simulated our network with the same data. A. Dataset Preparation Normalization of input data is used for ranging the values to fall within an acceptable scope, and range []. The features obtained from many specimens are actually considered as raw data. Some amount of pre-processing is always carried out on the input and output data in order to make it suitable for the network. These procedures help obtain faster and efficient training. If the neurons have nonlinear transfer functions (whose output range is from - to or to ), the data is typically normalized for efficiency [5]. Each feature in each dataset is normalized using column normalization in MATLAB. The normalized data are used as the inputs to the network. III. PROPOSED WORK In our work we have taken different datasets by gradually increasing the no of feature attributes. We have designed the neural network to train these datasets by taking different training functions such as GD, GDM, LM, RP (Train Resilient Back propagation) and SCG with different learning function such as LEARNGDM, LEARNGD and with transfer functions TANSIG, PURELIN and LOGSIG. The neural network was trained by back propagation algorithm. Gradient descent method (GDM) was used to minimize the mean squared error between network output and the actual error rate [2]. For different datasets the no of input and output nodes in the network architecture are going on changing with the following learning and training functions. The training error continues to decrease as the number of epoch s increases. Repeated experiments were performed to get the neural network converged. Weights were initialized to random values and networks were run until at least one of the following termination conditions was satisfied. Maximum Epoch 2. Minimum Gradient 3. Performance Goal For testing, the input data was presented to the ANN without weight adjustment. The output of the ANN was compared with the existing class attribute of the datasets. Then we have simulated the network with some of the inputs. We are measuring the efficiency of the network by taking the following parameters:. Rate of convergence 2. No of epochs taken to converge the network. 3. The calculated Mean Square Error (MSE). Training and learning functions are mathematical procedures used to automatically adjust the network's weights and biases. The training function dictates a global algorithm that affects all the weights and biases of a given network. Special Issue of IJCCT Vol. Issue 2, 3, 4; 2 for International Conference [ACCTA-2], 3-5 August 2

Functional Analysis of Artificial Neural Network for Dataset Classification 5 Neural Network Toolbox supports a variety of training algorithms. We have tested some training functions including LM(Train Levenberg-Marquardt) LM is a network training function that updates weight and bias values to Levenberg-Marquardt optimization.,scg (Train scaled conjugate gradient ).SCG is a network training function that updates weight and bias values according to the scaled conjugate gradient method, GD(Train Gradient Descent). GD is a network training function that updates weight and bias values according to gradient descent method and GDM (Train Gradient Descent Momentum). It trains the network with Gradient descent with momentum back propagation method. The learning function can be applied to individual weights and biases within a network. Learning algorithms are used to adapt networks. We have used LEARNGD and LEARNGDM. LEARNGD is the gradient descent weight and bias learning function and LEARNGDM is the gradient descent with momentum weight and bias learning function. IV. RESULTS AND DISCUSSION Data set No of Insta nces (size ) No of predicting attributes (No. of features) No of class attribut e Attribute types 5 4 3 Real 2 2 6 4 Real, integer 3 477 32 2 Categoric al, Integer, Real It was found that the datasets with less number of predicted attributes are quickly converged with training function LM, where as the datasets having large number of predicted attributes are converging with training function SCG with transfer function LOGSIG and learning function LEARNGDM. But it was observed that the SCG function works on datasets having less as well as large predictive attributes Transfer functions calculate a layer's output from its net input. We have tested different Transfer functions like Log sigmoid (LOGSIG), Hyperbolic tangent sigmoid transfer function(tansig) and Linear transfer function(purelin) and from the experiments we have found out that LOGSIG transfer function makes the network converged where as with the other two network is not reaching its convergence. Da ta se t 2 3 Trainin g Functio n LM, SCG LM, SCG SCG Learning function LEARNG DM LEARNG DM LEARN GDM Trans fer functi on LOG SIG LOG SIG LOG SIG No of epo chs 55, 977 34, 68 25 5 MSE.66 667,.33 45.937 5, 4935 332 9 Table-2 summarizes the information about the convergence of the neural network by taking account different parameters on different datasets The graphs below showing the results of the convergence of the network by taking the above said parameters with dataset, 2, 3 as inputs. In X-axis the graphs are showing the number of epochs to reach the convergence and in Y-axis these are showing the MSE(Mean Square Error) Special Issue of IJCCT Vol. Issue 2, 3, 4; 2 for International Conference [ACCTA-2], 3-5 August 2

52 Functional Analysis of Artificial Neural Network for Dataset Classification Fig- [Convergence of neural network with dataset- with training function LM network is converged with 56 epochs and with MSE as.666667.] Fig-3 [Convergence of neural network with dataset-2 with training function LM network is converged with 34 epochs and with MSE as.9375.] Fig-2 [Convergence of neural network with dataset- with training function SCG network is converged with 977 epochs and with MSE as.3345.] Fig-4 [Convergence of neural network with dataset-2 with training function SCG network is converged with 68 epochs and with MSE as 4.93526 e-7.] Special Issue of IJCCT Vol. Issue 2, 3, 4; 2 for International Conference [ACCTA-2], 3-5 August 2

Functional Analysis of Artificial Neural Network for Dataset Classification 53 V. CONCLUSION Fig-5 [Convergence of neural network with dataset-3 with training function SCG network is converged with 255 epochs and with MSE as 3.3297 e-7.] Data Set Actual output Simulated Output.8.... 2.... 3.8629.64.6824 Table-3 represents the actual outputs vs. the outputs generated by our neural networks The back propagation neural network can be used as a highly effective tool for dataset classification with appropriate combination of training, learning and transfer functions. t. In the proposed work we found that the combination of LM, LEARNGDM and LOGSIG works better for comparatively smaller datasets whereas the combination SCG LEARNGDM and LOGSIG is effective for larger datasets. However the efficiency of the neural network may be enhanced by using hybrid learning,training and transfer functions which is the future scope of the paper. REFERENCES [] Fabrizio De Nittis, Giampietro Tecchiolli, and Alessandro Zorat, Consumer Loan Classification Using Artificial Neural Networks, ICSC EIS conference, held in Tenerife,(998) [2] Janusz Starzyk,Jing Pang, Evolvable Binary Artificial Neural Network for Data Classification, Evolutionary Computation pp 5576-5783(2) [3] Guoqiang Peter Zhang, Neural Networks for Classification: A Survey, IEEE Transaction on systems management and Cybernatics- Applications and Reviews Vol-3, No-4,pp 45-462(2). [4] ] M.C. Sezgin, Z. Dokur, T. Ölmez, M. Korürek, Classification of Respiratory Sound by Using An Neural Network, Proceedings 23rd Annual Conference IEEE/EMBS,(2) [5] Sean Van Osselaer, Iris Data Analysis Using Back Propagation Neural Networks, Journal of Manufacturing Systemsvolume 3, issue -4, Page 262(23) [6] R. Sivakumar, Neural Network Based Diabetic Retinopathy Classification Using Special Issue of IJCCT Vol. Issue 2, 3, 4; 2 for International Conference [ACCTA-2], 3-5 August 2

54 Functional Analysis of Artificial Neural Network for Dataset Classification Phase spectral Periodicity components, ICGST-BIME Journal, Volume 7, Issue, (27) [7] Khaled Mohamed Almhdi, Paolo Valigi, Vidas Gulbinas, Rainer Westphal and Rainer Reuter Classification with Artificial Neural Networks and Suport Vector Machines: Application To Oil Fluorescence EARSeL eproceedings 6 pp 5-29( 27) [8]Waranyu Wongseree,Nachol Chaiaiyaratana a, Kanjana Vichittumaros, Pranee Winichagoon, Suthat Fucharoen, Thalassaemia classification by neural networks and genetic programming Science Direct Information Sciences 77 pp 77 786 (27) [9] Nuryanti Mohd. Salleh, Harsa Amylia Mat Sakim and Nor Hayati Othman, Neural Networks to Evaluate Morphological Features for Breast Cells Classification, IJCSNS International Journal of Computer Science and Network Security, VOL.8 No.9, pp 5-58 (28) Special Issue of IJCCT Vol. Issue 2, 3, 4; 2 for International Conference [ACCTA-2], 3-5 August 2