Character Recognition by Levenberg-Marquardt (L-M) Algorithm Using Back Propagation ANN

Similar documents
Artificial Neural Networks written examination

Python Machine Learning

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Test Effort Estimation Using Neural Network

Learning Methods for Fuzzy Systems

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Evolutive Neural Net Fuzzy Filtering: Basic Description

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

INPE São José dos Campos

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Artificial Neural Networks

Kamaldeep Kaur University School of Information Technology GGS Indraprastha University Delhi

Softprop: Softmax Neural Network Backpropagation Learning

Human Emotion Recognition From Speech

Word Segmentation of Off-line Handwritten Documents

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Axiom 2013 Team Description Paper

Time series prediction

SARDNET: A Self-Organizing Feature Map for Sequences

Knowledge Transfer in Deep Convolutional Neural Nets

Lecture 1: Machine Learning Basics

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

Knowledge-Based - Systems

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Soft Computing based Learning for Cognitive Radio

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

Seminar - Organic Computing

Calibration of Confidence Measures in Speech Recognition

Speech Emotion Recognition Using Support Vector Machine

Speaker Identification by Comparison of Smart Methods. Abstract

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Early Model of Student's Graduation Prediction Based on Neural Network

Modeling function word errors in DNN-HMM based LVCSR systems

Software Maintenance

Circuit Simulators: A Revolutionary E-Learning Platform

Classification Using ANN: A Review

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

A Reinforcement Learning Variant for Control Scheduling

(Sub)Gradient Descent

CS Machine Learning

arxiv: v1 [cs.cv] 10 May 2017

On the Formation of Phoneme Categories in DNN Acoustic Models

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Evolution of Symbolisation in Chimpanzees and Neural Nets

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach

A study of speaker adaptation for DNN-based speech synthesis

Reducing Features to Improve Bug Prediction

Modeling function word errors in DNN-HMM based LVCSR systems

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Probabilistic Latent Semantic Analysis

Learning to Schedule Straight-Line Code

Rule Learning With Negation: Issues Regarding Effectiveness

Issues in the Mining of Heart Failure Datasets

GACE Computer Science Assessment Test at a Glance

Rule Learning with Negation: Issues Regarding Effectiveness

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

Second Exam: Natural Language Parsing with Neural Networks

arxiv: v1 [cs.lg] 15 Jun 2015

Deep Neural Network Language Models

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

CSL465/603 - Machine Learning

An empirical study of learning speed in backpropagation

Model Ensemble for Click Prediction in Bing Search Ads

International Journal of Advanced Networking Applications (IJANA) ISSN No. :

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

WHEN THERE IS A mismatch between the acoustic

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

Improvements to the Pruning Behavior of DNN Acoustic Models

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

On-Line Data Analytics

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

A SURVEY OF FUZZY COGNITIVE MAP LEARNING METHODS

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Applying Fuzzy Rule-Based System on FMEA to Assess the Risks on Project-Based Software Engineering Education

arxiv: v2 [cs.ro] 3 Mar 2017

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

SIE: Speech Enabled Interface for E-Learning

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Using the Artificial Neural Networks for Identification Unknown Person

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS

Automating the E-learning Personalization

AUTOMATED TROUBLESHOOTING OF MOBILE NETWORKS USING BAYESIAN NETWORKS

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Lecture 1: Basic Concepts of Machine Learning

Laboratorio di Intelligenza Artificiale e Robotica

Transcription:

Character Recognition by Levenberg-Marquardt (L-M) Algorithm Using Back Propagation ANN Surabhi Varshney 1, Rashmi Chaurasiya 2, Yogesh Tayal 3 Department of Electronics and Instrumentation Engineering Institute of Technology and Management, Universe Gwalior, Madhya Pradesh, India varshney.49@gmail.com, Abstract The Author dedicatedly emphasis the character recognition that is applied vigorously on various techniques and the comparison of analysis has been done to justify the Network. Basis of complexity of task the network algorithm has been designed and developed and the recognition pattern is trained. The character recognition sequenced has been ranged on characters data and available technique. The emphasis has been given on the comparison and to increase the recognition accuracy and decreasing the recognition time. The character recognition interface includes the recognition of defined characters made available in the database and the integration of it. character recognition system is implemented for the characterisation of English alphabets with customised specific requirements using most contemporary optimisation algorithms ( Levenberg-Marquardt ) in back-propagation Multi layered Feed -forward network in Artificial Neural Network. The ANN training pattern has been done with most accuracy to the Characters from (A-J) are created with size of each character is n x n square matrix form. Keywords- Character recognition, Artificial Neural Network, Multi layered Feed -forward network, back-propagation, Levenberg-Marquardt Algorithms. ***** I. INTRODUCTION The Character Recognition has been the one of the most significant application in the development area of Artificial Neural Network field in the arena of Artificial Intelligence. It is a process of automatic computer recognition and identification of characters in which a scanned and digitized character image is converted into machine-editable format. Classical methods in pattern recognition were not sufficient for recognition of text characters due to varying shape, size and style of characters. So, intelligent systems are designed which are capable of recognizing characters of changing fonts. But still it is the case of less efficiency and accuracy. Character recognition is an important topic in pattern recognition which is the backbone of machine perception. The objective of character recognition is to translate human readable characters to machine readable codes. Character recognition has been applied to various areas such as automatic postal address readers, automations and automatic data entry from paper documents to computers. works in a manner similar to the biological system. It is not that the ANN must resemble the above said biological system, but the major learning steps can be reliably imitated. Mathematical model of ANN- A very simple mathematical model of a neuron was first presented by McCulloch and Pitts. In their model, there are n input nodes I 1, I 2, I 3...I n, and one output node O. The operation of the neuron is as follows: II. ARTIFICIAL NEURAL NETWORK Artificial Neural Network (ANN) are a kind of soft computing tools that help us recognizing patterns besides allowing us to accomplish a lot of other tasks. The essential idea of learning by an ANN is borrowed from the neural learning system of human being. The recognition capability of human being is due to the potential of nerve cells to store and propagate information. Keen observation over the biological neuro system enables us with the insight of the learning process. This knowledge has been applied in the design of ANN that Figure 1: Mathematical Model of ANN O i = f ( W ij I j ) Each of these inputs is multiplied by a connection weight. The weights are represented by w (n). These products are summed, fed to a transfer function (activation function) to generate a result, and this result is sent as output. This is also possible with other network structures, which used different threshold functions as well as different transfer function. A. 2368

B. Type of Artificial Neural Network- Else a) Single-Layer Feed-Forward l vl Return W end b) Multi-Layer Feed-Forward c) Recurrent d) Self-Organizing Map (SOM) e) BRANN f) Back-Propagation g) Hopfield Network h) Radial Basis Function. III. BACK PROPAGATION NETWORK The Back propagation Network is an widely accepted, specific tool that generalised and optimised the performance of real time system under the hood of multilayer network and non linear function. By using different set of step it is defined using specific rules and algorithms and the result is compared for nonlinear networks. These techniques classify & predict the network and to solve the wide ranging of problem basis of real time performance that has to be optimised for online and offline mode that could be beneficial. A successful character recognition methodology depends heavily on the choice of the alphabets used by the pattern classifier. The back-propagation is widely used learning algorithm in training multilayer perceptrons (MLP). the MLP refer to the network consisting of a set of sensory units (source nodes) that constitute one input layer, one or more than one hidden layers of computation nodes, and one output layer of computation nodes.the input signal propagates through the network in a forward direction, from left to right and on a layer by layer basis. IV. LLEVENBERG-MARQUARDT(L-M) ALGORITHM - It is the most sophisticated, widely used optimal and fastest algorithm although it takes large amount of Memory. this method is a second order approach and for training speed without having to compute the Hessian Matrix. it is safe and fast algorithm as compare to other algorithm,but it was found that the efficiency of the method decreases as the number of weight increases as the size of the Hessian increases and requires a large capacity memory. Levenberg-Marquardt Algorithm - V. MYTHOLOGY The application of L-M to neural network training is described this algorithm appears to be the fastest method for training moderate-sized feed-forward neural networks (up to several 1000 weight). It also has an efficient implementation in MATLAB software. Step1-The code creates training all the input vectors are placed in one matrix. Step2 - Create the feed-forward network. Step3- You might want to modify some of the default training parameters. If you want to use the default training parameters, the preceding commands are not necessary. Step4- Now you are ready to train the network. Step5- Simulate the trained network to obtain its response to the inputs in the training set. Input: R n R a function such that f (i)= Output: t, o local minimum of the cost function f Begin k 0 w w 0 whileupdate 0 and k < k max do findd such that ([H+l.I]ˉ1. ÑJ).d =J T *f W W lm If f(w lm ) < f(w) then W = W lm l l/v Flow chart 1 VI. RESULT This chapter discusses the use of artificial neural network for the purpose of character recognition. Characters from (A-J) are created with size of each character is 6 x 6. The matrix are represented as 2369

Figure 2: Characters for training (A-J) 0 0 1 0 1 0 1 0 0 0 0 1 For character recognition following neural network topology is used 0 1 0 0 0 1 A B 1. Multi layered feed forward architecture. 0 1 1 1 1 1 1 0 0 0 0 1 2. Recurrent neural network. 0 1 0 0 0 1 1 0 0 0 0 1 0 1 0 0 0 1 1 1 1 1 1 0 Multi layered feed forward architecture In multi-layered feed forward architecture, the following 1 1 1 1 1 1 1 1 1 1 0 0 training algorithms are used 1 0 0 0 1 0 1. Levenberg-Marquardt 1 0 0 0 0 1 2. Gradient descent C D 1 0 0 0 0 1 1 0 0 0 1 0 Table 1: ANN Parameter for Levenberg-Marquardt Learning Rule 1 1 1 1 1 1 1 1 1 1 0 0 S.No Multi-layered feedforward 1 1 1 1 0 0 1 1 1 1 1 1 ANN Architecture 1. Maximum iteration 1000 1 1 1 1 1 1 1 1 1 1 1 1 2. Learning rate 0.01 E F 3. Training goal 0.001 sigmoidal function and 4. Activation function 1 1 1 1 1 1 linear function 5. Number of input neurons 10 1 1 1 1 1 1 1 0 0 0 0 1 6. Number of hidden layer 1 1 0 0 0 0 1 Number of neurons in 1 0 0 0 0 1 7. 40 G hidden layer H 1 0 1 1 1 1 1 1 1 1 1 1 8. Error criteria MSE 1 0 0 0 0 1 1 0 0 0 0 1 9. Training Algorithm Levenberg- Marquardt 1 0 0 0 0 1 Number of output 10. 10 1 1 1 1 1 1 1 1 1 1 1 1 neurons 0 0 0 1 0 0 0 0 0 1 0 0 I J 1 0 0 1 0 0 1 0 0 1 0 0 1 1 1 1 1 1 0 1 1 0 0 0 Figure 3: 2370

Table 2: Testing results for Levenberg-Marquardt learning rule Character tested Regression Accuracy (%) 1 A 0.944 100 2 B 0.972 100 3 C 0.947 100 4 D 0.841 100 5 E 0.891 100 6 F 0.897 100 7 G 0.874 100 8 H 0.877 100 9 I 0.954 100 10 J 0.978 100 100% (Avg) Figure 4: Error v/s epoch curve for Levenberg-Marquardt learning rule Figure 5: Regression characteristics Levenberg-Marquardt of learning rule VII. CONCLUSION & FUTURE SCOPE In this research paper an attempt was made to develop an artificial neural network based system for character recognition. Using L-M algorithm that was used to train feedforward multilayer back-propagation neural network. The L-M method provide more stability in the error performance at different learning rates and better performance with 100% accuracy rate for recognition of printed English alphabet characters. In future scope, we plan to explore in the field of recognition of handwritten characters and textured characters in different languages with most efficient results. VIII. REFERENCES [1] Badi, m. Boutalline and s. Safi, The Neural Networks: Application and Optimization Application Of Levenberg- Marquardt Algorithm For Tifinagh Character Recognition, International Journal of Science, Environment and Technology, Vol. 2, No 5, 2013, 779 786, ISSN 2278-3687 (O). [2] Batsukh Batbayar and Xinghuo yu, An Improved Levenberg- Marquardt Learning Algorithm for Neural Networks Based on Terminal Attractors, 1-4244-2386-6/08/$20.00 2008 IEEE. [3] Gaganjot kaur and monika aggarwal, Artificial Intelligent System For Character Recognition Using Levenberg-Marquardt Algorithm,International Journal Of Advanced Research In Computer Science And Software Engineering, Volume 2, Issue 5, May 2012, ISSN: 2277 128x. [4] Suruchi G. Dedgaonkar, Anjali A. Chandavale and Ashok M. Sapkal, Survey of Methods for Character Recognition, International Journal of Engineering and Innovative Technology (IJEIT) Volume 1, Issue 5, May 2012, ISSN: 2277-3754. [5] Julio C Rodriguez-Quinonezn, Oleg Sergiyenko, Felix F. Gonzalez-Navarro, Luis Basaca-Preciado and VeraTyrsa, Surface recognition improvement in 3D medical laser scanner using Levenberg Marquardt method, 0165-1684/$ - seefrontmatter & 2012 Published byelsevierb.v. 2371

[6] Y. Fujisawa, Meng Shi, T. Wakabayashi and F. Kiruma, Handwritten numerical recognition using gradient and curvature of gray scale image, in Proc. 5 th Int. Conf. Document Analysis and Recognition, pp. 277-280, 1999. [7] U. Pal and B B Chaudhuri, Indian script character recognition: a survey, Pattern Recognition, 37 (9), 2004, pp. 1887-1899. [8] Y. Hamamoto, S. Uchimura, S. Tomia, On the behavior of artificial neural network classifiers in high dimensional spaces, IEEE Transactions Pattern Analysis and Machine Intelligence, 18, 5, pp. 571-574, 1996 [9] Rafel. C. Gonzalez and Richard E. Woods, Digital image processing, 3 rd Ed, Prentice Hall, 2006 [10] Yogesh Tayal, Ruchika Lamba, Subhransu Padhee, Automatic face detection using color based segmentation, International Journal of Scientific and Research Publication, vol. 2, no. 6, 2012, pp.1-7. 2372