J. Spanner EPRI NDE Center Charlotte, NC 28262

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Learning Methods for Fuzzy Systems

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Python Machine Learning

Evolutive Neural Net Fuzzy Filtering: Basic Description

Artificial Neural Networks written examination

Word Segmentation of Off-line Handwritten Documents

Artificial Neural Networks

INPE São José dos Campos

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Knowledge-Based - Systems

Axiom 2013 Team Description Paper

SARDNET: A Self-Organizing Feature Map for Sequences

A Reinforcement Learning Variant for Control Scheduling

Human Emotion Recognition From Speech

CS Machine Learning

Rule Learning With Negation: Issues Regarding Effectiveness

Test Effort Estimation Using Neural Network

Modeling function word errors in DNN-HMM based LVCSR systems

Circuit Simulators: A Revolutionary E-Learning Platform

Seminar - Organic Computing

Lecture 1: Basic Concepts of Machine Learning

A study of speaker adaptation for DNN-based speech synthesis

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

EGRHS Course Fair. Science & Math AP & IB Courses

On the Formation of Phoneme Categories in DNN Acoustic Models

Classification Using ANN: A Review

Speaker Identification by Comparison of Smart Methods. Abstract

GACE Computer Science Assessment Test at a Glance

Probabilistic Latent Semantic Analysis

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Modeling function word errors in DNN-HMM based LVCSR systems

(Sub)Gradient Descent

Ordered Incremental Training with Genetic Algorithms

Evolution of Symbolisation in Chimpanzees and Neural Nets

Australian Journal of Basic and Applied Sciences

Rule Learning with Negation: Issues Regarding Effectiveness

Development of an IT Curriculum. Dr. Jochen Koubek Humboldt-Universität zu Berlin Technische Universität Berlin 2008

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

arxiv: v1 [cs.cv] 10 May 2017

Softprop: Softmax Neural Network Backpropagation Learning

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

AUTOMATED FABRIC DEFECT INSPECTION: A SURVEY OF CLASSIFIERS

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Lecture 1: Machine Learning Basics

A Comparison of Two Text Representations for Sentiment Analysis

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

16.1 Lesson: Putting it into practice - isikhnas

AQUA: An Ontology-Driven Question Answering System

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Learning to Schedule Straight-Line Code

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Speech Recognition at ICSI: Broadcast News and beyond

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Speech Emotion Recognition Using Support Vector Machine

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Generative models and adversarial training

Computed Expert System of Support Technology Tests in the Process of Investment Casting Elements of Aircraft Engines

Introduction to Modeling and Simulation. Conceptual Modeling. OSMAN BALCI Professor

Automatic Speaker Recognition: Modelling, Feature Extraction and Effects of Clinical Environment

A Case Study: News Classification Based on Term Frequency

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

A Pipelined Approach for Iterative Software Process Model

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

arxiv: v1 [cs.lg] 15 Jun 2015

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

ME 443/643 Design Techniques in Mechanical Engineering. Lecture 1: Introduction

Infrared Paper Dryer Control Scheme

Second Exam: Natural Language Parsing with Neural Networks

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Henry Tirri* Petri Myllymgki

Knowledge Transfer in Deep Convolutional Neural Nets

Reinforcement Learning by Comparing Immediate Reward

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Automating the E-learning Personalization

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Device Independence and Extensibility in Gesture Recognition

Mathematics subject curriculum

LEGO MINDSTORMS Education EV3 Coding Activities

Time series prediction

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

An empirical study of learning speed in backpropagation

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Operational Knowledge Management: a way to manage competence

Issues in the Mining of Heart Failure Datasets

Radius STEM Readiness TM

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Data Fusion Models in WSNs: Comparison and Analysis

***** Article in press in Neural Networks ***** BOTTOM-UP LEARNING OF EXPLICIT KNOWLEDGE USING A BAYESIAN ALGORITHM AND A NEW HEBBIAN LEARNING RULE

Integrating simulation into the engineering curriculum: a case study

An Introduction to Simio for Beginners

Transcription:

ARTMAP NETWORKS FOR CLASSIFICATION OF ULTRASONIC WELD INSPECTION SIGNALS P. Ramuhalli, L. Udpa, S. S. Udpa Dept. of Electrical and Computer Engineering Iowa State University Ames, IA 50011 J. Spanner EPRI NDE Center Charlotte, NC 28262 INTRODUCTION Inverse problems in Nondestructive Evaluation (NDE) involve estimating the characteristics of flaws from measurements obtained during an inspection. Several techniques have been developed over the years for solving the inverse problem [1]. These techniques range from calibration approaches to numerical methods based on integral equations. Signal identification and classification is one of the more popular approaches for inverse problems encountered in many practical NDE applications. A variety of signal classification algorithms have been developed for the analysis of NDE data. A widely used implementation of signal classification systems in recent years is based on the use of neural networks. Neural networks are massively parallel interconnections of simple computing elements called neurons [2]. These networks acquire knowledge through a learning process. The information learned is stored in interneuron connection strengths also known as synaptic weights. A schematic of the overall signal classification strategy is shown in Figure 1. The signal is first mapped onto a lower dimensional vector composed of key discriminatory signal features which is then input to the neural network for classification. The learning process in neural networks takes the form of adapting the synaptic weights using an iterative algorithm such as the backpropagation algorithm [2,3], simulated annealing [2] and self-organization [2,3]. During training in a supervised manner, the network is provided with a set of input patterns and the corresponding target outputs. The network then uses the learning algorithm to adapt its synaptic weights to reflect the mapping between the input and output. Unsupervised training uses only the input patterns and tries to extract discriminatory information through some form of self organization. Review of Progress in Quantitative Nondestructive Evaluation, Vol 17 Edited by D.O. Thompson and D.E. Chimenti, Plenum Press, New York, 1998 751

Input ignal ~, Featun' sl'il'ctiull! I -------" :\l'ural :\d\\urk I ~-.--J ~Cla sification Re ult Figure 1. Overall signal classification scheme. A desirable quality for neural networks to have is the ability to learn in an incremental fashion. This is the property of retaining previously learned knowledge while learning new information. Conventional networks like the MLP are not suited for incremental learning due to the fact that a single set of weights is used to learn all possible mappings. Use of a separate set of weights for each part of the input space provides the ability for incremental learning. Several studies related to incremental learning have been performed. Among these is the design of a modular network by Jacobs et al [4] in which a collection of networks is moderated by means of a gating network. Each module learns only a part of the input space and the cumulative output of the network is gated by the gating network to produce the final decision. The gating network is also responsible for deciding which module gets the current input pattern. Other designs use a number of modules of multilayer perceptrons (MLP), a commonly used network for pattern recognition. These modules are then combined (either through adding their outputs or by making use of a more sophisticated scheme) to produce the final output. Incremental learning can be accomplished by way of including additional modules as patterns from new classes arrive (for example, [5]). While these networks work well in practice, they also suffer from certain inherent disadvantages. Chief among these is the time required for training. These networks are not suited for learning in real time. Moreover, the use of separate modules means that the time for training increases as the number of modules increases. This paper investigates the capability of Adaptive Resonance Theory (ART) Networks for incremental learning. The network is tested using ultrasonic signals obtained during the inspection of welds in nuclear power plant piping. The organization of this paper is as follows. The next section introduces ART networks and a common variant, the ARTMAP. The architecture and learning strategy are discussed. Section 3 describes the databases used for classification followed by results and discussion. Finally, section 4 draws conclusions and suggests issues for future investigation. ADAPTIVE RESONANCE THEORY (ART) NETWORKS Adaptive resonance theory (ART) networks are self organizing neural networks that may be trained in either supervised or unsupervised manner. There are several versions of the ART network. Among them are the ARTl network [6], used for binary input patterns, the ART2 and ART3 networks, used for analog patterns and the Fuzzy ART [7], used with fuzzy systems. All these networks learn in an unsupervised manner. The family of ART networks that are trained in a supervised manner are often referred to as ARTMAP networks. A description of both types of networks follows. 752

j g, 1'2 lu or ~ -J- -t ~~:::;~ Top down - 'Highl T.~ 1:, 1'1 layer Reset - -. Figure 2. Schematic of the ARTl network. The schematic of an ARTl network is shown in Figure 2. The network has two layers of nodes fully interconnected by two sets of directional weights. The bottom layer is called the Fllayer while the top layer is called the F2 layer. The set of interconnection weights from the Fl nodes to the F2 nodes are known as the bottom-up weights while the interconnection weights in the opposite direction are called the top-down weights. The nodes in the F2 layer are used to store the classification information. The top-down and bottom-up weights store the representative patterns (also called the exemplars) for the classes. In addition, a control parameter called the vigilance (p), is provided as shown in Figure 2. The ART algorithm resembles a clustering algorithm. The input pattern is presented at the Fllayer. The activation at the ith node of the output F2layer is computed as (1) where Bi is the bottom-up weights and X is the input pattern and IX is usually set much less than 1. The node with the maximum activation represents the class in which the pattern X will be clustered. However. learning does not take place yet. Once the representative class has been selected, the exemplar for that class is read and presented at the Fl layer as a second set of inputs. This exemplar is then compared with the input pattern. A quantitative measure of the similarity between the exemplar and the input pattern is computed using the expression,tjt(k)nxll IIT(k)11 where T(k) is the top-down weights for node k in the F2 layer. Learning takes place if the stored exemplar and the input pattern are similar (A > p). If the stored exemplar and the input are not similar, the current node is held low and the network searches for a better representation. This cycle is repeated till (a) a suitable node is (2) 753

found or (b) a new category is formed with an unused node or (c) the network runs out of nodes. In the first two cases, the network weights are adapted using B(k)=B(k)nX; T(k)=T(k)nX. (3) In the last case described by (c) above, the network starts over again with a new pattern. The expressions shown above describe the ARTl network for binary patterns. The Fuzzy ART [7] network operation is also similar to the one described above. The only changes are in the expressions where the intersection operation (n) is replaced by the fuzzy intersection operation (represented by 1\.). The fuzzy intersection operation between two numbers selects the minimum of the two numbers. Use of the fuzzy intersection operation requires that the numbers be normalized to the range [0,1]. The ART networks described above are unsupervised learning networks. However, many pattern recognition applications require supervised learning particularly when we need to associate a name or a class with a pattern. For such applications, the family of supervised learning ART networks (the ARTMAP [8]) is used. A schematic of the ARTMAP is shown in Figure 3. Two ART modules are connected by a layer of nodes known as the map layer. The individual ART modules (ART. and ARTb) may be binary or fuzzy in nature. The map layer is associated with a second vigilance parameter pab. The learning algorithm for the ARTMAP is as follows. (l) Initialize all weights. (2) Present input at ARTa and target output at ARTb. Allow category formation in both modules. The F21ayer nodes in ARTb encode the class information. (3) Get predicted output class information from ARTa by using the map field weights Wab. (4) Compare predicted class with the with actual class information. If they are the same, go to step 6. (5) Reset current representation in ART. and search for a better representation. Repeat steps 4 and 5 until a correct match is found. (6) Update all weights in both modules. (7) Go to step 2. w \1.", h, hl h I - ---- Figure 3. ARTMAP schematic. 754

~ -r-~hc Affected Zone (HAZ) OUtsuie diameter PIPE Inside dujillctcr upst=m FLOW down'tream direction -_:.=:.;.;-. dirccllon Figure 4. Inspection geometry for ultrasonic weld inspection. RESULTS AND DISCUSSIONS The fuzzy ARTMAP network was used for the classification of signals obtained from ultrasonic inspection of welds in nuclear power plants. The inspection geometry is shown in Figure 4. A cross-section of the weld with the three possible classescracks, counterbores and root welds- from which reflections are obtained is shown. Two different databases were used to test the network. DATABASE I The first database consisted of A-scans from three different classes. Two different inspection frequencies, 1 MHz and 2.25 MHz, were used at a sampling frequency of 10 MHz. In both cases, the discrete wavelet transform (DWT) coefficients of the A-scans were used as the features based on which the classification was done. The network performance was recorded for both non-incremental and incremental learning cases. In the first test, the network was provided training signals from all three classes. There were 90 signals in the training database for I MHz signals and 160 for the 2.25 MHz signal database. In both cases, the network learned in one epoch (one presentation of the entire training set is called an epoch). A vigilance value of 0.8 and a equal to 0.001 was used. The performance of the trained network on signals in a test set is shown in Table I. In order to test the incremental learning capability, the network was first trained with signals from cracks and counterbores only. The network was then trained again with signals from rootwelds. (In the case of 1 MHz signals, the network was trained with cracks and rootwelds first, with counterbores added later). Training was accomplished in 2 epochs Table I Fuzzy ARTMAP classification results for database I with non-incremental learning. Inspection Frequency Cracks C ou nterbores Rootwelds Total classification 1 MHz 2.25MHz 100 % 100 % 100 % 100 % (43/43) (68/68) (45/45) (156/156) 64.06 % 62.96 % 95.91 % 73.05 % (41/64) (34/54) (47/49) (131/167) 755

Table 2 Fuzzy ARTMAP results for database I with incrementalleaming. In s pection F req u e n c y C rac k s Counterbore s Rootwelds Total classification I MHz 2.25MH z 100 % 100 % 100 % 100 % (43/43) (68/68) (45 145) (156/156) 75 % 68.5 % 93.8 % 78.44 % (48/64) (37/54) (46/49) (131/167) for each set of training data. The results on a validation data set are shown in Table 2. The value of vigilance used was 0.8 while <X was set at 0.001. The signals used for training are the same as those used in the non-incremental case to provide ease of comparison. The tables clearly demonstrate that the network performs exceedingly well in classifying the 1 MHz signals. DATABASE II Database II consisted of C-scan images obtained by automatic scanning of welds. Again, two inspection frequencies were used (l MHz and 2.25 MHz) but the sampling frequency was 25 MHz. The network was trained with 1200 samples in the 2.25 MHz and 1000 samples in the 1 MHz training databases. The vigilance factor was set at 0.8 and <X was set at 0.001. The results are shown in Figures 5 and 6 in the form of classification images. In each of the two classification images, the light gray region represents the counterbore while the white region is the rootweld. Cracks are indicated as a dark gray region. In both the images, the background is black in color. CONCLUSIONS Results obtained using the fuzzy ARTMAP indicate that the network is capable of providing very accurate classification results. The network can train very fast with very few presentations of the input set. Results for both non-incremental and incrementalleaming are given. The ART algorithm is naturally suited for incrementalleaming. The reasons for the increased classification errors in the case of 2.25 MHz signals is being studied. I t. I 1 1>MfV» ---- Figure 5. Original and Fuzzy ARTMAP generated C-scan images (2.25 MHz). I 756

()IgInoIC-_ ---- eo I I ~ ~ j J J I :to.. ~~Ift. I Figure 6. Original and Fuzzy ARTMAP generated C-scan images (I MHz). Another aspect of neural network based signal classification systems is reliability. Reliability of the network decision can be quantified by a confidence level. This information can in turn be used by operators to take appropriate actions. A low value of reliability may require re-scanning at a different frequency or inspection angle. This work is currently under way. ACKNOWLEDGEMENTS This work was supported by EPRI under contract RP 3148-06. REFERENCES 1. L. Udpa and S. S. Udpa, "Application of signal processing and pattern recognition techniques to inverse problems in non-destructive evaluation", in Int. 1. of Applied Electromagnetics and Mechanics 8 (1997), pp. 99-117. 2. S. Haykin, Neural Networks: A Comprehensive Foundation, Macmillan College Publishing Co., New York, 1994. 3. R. P. Lippmann, "An introduction to computing with neural nets", in IEEE Acoustics, Speech and Signal Processing Magazine, (April 1987). 4. R. A. Jacobs and M. 1. Jordan, "A competitive modular connectionist architecture", in Adv. In Neural Information Processing Systems 3 (1991), pp. 767-773. 5. Y. W. Huang et ai, "Modular neural networks for identification of starches in manufacturing food products", in Biotechnology Progress 9 (1993), pp. 401-410. 6. G. A. Carpenter and S. Grossberg, "A massively parallel architecture for a self organizing neural pattern recognition machine", in Computer Vision, Graphics and Image Processing 37 (1987), pp. 54-115. 7. G. A. Carpenter et al, "Fuzzy ART: Fast stable learning and categorization of analog patterns by an adaptive resonance system", in Neural Networks 4 (1991), pp. 759-771. 8. G. A. Carpenter et ai, "Fuzzy ARTMAP: A neural network architecture for incremental learning of analog multidimensional maps", in IEEE Trans. Neural Networks vol. 3, number 5 (1992), pp. 698-713. 757