INPE São José dos Campos

Similar documents
QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Artificial Neural Networks written examination

Evolutive Neural Net Fuzzy Filtering: Basic Description

Softprop: Softmax Neural Network Backpropagation Learning

Python Machine Learning

Test Effort Estimation Using Neural Network

A Reinforcement Learning Variant for Control Scheduling

Learning Methods for Fuzzy Systems

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

An Automated Data Fusion Process for an Air Defense Scenario

WHEN THERE IS A mismatch between the acoustic

Word Segmentation of Off-line Handwritten Documents

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Human Emotion Recognition From Speech

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Lecture 1: Machine Learning Basics

Speech Recognition at ICSI: Broadcast News and beyond

Generative models and adversarial training

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Rule Learning With Negation: Issues Regarding Effectiveness

Knowledge Transfer in Deep Convolutional Neural Nets

Calibration of Confidence Measures in Speech Recognition

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

SARDNET: A Self-Organizing Feature Map for Sequences

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Modeling function word errors in DNN-HMM based LVCSR systems

Speaker Identification by Comparison of Smart Methods. Abstract

Classification Using ANN: A Review

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

Rule Learning with Negation: Issues Regarding Effectiveness

CS Machine Learning

UNIVERSITY OF CALIFORNIA SANTA CRUZ TOWARDS A UNIVERSAL PARAMETRIC PLAYER MODEL

Soft Computing based Learning for Cognitive Radio

Lecture 10: Reinforcement Learning

Time series prediction

On the Combined Behavior of Autonomous Resource Management Agents

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Modeling function word errors in DNN-HMM based LVCSR systems

A student diagnosing and evaluation system for laboratory-based academic exercises

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

Issues in the Mining of Heart Failure Datasets

A study of speaker adaptation for DNN-based speech synthesis

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Speech Emotion Recognition Using Support Vector Machine

A deep architecture for non-projective dependency parsing

Mandarin Lexical Tone Recognition: The Gating Paradigm

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

CSL465/603 - Machine Learning

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method

On the Formation of Phoneme Categories in DNN Acoustic Models

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Learning to Schedule Straight-Line Code

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Axiom 2013 Team Description Paper

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Moderator: Gary Weckman Ohio University USA

Software Maintenance

An empirical study of learning speed in backpropagation

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Handling Concept Drifts Using Dynamic Selection of Classifiers

Evolution of Symbolisation in Chimpanzees and Neural Nets

Beyond the Pipeline: Discrete Optimization in NLP

XXII BrainStorming Day

(Sub)Gradient Descent

Artificial Neural Networks

Strategy for teaching communication skills in dentistry

Analysis of Enzyme Kinetic Data

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Semi-Supervised Face Detection

Henry Tirri* Petri Myllymgki

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

Deep Neural Network Language Models

arxiv: v2 [cs.cv] 30 Mar 2017

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Seminar - Organic Computing

A Model to Detect Problems on Scrum-based Software Development Projects

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Using focal point learning to improve human machine tacit coordination

TD(λ) and Q-Learning Based Ludo Players

Intervening to alleviate word-finding difficulties in children: case series data and a computational modelling foundation

European Cooperation in the field of Scientific and Technical Research - COST - Brussels, 24 May 2013 COST 024/13

A SURVEY OF FUZZY COGNITIVE MAP LEARNING METHODS

A Pipelined Approach for Iterative Software Process Model

Transcription:

INPE-5479 PRE/1778 MONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND COVER CLASSIFICATION IN A NEDRAL NETWORK ENVIRONNENT Maria Suelena S. Barros Valter Rodrigues INPE São José dos Campos 1993

SECRETARIA DA CIÊNCIA E TECNOLOGIA INSTITUTO NACIONAL DE PESQUISAS ESPACIAIS INPE-5479-PRE/1778 MONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND-COVER CLASSIFICATION IN A NEURAL NETWORK ENVIRONMENT Maria Suelena S. Barros Valter Rodrigues Paper accepted for presentation at the 29th Plenary Meeting of the Committee on Space Research-COSPAR/The World Space Congress, Washington, DC, 1992, and accepted for publication in Advances in Space Research. INPE São José dos Campos 1993

CDU: 528.711.7:621.376.5 KEY WORDS: Neural Network; Land-Cover Classification; Satellite Imagery.

NONLINEAR ASPECTS OF DATA INTEGRATION FOR LAND-COVER CLASSIFICATION IN A NEURAL NETWORK ENVIRONMENT Maria Suelena Santiago Barros and Valter Rodrigues National Institute for Space Research - INPE. PO BOX 515, 12 201-907 - Sao Jose dos Campos - SP- Brazil ABSTRACT Some results of exploring nonlinear aspects of a neural network methodology to provide land-cover in satellite imagery are presented. All required images are used in a Back-Error Propagation (BEP) network which is a nonlinear data integrator for spatial patterns classification. The network is trained to give the basic categories: grass, moisted sou, bare sou, forest, water and built-up areas. The results of a partial classification are used in a posterior analysis which is done to get the final classification in more detailed classes of land use. The performance results show how powerful is a neural-network based methodology for sattelite imagery integration and classification.

INTRODUCTION Land-cover classification can be considered as a conventional classification process whose goal is to separate data into discrete groups of known identity. Usually in this process techniques of classification are used to generate land-cover classes and sub-classes from sattelite imagery. Much of the earliest conventional classification techniques were based on the probability density function of data belonging to each class /1/. The main problem is to find an efficient classification algorithm which could define bounderies among classes which are often not well separated. In addition, a classification algorithm has to take into account the existence of different textures within each class. In order to get the maximum information from the multivariate data, the classification algorithm should be isotropic, in the sense that each weighted source data is equally accounted. For pratical results, the classification algorithm has to be robust to the data variability, generating correct pixel classification under noisy data. Statistical classification methods usually assign the most likely class to the observed data. Although optimal theoretical results are obtained for the assumption about probability density functions, in real situations the true probability density functions are different from their theoretical models. The main advantage of the neural network approach for -2-

classification tasks is its distribution free characteristics, besides its potential to weight each data /2/. So, neural networks seem to be a compromise between classification techniques based on statistical methods and heuristics procedures. The learning capabilities of neural network provide a mechanism to assimilate the statistical information of the observed data. On the other hand, the neural network topology could be considered as a synthetic composition of the observed data /3/. In this paper we present some results in land-cover classification experiments using two neural network Back- Propagation models. One of them is based on a "Monolithic" architecture in which ali integrated data are used for a exclusive class activation, and the other is based on "classdistributed" architecture where each neural network is trained to recognize only one class characteristic. DEFINITIONS A neural network consists of nodes called neurons, and weighted links between these neurons simulating synaptic activities. Mathematically speaking, a neural network maps the values of inputs neurons to output neurons. In a formal model the output value is typically computed as some nonlinear bounded function of a weighted sum of activities of the neuron inputs. These inputs are the output values of other neurons. One of the most known neural network model is the Back- Propagation that has been under experiment in land-cover classification problems. The Back-Propagation neural network has three or more processing layers: an input layer, one or more hidden layers and an output layer. -3-

Dl LAVIR INRI uiva Fig. 1. Back-error Propagation Model Each nade has an activity represented by the following equation: oj = f ( Ei wii xi -lj ) (1) where: o is the output value of the node j; f is a nonlinear function, such as a sigmoid: f= 1 / ( 1 +e-x ) (2) wij are the weights between the nades of two layers linked, i and j; x3 is the input value of the nade i; 13 is a threshold. In the learning phase of the Back-Propagation algorithm the way of adjust the weights is according to the following - 4 -

equation: 6w = n 6 f' 1.3 (3) where n is a learning rate, 6 is the difference between the desired ouput of neuron j and its actual output, and f' is the partial derivative of the sigmoid function f(.). For classification, usually a neural network operates as a class identifier which receives a set of input vectors and produces responses at each output unit associated to each class. Monolithic classifier. Back-Propagation networks can form arbitrarily complex decision boundaries to separate very meshed classes /4/. The Monolithic classifier, as shown in figure 2, is a Back- Propagation neural network whose output nodes are associated to each class. An output node is activated every time the input x of the network belongs to the associated class. The output nodes have as activities a weighted function of the same hidden node activities in the previous layer. The decision rule is to select that cia ss corresponding to the output node with the largest output. The supervised learning algorithm specifies, for each possible input, an associated output vector. The function of the learning algorithm is to choose the best values of the weights so the output units give the correct class indication when it is in the classification procedure. In the learning procedure the algorithm consider an exclusive class labeling. This means ali classes are considered in the learning process, but each class is labeled one at a time, and the -5-

same weight set has to adapt itself to ali classes characteristics. Often that limits the neural network performance. 1 hidden layer: 10 nodes channel 3 o o channel 2 channel 1 Fig. 2. Monolithic Classifier Class-distributed classifier. As shown in the figure 3, this type of classifier consists of a set of networks. Each network is specialized in classifying one kind of class. The decision rule is also to select that class corresponding to the output node, or network output, with the largest output. Differently to the Monolithic approach the boundaries to be determined by this classifier are a competition of individual boundaries defined by each network. A class-distributed architecture permits the use of the simplest neural networks for each class learning. It makes easier the learning task, because there is no interference among networks during the learning procedure. -6-

land cover class (1 hidden layer: 6 nodes) NW 4 ÍNW MT/ Fig. 3. Class- distributed classifier EXPERIMENTS The research discussed in this paper concerns the determination of the most appropriate approach to land-cover classification tasks,when there are more than two classes to be identified, in between the previously described. Relative performance was estimated by comparing classification results of the Monolitic approach to the class-distributed, using the same imagery, same learning sites and a learning window of (3 x 3) pixels. This size permitted a fine consideration of texture details in both the training and classification procedures. Both approaches were used to classify a data set consisting -7-

of a Landsat Thematic Mapper imagery (channels 3, 4 and 5). Channels 3, 5 and 7 have been indicated as good information sources for sites visualization when seen separately. In case those sources are considered superposed channels RGB TM 435 are better for urban studies. Each channel comprises an image of (512 x 512) pixels. The area used for classification is Sao Jose dos Campos, Sao _ Paulo State/Brazil. Only six basic classes were considered:grass, moisted sou, bare sou, forest, water and built-up areas. Fig. 4. Landsat image (Channel 5) with the training areas considered. -8-

The picture shown in the figure 4 corresponds to the channel 5 of the considered Landsat image. The neural networks in the two approaches had the following architectures: - Monolitic: 27 input nodes ( three input layers each containing (3x3) input nodes ); one hidden layer with 10 nodes and an output layer with 6 nodes, one for each class. - Class-distributed: Six identical architectures each consisting of 27 input nodes ( three input layers each containing (3x3) input nodes); one hidden layer with 6 nodes and an output layer with one node. RESULTS The training procedure for the monolithic approach stopped at 1500 epochs (each epoch corresponds to a training set). For the class-disstributed approach different numbers of epochs were established for the stopping error condition to each network. In this case, 250 epochs were used for water and bare sou l networks; 500 epochs for forest and 2000 epochs for the others landcover classes. The Back-Propagation learning parameters were: learning rate = 0.8 and momentum = O. Experimentally we have observed that after those epochs, usually the error is less than 25 % of the initial error. The classification performance was measured comparing the networks results to the true class indication which was provided by an expert in the domain, by maps and aerial photographs. - 9 -

In the figure 5 results of Monolitic approach for ali the land-cover classes are shown, and in figure 6, the classdistributed approach CONCLUSIONS The land-cover classification results demonstrate that the Monolithic approach adequately retrieve classes' type better than the class-distributed one. However both of the two approaches are robust in land-cover discrimination, combining spatial and spectral information. In the class-distributed approach the decision process is based on a competition between ali the six class-distributed networks. Some classifying distortions were observed due to different network sensitivity. In addition it was noted that for small number of hidden nodes misclassifications occured in the class-distributed approach. Although observing each class-distributed network activation it is noted a good performance, as shown in the figure 7 for the network specialized for the water class. It was also observed that both monolithic and classdistributed networks are sensitive to the quality and number of data samples. It implies a special care in the data sample for the learning phase. Finally it was identified that neural networks seem to be an efficient tool for incremental learning in different scenarios where a multitude of classes characteristics can be assimilated. ACKNOWLEDGMENTS The authors wish to thanks Prof. Josef Skrzypek from UCLA,

Fig. 5 - Results of monolithic approach.

Fig. 6. Results of class-distributed approach. - 12 -

who permitted to use the UCLA-SFINX (Structure and Function in Neural Connections) simulator to perform all experiments, and Dra. Maria L.N. de O. Kurkdjian for helpful discussions. 6. 1. J. Lee; R.C.Weger; S.K.Sengupta and R.M Welch, "A Neural Network approach to cloud classification," IEEE Trans.on GRS, Vol.28 No.5, pp. 846-855, Sept. 1990. 2. J. A. Benediktsson; P.H. Swain and O.K. Ersoy, 4Neural Network approaches versus statistical methods in classification of multisource remote sensing data, IEEE Trans. on GRS,Vol. 28, No. 4, July 1990, pp. 540-552. 3. R. Hecht-Nielsen, Neurocomputing. Addison-Wesley Publishing Co.,1990. 4. G. F. Hepner; T. Logan; N. Ritter, and N. Bryant, "Artificial Neural Network classification using a minimal training set: comparison to conventional supervised classification," Photogrammetric Engineering and Remote Sensing, Vol. 56, no. 4, pp. 469-473, April 1990. - 13 -