Time Series Prediction by means of GMDH Analogues Complexing and GAME

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Learning Methods for Fuzzy Systems

Python Machine Learning

Artificial Neural Networks written examination

Evolutive Neural Net Fuzzy Filtering: Basic Description

CS Machine Learning

Human Emotion Recognition From Speech

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Lecture 1: Basic Concepts of Machine Learning

INPE São José dos Campos

Learning From the Past with Experiment Databases

(Sub)Gradient Descent

A Case Study: News Classification Based on Term Frequency

Multimedia Application Effective Support of Education

Test Effort Estimation Using Neural Network

Reinforcement Learning by Comparing Immediate Reward

Evolution of Symbolisation in Chimpanzees and Neural Nets

WHEN THERE IS A mismatch between the acoustic

Time series prediction

Learning Methods in Multilingual Speech Recognition

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

A student diagnosing and evaluation system for laboratory-based academic exercises

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Seminar - Organic Computing

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Rule Learning with Negation: Issues Regarding Effectiveness

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

On the Combined Behavior of Autonomous Resource Management Agents

Diploma in Library and Information Science (Part-Time) - SH220

A study of speaker adaptation for DNN-based speech synthesis

Word Segmentation of Off-line Handwritten Documents

Rule Learning With Negation: Issues Regarding Effectiveness

UNIVERSITY OF CALIFORNIA SANTA CRUZ TOWARDS A UNIVERSAL PARAMETRIC PLAYER MODEL

Generative models and adversarial training

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

CSC200: Lecture 4. Allan Borodin

School of Innovative Technologies and Engineering

DICTE PLATFORM: AN INPUT TO COLLABORATION AND KNOWLEDGE SHARING

Softprop: Softmax Neural Network Backpropagation Learning

On-Line Data Analytics

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Introduction to Simulation

Practical Research. Planning and Design. Paul D. Leedy. Jeanne Ellis Ormrod. Upper Saddle River, New Jersey Columbus, Ohio

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Learning to Schedule Straight-Line Code

STA 225: Introductory Statistics (CT)

Australian Journal of Basic and Applied Sciences

Lecture 1: Machine Learning Basics

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Laboratorio di Intelligenza Artificiale e Robotica

Applications of data mining algorithms to analysis of medical data

Knowledge-Based - Systems

Issues in the Mining of Heart Failure Datasets

A Case-Based Approach To Imitation Learning in Robotic Agents

Speech Recognition at ICSI: Broadcast News and beyond

Speaker Identification by Comparison of Smart Methods. Abstract

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Speech Emotion Recognition Using Support Vector Machine

A Neural Network GUI Tested on Text-To-Phoneme Mapping

PRODUCT COMPLEXITY: A NEW MODELLING COURSE IN THE INDUSTRIAL DESIGN PROGRAM AT THE UNIVERSITY OF TWENTE

Mining Association Rules in Student s Assessment Data

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Abstractions and the Brain

Laboratorio di Intelligenza Artificiale e Robotica

SURVIVING ON MARS WITH GEOGEBRA

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Modeling function word errors in DNN-HMM based LVCSR systems

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Circuit Simulators: A Revolutionary E-Learning Platform

SARDNET: A Self-Organizing Feature Map for Sequences

KUBAN STATE UNIVERSITY: DOUBLE-DEGREE MASTER S PROGRAMME INNOVATION FOR THE INSTITUTION ENVIRONMENT

Early Model of Student's Graduation Prediction Based on Neural Network

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

An empirical study of learning speed in backpropagation

GACE Computer Science Assessment Test at a Glance

Artificial Neural Networks

Certified Six Sigma Professionals International Certification Courses in Six Sigma Green Belt

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Using the Artificial Neural Networks for Identification Unknown Person

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

Electronic Reserves: A Centralized Approach to the Scanning Process

Mandarin Lexical Tone Recognition: The Gating Paradigm

Ricopili: Postimputation Module. WCPG Education Day Stephan Ripke / Raymond Walters Toronto, October 2015

Defragmenting Textual Data by Leveraging the Syntactic Structure of the English Language

Constructing a support system for self-learning playing the piano at the beginning stage

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

Reducing Features to Improve Bug Prediction

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Research Design & Analysis Made Easy! Brainstorming Worksheet

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Data Fusion Through Statistical Matching

Transcription:

Time Series Prediction by means of GMDH Analogues Complexing and GAME Josef Bouška, Pavel Kordík Dept. of Computer Science and Engineering, Karlovo nam. 13, 121 35 Praha 2, Czech Rep. {bouskj1,kordikp}@fel.cvut.cz Abstract. For time series prediction we can use either parametric or nonparametric models. In this paper we study properties of both approaches for short and medium term prediction intervals. We compare the accuracy of GMDH Analogues Complexing as typical nonparametric method and the Group of Adaptive Models Evolution (GAME) as a parametric method. In our study, we focus on medical data from Motol hospital in Prague and horticulture data from Hort Research New Zealand. Keywords Inductive modeling, Analogues Complexing GMDH, GAME, Time series prediction. 1 Introduction Prediction of time series determines future values based on measuring previous values of that series. The goal is to predict unknown future values from available data. There are many methods for prediction time series, ranging from statistical methods to neural networks as typical black box methods. Many of these methods are parametric, meaning that some parameters are being adjusted to fit the time series and to estimate future values. In this article we focus on two different methods based on inductive modeling. As a nonparametric method we describe our implementation of well known GMDH Analogues Complexing (AC) inductive method [1, 3, 4]. We compare the performance of the AC method with the performance of the GAME method [5] which is parametric. The comparison is performed on real data sets, first on horticulture data (water consumption of mandarin tree) and second on medical data from hospital (progression CO2 in brain). We use these two data sets, because of their different properties. The Mandarin data demonstrates the periodic behavior (see Figure 1), whereas the Brain data set is strictly aperiodic (see Figure 5) from its nature. The setup of the experiments can be found in the Section 4.1 and 5.1 of this paper. In sections 4 and 5 we are experimenting with short and medium term prediction and the final comparison and evaluation of results are given in Section 7. 2 The Analogues Complexing GMDH Analogues Complexing is non-parametrical algorithm of GMDH. It is a multidimensional pattern search method that can be used for clustering, classifying, and predicting most fuzzy objects. For 278

prediction, for example, it self-selects several similar patterns relative to a given reference pattern and then uses their known continuations to form a prediction for the reference pattern. Fig. 1. The Analoques Complexing algorithm stores samples of historical signal (training set) in a memory. For the purpose of prediction, a reference sample is compared to samples in the memory. The continuations of selected sample(s) form the final prediction. The AC method is frequently applied to short noisy data sets with fuzzy properties and aperiodic behavior. It is possible that for such data sets the AC method has better properties than parametric methods. In this paper we would like to find out, if the AC method is superior to parametric methods (e.g. GAME) when applied to time series with periodic and aperiodic properties. 3 The GAME method for time series prediction The GAME states for the Group of Adaptive Model Evolution. This method is proceeds from GMDH theory specifically the Multilayer Iterative Algorithm (MIA) [2]. The GAME generates models inductively from a data set. The model grows from a minimal form during the learning phase, until the optimal complexity is reached. Starting from the first layer, a special genetic algorithm [7] evolves units in the layer. Units can be of several types differing in the function transferring input signals to their output (linear, polynomial, sigmoid, etc.). The most successful units (fitness is computed using an external criterion e.g. performance on the validation set) from the population of the genetic algorithm are frozen and form the first layer of the GAME model. The method proceeds with next layers, evolving units which are most suitable for the given data set. The resulting model consists of heterogeneous units effectively interconnected in a feedforward network manner (see Figure 2). The GAME method is primary designed for regression and classification tasks. The prediction is not so straightforward. A training data set has to be prepared from historical signal using the Sliding window approach [6]. This training data set is used to evolve GAME models capable of single value prediction (model has just single output). To predict more values, it is necessary to evolve more models with the same inputs and different outputs. 279

The Figure 3 shows how the GAME model can be evolved and used for prediction of single future value (t+1). Input variables...... First layer of units 2 nd layer of units x 1 x 2... x n n i= 1 Line a r unit y = aixi + an + 1 Output variable Units in layer evolved by genetic algorithm x 1 x 2... x n y n m r a x i j + i= 1 j= 1 = Polynomia l unit a Fig. 2. An example of GAME model consisting of heterogeneous units interconnected into a feedforward network. Training data Time series signal Prediction Trained GAME model Evolution of GAME model Sliding window Evolution of GAME model Training Training Testing data Fig.3. The scheme of time series prediction by means of GAME model. The model is first evolved using training data prepared from the time series using the sliding window approach, and then the prediction of the model is evaluated on testing data set. 4 Setup of experiments We designed the experiments to find out whether the AC method predicts better than parametric models evolved by the GAME method. The experiments were performed on two different time series data sets one with periodic behavior and the second with aperiodic development of signal. 28

We used the implementation of the AC method implemented by Radek Pinc [1] and our open source environment FAKE GAME [8] allowing building GAME models. At first, we used the same training sample for both methods and compared their prediction on a testing sample. Then we compared the performance of methods on several testing samples to get more accurate results. 5 Results on Mandarin data (periodic behavior) The Mandarin data set (periodic time series) was provided by Hort Research Company, New Zealand. It contains measurements of mandarin tree water consumption. During a day, a tree consumes much more water than during the night (see Figure 4). Fig. 4. Training set of mandarin data for prediction in both of methods The training data set containing 5 measurements was used for both the GAME and the AC methods. The data set for testing contained 12 subsequent measurements. For medium and longer term prediction, we built several GAME models, each trained to predict the signal at certain time horizon. The differences of model output and the target values were measured as the RMS error, which can be computed as: RMS = 1 ( y d ) 2, (1) n where n is number of target values, y is real value from the testing set and d is a predicted value. Tab.1. RMS error of prediction by GAME method on mandarin data, on the sample 51 GAME Index Real data prediction ( y d ) 2 prediction ( y d ) 2 51,273,227599,2612,124417,2277 53,534,595755,38137,2188,144528 55,7295,653557,57673,4572,74245 57,928,873399,29813,617773,96245 59,926,861338,41811,663593,688572 61,877,658491,477461,77542,13184 63,916,86168,31172,837687,61329 65,743,642948,113,83242,79959 67,558,47969,75744,81892,68792 69,38,391431,6966,644293,113932 AC RMS =,97637 RMS =,239591 281

The Table 1 1 explains, how the RMS error was computed for both GAME and AC methods and how these errors can be compared. 5.1 Setup of experiments The GAME models had 3 historical values in their inputs (see Fig. 3). It means that the prediction by GAME is based on window of 3 reference samples in testing set and 2 future unknown samples which are predicted (for example: on the Table 1, there is a prediction of 2 future values from sample 51, and for inputs the reference samples 3-5 are used). The same at the Figure 5 thick blue line 16-46 are the reference samples on the input, and outputs of models are samples 47-67 red line. The thin blue line from sample 46 connects target values. Fig. 5. Prediction of 2 future values with GAME on mandarin data on sample 46 In setup of GAME are used linear and polynomial neurons and perceptrons. As training method the Quasi Newton method was preselected. The prediction by GMDH Analogues Complexing is based on the range 15 3 historical values. This method tries to predict the whole rest of the testing set (see Figure 6), but for the purpose of our experiment only 2 future values are used for comparison with the GAME method. In the setup of the GMDH AC method, we used the Normalization of samples and the Euclid metric for selection of the best samples from the training set (the most similar to the samples in reference window of the testing set). The prediction displayed on the Figure 6 shows that the AC method has in some parts of the time series tendency to delay the real signal. This shift of signal significantly contributes to the RMS error and therefore the results of the AC method are much worse than that of the GAME method. However it is clear that we cannot make a conclusion from one specific observation, therefore we performed additional experiment described in later sections to verify this preliminary result. 1 There are only odd indexes in tables, because of constrained number of out values from GAME and effort to get longer time of prediction. The odd indexes in table of GMDH are there for better comparison with results of GAME. 282

Fig. 6. Medium-term prediction with GMDH Analog Complexing on mandarin data. 6 Results on Brain injury data (aperiodic behaviour) In this data set, we used the same methodology as for the Mandarin data set. The training sequence displayed on the Figure 7 clearly demonstrates the aperiodic properties of the time series. Fig. 7. Training set of medical data for prediction in both of methods 6.1 Setup of experiments In the experiments on aperiodic data training set of 9 samples is used (Fig. 7). The setup of GAME and GMDH method is the same as the setup for the Mandarin data set in the previous section. At first, we performed the same experiments as in the previous section the prediction of 2 future values. The testing set consisted of 1 samples continuing from training set. Results for the GAME and the AC method are displayed on Figures 8 and 9 respectively. The RMS error of the GAME model computed from prediction of 2 future values was,1949. For the GMDH AC method, the error was approximately twice as high (,38889). Again, more confident results are given in the next section. The Figure 1 demonstrates that the AC method is capable of medium and long term prediction. However the error of the prediction tends to increase with longer prediction horizon. We will verify this assumption in the next section. 283

Fig. 8. Prediction of 2 future values with GAME on Brain Injury data, on sample 31 Fig. 9. Prediction with GMDH Analog Complexing on Brain Injury data, on sample 47 Fig. 1. Example of medium-term prediction with GMDH Analog Complexing 284

7 Evaluation of methods and results discussion The Table 2 summaries average errors of prediction on the Mandarin and the Brain Injury testing data for several different testing sets. A sliding window method was used to generate 2 testing sets and reference sets from testing data. The same results are depicted on the Figure 11. Tab.2. Comparison of RMS error of prediction 2 future values by GAME and GMDH (averaged from 2 measurements for different testing data sets) RMS error Mandarin data Brain injury data GAME,73994,38623 GMDH,158162,743 The difference is evident. The prediction on periodic data using GAME method has better results than GMDH AC. The big RMS error of GMDH method is caused by translation predicted from real values along time axis, but the prediction using GMDH is easy for long term prediction. The RMS errors on Mandarin data and the Brain injury data cannot be compared. Each data set has different properties and therefore RMS errors can be used just to compare the performance of individual methods on single data set. RMS Mandarin data (periodic behavior) RMS.12 Brain injury data (aperiodic behavior).25.1.2.8.15.6.1.4.5.2 GAME AC GAME AC Fig. 11. The RMS errors box plot: prediction by GAME and GMDH AC on Mandarin and Brain Injury data collected on 2 different testing sets. 285

We were interested, if there is a relationship between the error of the prediction and the time horizon (how many time steps to the future we are predicting). The Figure 12 shows that for the Analogues Complexing method, the error naturally increases with the distance of the target value in the future. Also the dispersion of errors increases. This behavior is apparent for both Mandarin and Brain injury data sets. Very interesting are the results of the GAME method. For the Brain injury data, the error of the prediction stays on the same level even for the prediction of target value 1 time steps in future. Even more surprising is the decreasing trend of the errors dispersion. We will investigate this behavior in the future. For the Mandarin data set the behavior is similar except that the trend of errors dispersion is the opposite (that is more natural). Brain injury data set (aperiodic) Mandarin data set (periodic) RMS error.12.1.8.6 GAME RMS error.3.25.2.15 GAME.4.1.2.5 Prediction t + 1 2 3 4 5 6 7 8 9 1 Prediction t + 1 2 3 4 5 6 7 8 9 1 RMS error.5.4.3.2.1 AC Prediction t + 1 2 3 4 5 6 7 8 9 1 Prediction t + 1 2 3 4 5 6 7 8 9 1 RMS error.35.3.25.2.15.1.5 AC Fig. 12. Average, maximal and minimal RMS errors for 2 testing samples and different prediction intervals. The Analogues Complexing GMDH method has similar behavior for both aperiodic and periodic data. The average and the dispersion of errors have growing trends with the time distance predicted to future. 8 Conclusion In this paper we were looking for the answer to the question if parametric method (GAME) is better than non parametric method (GMDH AC) for the prediction of time series. We experimented with both periodic and aperiodic time series. Our results showed that the GAME models are superior to the AC method for both periodic and aperiodic time series (GAME is twice as accurate as GMDH AC). For the short-term and medium-term prediction the GAME is more accurate than the GMDH AC. The questionable remains the prediction for longer time, which has not been tested yet. The performance of the GAME method is quite promising for longer time horizon, because the error of this method is not increasing much with the distance of the prediction (see Figure 12). This will be subject of our further research. 9 Acknowledgements We would like to thank to Dr. Richard Brzezny from Dept. of Neurosurgery, The Motol University Hospital in Prague, Czech Republic for the Brain Injury data set and to Dr. Phil Prendergast from Hort 286

Research company, Kerikeri, New Zealand for the Mandarin data set. Thanks to Radek Pinc for his implementation of the Analoques Complexing GMDH method. This research is partially supported by the internal grant of the Czech Technical University in Prague (CTU715313), by the grant Automated Knowledge Extraction (KJB212171) of the Grant Agency of the Academy of Science of the Czech Republic and the research program "Transdisciplinary Research in the Area of Biomedical Engineering II" (MSM6847712) sponsored by the Ministry of Education, Youth and Sports of the Czech Republic. References [1] Pinc R.: Diploma thesis: Implementation of GMDH Analog Complexing, 25 (in Czech language) [2] The short description of the Analoques Complexing Algorithm (AC GMDH) online at http://www.gmdh.net/gmdh_ana.htm [3] The short description of the Multi-layered Iterative Algorithm (MIA GMDH) online at http://www.gmdh.net/gmdh_mia.htm [4] Ivakhnenko,A.G. An Inductive Sorting Method for the Forecasting of Multidimensional Random Processes and Events with the Help of Analogues Forecast Complexing. Pattern Recognition and Image Analysis, 1991, vol.1, no.1, pp.99-18. [5] Kordík P.: GAME - Group of Adaptive Models Evolution, dissertation thesis proposal DCSE- DTP-25-7, FEE, CTU Prague, 25 [6] Chu, C.J.: Times series segmentation: a sliding window approach, Information Sciences Informatics and Computer Science: An International Journal archive 85, p.147-173, 1995. [7] Mahfoud, S.W.: Niching Methods for Genetic Algorithms (951), Technical report, Illinois Genetic Algorithms Laboratory (IlliGaL), University of Ilinios at Urbana-Champaign, 1995 [8] The FAKE GAME environment for the automatic knowledge extraction, available online at: http://neuron.felk.cvut.cz/game 287