Artificial Neural Networks for Storm Surge Predictions in NC. DHS Summer Research Team

Similar documents
Lecture 1: Machine Learning Basics

Python Machine Learning

(Sub)Gradient Descent

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Artificial Neural Networks written examination

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Evolutive Neural Net Fuzzy Filtering: Basic Description

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Test Effort Estimation Using Neural Network

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Learning Methods for Fuzzy Systems

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

arxiv: v1 [cs.lg] 15 Jun 2015

INPE São José dos Campos

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

CS Machine Learning

Knowledge Transfer in Deep Convolutional Neural Nets

Softprop: Softmax Neural Network Backpropagation Learning

arxiv: v1 [cs.cv] 10 May 2017

Calibration of Confidence Measures in Speech Recognition

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Human Emotion Recognition From Speech

Axiom 2013 Team Description Paper

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Model Ensemble for Click Prediction in Bing Search Ads

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

CSL465/603 - Machine Learning

A study of speaker adaptation for DNN-based speech synthesis

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Artificial Neural Networks

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Attributed Social Network Embedding

Evolution of Symbolisation in Chimpanzees and Neural Nets

Time series prediction

On the Formation of Phoneme Categories in DNN Acoustic Models

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Forget catastrophic forgetting: AI that learns after deployment

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

A Reinforcement Learning Variant for Control Scheduling

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Classification Using ANN: A Review

Deep Neural Network Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

arxiv: v1 [cs.lg] 7 Apr 2015

Assignment 1: Predicting Amazon Review Ratings

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Lecture 10: Reinforcement Learning

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Generative models and adversarial training

Data Fusion Through Statistical Matching

Word Segmentation of Off-line Handwritten Documents

Introduction to Simulation

Software Maintenance

Syntactic systematicity in sentence processing with a recurrent self-organizing network

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Soft Computing based Learning for Cognitive Radio

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

Adaptive Learning in Time-Variant Processes With Application to Wind Power Systems

Issues in the Mining of Heart Failure Datasets

stateorvalue to each variable in a given set. We use p(x = xjy = y) (or p(xjy) as a shorthand) to denote the probability that X = x given Y = y. We al

Lecture 1: Basic Concepts of Machine Learning

Kamaldeep Kaur University School of Information Technology GGS Indraprastha University Delhi

FF+FPG: Guiding a Policy-Gradient Planner

Second Exam: Natural Language Parsing with Neural Networks

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

TD(λ) and Q-Learning Based Ludo Players

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

Applications of data mining algorithms to analysis of medical data

Probabilistic Latent Semantic Analysis

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

Reinforcement Learning by Comparing Immediate Reward

An OO Framework for building Intelligence and Learning properties in Software Agents

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Application of Virtual Instruments (VIs) for an enhanced learning environment

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Device Independence and Extensibility in Gesture Recognition

A SURVEY OF FUZZY COGNITIVE MAP LEARNING METHODS

Modeling function word errors in DNN-HMM based LVCSR systems

Top US Tech Talent for the Top China Tech Company

Moderator: Gary Weckman Ohio University USA

Predicting Future User Actions by Observing Unmodified Applications

Name Class Date. Graphing Proportional Relationships

Using focal point learning to improve human machine tacit coordination

Australian Journal of Basic and Applied Sciences

SARDNET: A Self-Organizing Feature Map for Sequences

Dropout improves Recurrent Neural Networks for Handwriting Recognition

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

A deep architecture for non-projective dependency parsing

Early Model of Student's Graduation Prediction Based on Neural Network

Rule Learning With Negation: Issues Regarding Effectiveness

Hierarchical Linear Modeling with Maximum Likelihood, Restricted Maximum Likelihood, and Fully Bayesian Estimation

Transcription:

Artificial Neural Networks for Storm Surge Predictions in NC DHS Summer Research Team 1

Outline Introduction; Feedforward Artificial Neural Network; Design questions; Implementation; Improvements; Conclusions; 2

Brief Introduction Anton Bezuglov, Ph.D. in Computer Science and Engineering, University of South Carolina, Columbia, 2006 Assoc. Professor of Computer Science at Benedict College Areas of interests: Machine learning, neural networks, algorithms, etc Summer Research Team 2016, sponsored by DHS Artificial Neural Networks for Storm Surge Prediction 3

Brief Introduction, contd. Motivation: accurate method for storm surge prediction; Parametric vs. Nonparametric approaches (Bishop, 2006) Parametric models are computationally expensive; Nonparametric models are cheap, but need training; Problem: need large datasets for training; Synthetic hurricanes; 4

Dataset 324 synthetic hurricanes; 193 samples per hurricane 6 inputs, 10 outputs Inputs: hurricane parameters Outputs: water levels at 10 locations inputs outputs 5

Assumption Based on previous studies; Suppose input -- x(t), output -- y(t), t - time; x(t) contains all information to make predictions y(t) depends on x(t) only y(t) does not depend on x(t-1), y(t-1), etc. 6

Outline Introduction; Feedforward Artificial Neural Network; Design questions; Implementation; Improvements; Conclusions; 7

Regression with a FF ANN Problem: find a function f(.), so that: yp = f(x), yp-- storm surge predictions f(.) -- can be a Feed Forward Artificial Neural Network (FF ANN); Train FF ANN to minimize the error between y and yp Use synthetic storms to train; 8

FF ANN s One hidden layer ANN, two layer model; Information travels from left to right; Nodes are variables (inputs, outputs, and hidden); Edges -- independent parameters; Nonlinear function; Complexity determined by # of multiplications Approx. O(N2), N - # of hidden nodes Backpropagation algorithm 9

Design Questions Architecture? Number of hidden layers? Size of each layer? Choice of nonlinear function? Initial weights/biases? Learning rate? Learning rate decay? Algorithm for training? Clipping gradients? Dealing with overfitting? Loss function? 10

Design Questions, contd. Architecture? -- Two hidden layer multiple outputs Number of hidden layers? -- two hidden layers Size of each layer? -- 16-64 neurons, second layer larger Choice of nonlinear function? -- TanH Initial weights/biases? -- N(0, 0.01) Learning rate? -- 0.001 -- 0.01 Learning rate decay? -- 0.5 Algorithm for training? -- ADAM optimization algorithm Clipping gradients? -- yes, 1.25-1.5 norm Dealing with overfitting? -- validation set, 15% Loss function-- Mean Squared Error (MSE) 11

Design Questions, contd. Stochastic optimization Use portions of the training dataset: batches Training dataset: 228 storms, batches: 19, 57, 114 Or Training dataset: 225 storms, batches: 3, 5, 9, 15, 45, 225 Inputs normalization Inputs vary by 2-3 orders of magnitude Too long to converge Calculate moments for each input param in the training dataset Normalize inputs Store the moments along with the model 12

Design Summary Split dataset into training (70%), validation (15%), and testing (15%); Two hidden layer FF ANN (N1 < N2, less inputs than outputs); Train to minimize MSE; Check for overfitting on the validation dataset; Evaluate performance on the testing dataset; 13

Outline Introduction; Feedforward Artificial Neural Network; Design questions; Implementation; Improvements; Conclusions; 14

Implementation: TensorFlow TensorFlow -- Open Source Library for Machine Intelligence; Algorithms are graphs, nodes -- operations, edges -- tensors x Wh bh tanh W b MSE loss y 15

Implementation: Training and Evaluation x Graph variables can be evaluated/called To train -- call optimizer variable To evaluate -- call loss variable etc. Wh bh tanh W List of graph variables to evaluate b Inputs for placeholders MSE loss y 16

Implementation: Dealing with Gradients Graph variables to evaluate Calculate gradients Clip Apply gradients Evaluate train_op to perform a single train iteration 17

Implementation: Multiple GPU s Each GPU has same graph but individual inputs/outputs; Calculate gradients on each GPU; Average gradient; Apply gradients; Update graphs; 18

Implementation: Restore ANN Save model: weights, biases, and input moments; Train/Run modes; Train -- open file, train ANN, save ANN; Run -- open file, open model, run, save outputs; Train, approx. 1-20 minutes; Run, 0.11 sec (324x193 samples); 19

FF ANN: Performance Two hidden layer FF ANN (32,64) Before and after landfall Landfall only Easy Difficult 20

FF ANN: Performance ADCIRC FF ANN Easy Underpredictions 21

FF ANN: Summary Multi-output ANN: one model for several locations MSE s are approx. 0.006 m^2 CC s are 0.95 ANN has no error before and after the storm surge; Larger errors at storm surge; Low MSE s b/c of zeros; Does y(t) depend on x(t) and something else? Does x(t) miss information? 22

Acknowledgements Many thanks to Brian Blanton, Ph.D. RENCI, CRC DHS SRT Program This research was performed under an appointment to the U.S. Department of Homeland Security (DHS) Science & Technology (S&T) Directorate Office of University Programs Summer Research Team Program for Minority Serving Institutions, administered by the Oak Ridge Institute for Science and Education (ORISE) through an interagency agreement between the U.S. Department of Energy (DOE) and DHS. ORISE is managed by ORAU under DOE contract number DEAC05-06OR23100. All opinions expressed in this presentation are the author s and do not necessarily reflect the policies and views of DHS, DOE or ORAU/ORISE. 23