Meta-Learning with Backpropagation

Similar documents
System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Learning Methods for Fuzzy Systems

Python Machine Learning

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Evolutive Neural Net Fuzzy Filtering: Basic Description

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Artificial Neural Networks written examination

A Reinforcement Learning Variant for Control Scheduling

Knowledge Transfer in Deep Convolutional Neural Nets

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

SARDNET: A Self-Organizing Feature Map for Sequences

An empirical study of learning speed in backpropagation

Evolution of Symbolisation in Chimpanzees and Neural Nets

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Softprop: Softmax Neural Network Backpropagation Learning

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Reinforcement Learning by Comparing Immediate Reward

Artificial Neural Networks

Modeling function word errors in DNN-HMM based LVCSR systems

arxiv: v1 [cs.lg] 15 Jun 2015

Axiom 2013 Team Description Paper

Lecture 1: Machine Learning Basics

Human Emotion Recognition From Speech

Modeling function word errors in DNN-HMM based LVCSR systems

Visual CP Representation of Knowledge

Knowledge-Based - Systems

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Test Effort Estimation Using Neural Network

Assignment 1: Predicting Amazon Review Ratings

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Rule Learning With Negation: Issues Regarding Effectiveness

Soft Computing based Learning for Cognitive Radio

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Guru: A Computer Tutor that Models Expert Human Tutors

Word Segmentation of Off-line Handwritten Documents

A study of speaker adaptation for DNN-based speech synthesis

Second Exam: Natural Language Parsing with Neural Networks

Lecture 1: Basic Concepts of Machine Learning

Using focal point learning to improve human machine tacit coordination

Learning to Schedule Straight-Line Code

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

The dilemma of Saussurean communication

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

A Case Study: News Classification Based on Term Frequency

A Case-Based Approach To Imitation Learning in Robotic Agents

Education: Integrating Parallel and Distributed Computing in Computer Science Curricula

LEGO MINDSTORMS Education EV3 Coding Activities

INPE São José dos Campos

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

arxiv: v1 [cs.cv] 10 May 2017

Radius STEM Readiness TM

Data Fusion Models in WSNs: Comparison and Analysis

Georgetown University at TREC 2017 Dynamic Domain Track

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

(Sub)Gradient Descent

Physics 270: Experimental Physics

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

Mathematics process categories

On the Combined Behavior of Autonomous Resource Management Agents

WHEN THERE IS A mismatch between the acoustic

Seminar - Organic Computing

Xinyu Tang. Education. Research Interests. Honors and Awards. Professional Experience

Forget catastrophic forgetting: AI that learns after deployment

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Agent-Based Software Engineering

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

How People Learn Physics

Speaker Identification by Comparison of Smart Methods. Abstract

ICTCM 28th International Conference on Technology in Collegiate Mathematics

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

On the Formation of Phoneme Categories in DNN Acoustic Models

Firms and Markets Saturdays Summer I 2014

COMPUTER-AIDED DESIGN TOOLS THAT ADAPT

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Rule Learning with Negation: Issues Regarding Effectiveness

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

CSL465/603 - Machine Learning

Probabilistic Latent Semantic Analysis

Time series prediction

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Laboratorio di Intelligenza Artificiale e Robotica

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

Increasing the Learning Potential from Events: Case studies

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

This scope and sequence assumes 160 days for instruction, divided among 15 units.

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Deep Neural Network Language Models

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

TD(λ) and Q-Learning Based Ludo Players

Extending Place Value with Whole Numbers to 1,000,000

Syntactic systematicity in sentence processing with a recurrent self-organizing network

Transcription:

Meta-Learning with Backpropagation A. Steven Younger Sepp Hochreiter Peter R. Conwell University of Colorado University of Colorado Westminster College Computer Science Computer Science Physics Department Boulder, CO 80309 USA Boulder, CO 80309 USA Salt Lake City, UT 84105 USA syounger 0 boulder.net hochreit @cs.colorado.edu conwellp @ xmission.com Abstract This paper introduces gradient descent methods applied to meta-leaming (leaming how to leam) in Neural Networks. Meta-leaning has been of interest in the machine leaming field for decades because of its appealing applications to intelligent agents, non-stationary time series, autonomous robots, and improved leaming algorithms. Many previous neural network-based approaches toward meta-leaming have been based on evolutionary methods. We show how to use gradient descent for meta-leaming in recurrent neural networks. Based on previous work on Fixed- Weight Leaming Neural Networks, we hypothesize that any recurrent network topology and its corresponding leaming algorithm(s) is a potential meta-leaming system. We tested several recurrent neural network topologies and their corresponding forms of Backpropagation for their ability to meta-leam. One of our systems, based on the Long Short-Term Memory neural network developed a leaming algorithm that could leam any two-dimensional quadratic function (from a set of such functions} after only 30 training examples. 1 Introduction This paper reports on our work utilizing gradient descent methods (i.e. Backpropagation) to search out and find learning algorithms tailored to specific learning tasks (meta-learning). After a brief review previous meta-learning systems, we will discuss Fixed-Weight Learning Neural Networks, which motivates our method. We will also review the Long-Short Term Memory Network. Section 3 describes our metalearning evaluation experimental set-up. In Sec- tion 4, we summarize our results. Finally, we will discuss some of the questions raised by our work. 2 Previous Work In meta-learning, there are two learning processes proceeding simultaneously. There is a supervisory system, which is attempting to learn a good learning algorithm for a set of problems with similar characteristics. There is also a subordinate learning algorithm, which is attempting to learn a specific problem. Periodically, the supervisor alters the subordinate algorithm slightly to improve its learning performance. Mostly, these two algorithms must perform the same task: they must leverage the regularities of their respective problems in order to efficiently solve them. However, there are differences in the time scale and scope of their problems. The supervisory process has a broader scope. It must ignore the details unique to.specific problems, and look for symmetries over a long time scale, while the opposite is true for a subordinate learning scheme. 2.1 Review of Meta-Learning Several researchers have used meta-learning techniques to derive or improve learning algorithms [ 1,2,3]. For example, Runarsson and Jonsson in [2] used a genetic algorithm to evolve neural networks that implemented sophisticated learning rules. Some conclusions of the study were that the evolved networks are fast learners; and the derived learning rule is biased, i.e. it is tuned to solve a given problem class fast. 0-7803-7044-9/01/$10.00 02001 IEEE 200 1

The self-modifying neural networks of Schmidhuber et al. [3], which run their own learning algorithms, are similar to our meta-learning method. Unlike our networks, their networks required special units to read and modify their synaptic weights during learning. 2.2 Fixed-Weight Learning Neural Networks For lalrge networks, genetic-based meta-learning can become intractable due to the number of computations required. We used Fixed-Weight Learning Neural Networks (FWNNs) [4-71 to motivate how to use of gradient descent to speed up meta-learning. FWNNs are recurrent networks that have a learning algorithm encoded or wired into their synaptic weights. Recurrent signal loops store information about the particular mapping being learned by the network. Thus, they can learn without changing any of their synaptic weights. Figure 1 illustrates the conceptual steps involved in converting a single synapse neural network and its attendant learning algorithm into an equivalent FWNN. (This example ignores certain timing issues that change the details of the conversion, but not the overall concept.). We will use the term embedded leaming algorithm to refer to a learning algorithm encoded in synaptic weights. FWNNs move the adaptation associated with learning a particular mapping to the dynamics of the networks. The adaptation is manifest in the changing signals in the recurrent loops. On the other hand, the weights in a FWNN network represent the learning algorithm. Since the networks output error is continuous with respect to changes in the synaptic weights, gradient decent applied to these weights is meta learning. The new idea we bring with this paper is that any recurrent network can be considered a potential fixed weight learning network. In other words, a recurrent network with random weights is simply a very inefficient learning machine. By applying standard gradient decent to these synaptic weights we improve the embedded learning algorithm associated with these weights. Furthermore we can perform meta-learning without any modifications to the training algorithm(s) normally used for that network. A fully recurrent network trained with the Williams and Zipser algorithm [8] can, in principle, be used for meta- learning. However, the training set must include exemplars from many different types of functional mappings. We have found certain recurrent architectures to be better than others at meta learning. One architecture in particular is the Long Short-Term Memory (LSTM). 2.3 The Long Short-Term Memory Network The LSTM Network [9] is a type of recurrent network that was designed to overcome the problems that appear when trying to learn to store information long time intervals. In addition to standard neurons, the LSTM has special memory cells, shown in Figure 3. The memory cells consist of three main components: a self-recurrent linear neuron, input and output gate units controlled by gatekeeper neurons, and a non-linear output squash unit. A LSTM can have either one or two hidden layers. Neurons within each layer are fully interconnected. A special LSTM Truncated Backpropagation is used to train the network. We included the LSTM in our study because Shitoot [lo] noticed strong similarities between the LSTM and the FWNNs reported in VI. 3 The Key to Meta-Learning: Preparing the Meta-Training Data Set The selection of the training data is what determines the difference between regular (non meta-) learning and meta-learning. Regular learning uses several examples of inputs and the associated target outputs from a single functional mapping. For meta-learning, we need many training pairs from many different functional mappings from a given set of such mappings. We will illustrate by giving a specific example: the set or class of all Boolean mappings with two arguments and one result. This set of sixteen functions includes the standard AND, OR, and XOR. Our training corpus consisted of 100 instances the Boolean maps, selected in random order. For each instance of a Boolean map, there was a sequence of 256 randomly generated training vectors. During each training cycle, we presented one of these vectors to the network. A training epoch consisted of a complete pass through all 25,600 vectors. Each training vector also contains the target output for the inputs of the current cycle. However, this value was only used to maintain a running tally of the mean 2002

squared error. The tally was used by the supervisory program for the meta-learning. The FWN s embedded learning algorithm needed a supplementary input so it can learn the presented mapping. We could have used the error of the network s output associated with the previous cycle s input vector. Another possible supplementary input was the target (i.e. the function s result) associated with the previous cycle s input vector. We choose the latter approach. Thus, each training vector had three input values: the two Boolean arguments, and the target for the previous training cycle. 4 Experimental Results In [l 13 we detailed results of our experiments with gradient-based meta-learning. We evaluated several different recurrent network topologies (with their corresponding versions of Backpropagation) for their meta-learning ability. After meta-training, we evaluated the resulting learning networks on separately generated test data. We tested the potential meta-learning topologies and algorithms on three sets of functional mappings. The first is the set of Boolean mappings described above. The second was a set of semilinear function mappings given by the expression y=f.(l+tanh(w,.x, +w2 x, +w,)), where x,, w, E [-1,1]. The x s are the network inputs, they is the target function value. The w s parameterize or specify the particular mapping. This is the set of all mappings that a single neuron with one bias and two inputs can learn exactly (with weights in the range [-1,+1]). The third set of mappings was the set of twoparameter quadratic functions given by: y = ax: + bx: + cx,x, + dx, +ex, + f, where a,..., f E [-1,1] parameterize the particular mapping. The x s were as above, and the y was scaled to the interval [0.2,0.8] before being used as the target value. The only fully successful topology in metalearning was the LSTM neural network and its associated LSTM Backpropagation. The LSTM meta-learning could successfully derive a learning network for all mapping sets attempted. We used two versions for the LSTM, the standard three-layer version, and a modified four-layer version. The latter was required to derive a learning network for the quadratic problem set. Table 1 summarizes the LSTM results. The first column shows the structure of the hidden layers. The first LSTM had one hidden layer with six memory cells and six standard neurons. The second meta-learning network had 12 memory cells and 6 regular neurons in its first hidden layer. It also had a second hidden layer with 40 standard neurons. The second column is the set of mappings that were to be meta-learned. The third column is the number of examples for each mapping s presentation sequence. The fourth column is the number of epochs that the meta-learning program required to derive the learning algorithm. The fifth column is the Mean Squared Error on test data after meta-training has occurred. The final column is the average number of steps that the derived learning algorithm required to converge. Figure 3 shows a plot of absolute error versus time, after meta-learning was successful. The plot is for the Boolean set of functional mappings. The peaks at 512, 768, and 1024 indicate large error when a new mapping begins. Note that the error rapidly reduced after each change, indicating that the learning network performed successfully. Note that the resulting learning networks were rapid learners. The Boolean learning network, for instance, took only about ten steps to learn a new mapping - including XOR and NOT XOR. The most important aspect of our work was that effective learning networks were automatically derived by the LSTM meta-training, not the specific learning networks that were generated. 5 Discussion Why could the LSTM -meta-learn while other architectures could not? We believe that there were two necessary features. We showed in [ 111 that the recurrent loop-back synaptic weights must be 1.0 and the neuron must have a linear squashing function into store information longterm. (Actually, the constraint is slightly less restrictive than this.) We also showed this experimentally in [7]. The second necessary feature was the input gatekeeper units, which control the input to the loop cell. By learning when to allow and (perhaps more importantly) when to disallow new information into the memory cell, the 2003

LSTM can store information for the longer periods of time needed to do meta-learning. The l units could be replaced by an equivalent (standard neuron) network at the expense of more complexity. How did the resultant learning networks work? Were they similar to known methods? It is very difficult to take apart a neural network (especially a recurrent network) and extract the rules that are encoded in its synaptic weights. However, examination of the output of the memory cells revealed that the Boolean problem learner encoded the sixteen possible functions by a fourneuron binary encoding scheme. Obviously, this way of enumerating the mappings would only work for small sets of mappings, each with a small number of possible results (in this case 0 or 1). The meta-learning correctly extracted these properties from the meta-training data set. This is similar to the way a human being may try to solve the problem. Meta-learning on the set of Semi-linear functions resulted in a learning network that stored three continuous values in the memory cells. This reflects the continuous, three-parameter nature of the set of mappings. The Quadratic problem learner also generated continuous values in its memory cells. Another signal it generated was approximately inversely proportional to the cycle step number within a sequence. We believe that the network used this signal to increase the influence of the errors near the beginning of the sequence, speeding up learning. References [ 13 David J. Chalmers, The Evolution of Learning: Experiments in Genetic Connectionism in Proceedings of the I990 Connectionist Models. Summer, School. Editors D.S. Touretsky, J.L. Elman, T.J. Sejnowski & G.E. Hinton, Morgan Kauffmann; San Mateo, CA [2] Thomas Philip Runarsson and Magnus Thor Jonsson. Evolution and Design of Distributed Learning Rules 2000 IEEE Symposium of Com- binations of Evolutionary Computing and Neural Networks. San Antonio, Texas (2000) p. 59 [3] J. Schmidhuber. A neural network that embeds its own meta-levels. In Proc. Of the International Conference on Neural Networks 93, San Fransisco, IEEE 1993 [4] N. E. Cotter and P. R. Conwell. Fixed- Weight Networks Can Learn. In Intemational Joint Conference on Neural Networks held in San Diego 1990, IEEE, New York, 1990, pp. II- 553-559. [5] N. E. Cotter and P. R. Conwell. Learning Algorithms and Fixed Dynamics. In Intemational Joint Conference on Neural Networks held in Seattle 1991 by IEEE. New York: IEEE 1991, I- 799-804. [6] A. Steven Younger, Learning in Fixed- Weight Recurrent Neural Networks. Ph.D. Dissertation, University of Utah 1996 [7] A. Steven Younger, P. R. Conwell, and N. E. Cotter. Fixed-Weight On-Line Learning. IEEE Transactions on Neural Networks. Vol.10 No. 2, March 1999 pp. 272-283 [8] R. J. Williams and D. Zisper, A learning algorithm for continually running fully recurrent neural networks, Univ of California, San Diego, La Jolla, CA. Tech Report TR-8805. [9] Sepp Hochreiter and J. Schmidhuber, Long Short-Term Memory. Neural Computation 9(8) pp. 1735-1780,1997 LSTM source code can be obtained by: ftp ftp.cs.colorado.edu cd users/hochreit/software get hochreiter.lstm.tar.gz [ 101 Yashwant Shitoot, Private Communication, 1996 [ll] Sepp Hochreiter, A. Steven Younger and Peter R. Conwell. Learning To Learn Using Gradient Descent. to appear in Proceedings of the Intemationul Conference on Artificial Neural Networks, Springer Verlag 2001 2004

X 6 a -b Y x o - W 'Y Figure 1 : Construction of an equivalent FW" for a single synapse and its attendant learning algorithm Clockwise from the upper left: (1) Conventional network with learning algorithm neww = f (x, y. 6, ozdw). (2) Universal approximation allows us to replace the learning algorithm with an equivalent recurrent network. Note that recurrence is necessary to store the oldw information dynamically in signal loops. (3) Replace the synapse with a Il unit, removing the requirement to change the synaptic weight. (4) If required, replace the l unit with an equivalent non-il network. I 1.0 I squash neuron output gate memory loop gatekeeper neuron Figure 2: LSTM memory cell. Key features are the input and output gates controlled by gatekeeper neurons, the linear memory loop neuron and the output squash neuron. The gatekeeper neurons learn when to allow data in and out of the memory loop neuron. 2005

Table 1: Performances of Automatically Derived LSTM-Based Learning Networks Hidden Neu- Problem Set Examples per Epochs MSEt Cycles to rons Mapping Learn HI: 6 Memory Boolean 256 800 0.0058 10 + 6 Standard - Semi-Linear 64 10000 0.0008 10 Semi-Linear 1000 5000 0.0025 50 HI: 12 Memory + 6 Standard Quadratic 100 25000 0.00068 35 H2:40 Standard 1 0.8-1 0.6 0.4 0.2 - - - n io0 500 600 700 800 900 1000 1100 Figure 3: Absolute error versus time, after meta-leaming was successful. The plot is for the Boolean set of functional mappings. The peaks at 5 12,768, and 1024 indicate a large error when a new mapping begins. The rapid reduction of the error after the peaks shows that the net mapping was learned quickly. Before meta-learning, this entire plot would have consisted of errors the size of the peaks. 2006