Connectionist Learning Procedures. Siamak Saliminejad

Similar documents
Artificial Neural Networks written examination

Python Machine Learning

Lecture 10: Reinforcement Learning

Lecture 1: Machine Learning Basics

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Softprop: Softmax Neural Network Backpropagation Learning

(Sub)Gradient Descent

Learning Methods for Fuzzy Systems

Artificial Neural Networks

Learning to Schedule Straight-Line Code

An empirical study of learning speed in backpropagation

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Evolution of Symbolisation in Chimpanzees and Neural Nets

Knowledge-Based - Systems

Human Emotion Recognition From Speech

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Axiom 2013 Team Description Paper

Reinforcement Learning by Comparing Immediate Reward

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

A Reinforcement Learning Variant for Control Scheduling

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Time series prediction

Evolutive Neural Net Fuzzy Filtering: Basic Description

Knowledge Transfer in Deep Convolutional Neural Nets

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

The Good Judgment Project: A large scale test of different methods of combining expert predictions

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

On the Combined Behavior of Autonomous Resource Management Agents

Word Segmentation of Off-line Handwritten Documents

INPE São José dos Campos

A Stochastic Model for the Vocabulary Explosion

Lecture 1: Basic Concepts of Machine Learning

arxiv: v1 [cs.cv] 10 May 2017

A Comparison of Annealing Techniques for Academic Course Scheduling

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Test Effort Estimation Using Neural Network

Speaker Identification by Comparison of Smart Methods. Abstract

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition

WHEN THERE IS A mismatch between the acoustic

Speech Recognition at ICSI: Broadcast News and beyond

phone hidden time phone

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Calibration of Confidence Measures in Speech Recognition

FF+FPG: Guiding a Policy-Gradient Planner

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Neural pattern formation via a competitive Hebbian mechanism

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Generative models and adversarial training

CSL465/603 - Machine Learning

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

arxiv: v1 [cs.lg] 15 Jun 2015

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

How People Learn Physics

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

SARDNET: A Self-Organizing Feature Map for Sequences

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Probabilistic Latent Semantic Analysis

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Issues in the Mining of Heart Failure Datasets

arxiv: v1 [cs.cl] 2 Apr 2017

Soft Computing based Learning for Cognitive Radio

Unsupervised Acoustic Model Training for Simultaneous Lecture Translation in Incremental and Batch Mode

On the Formation of Phoneme Categories in DNN Acoustic Models

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

International Journal of Advanced Networking Applications (IJANA) ISSN No. :

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Neural Representation and Neural Computation. Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990),

Seminar - Organic Computing

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

Using focal point learning to improve human machine tacit coordination

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Laboratorio di Intelligenza Artificiale e Robotica

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

A study of speaker adaptation for DNN-based speech synthesis

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

TD(λ) and Q-Learning Based Ludo Players

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

CS Machine Learning

Firms and Markets Saturdays Summer I 2014

AMULTIAGENT system [1] can be defined as a group of

Mathematics process categories

An OO Framework for building Intelligence and Learning properties in Software Agents

English Language and Applied Linguistics. Module Descriptions 2017/18

THE world surrounding us involves multiple modalities

Corrective Feedback and Persistent Learning for Information Extraction

Introduction to Simulation

Transcription:

Connectionist Learning Procedures Siamak Saliminejad

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Introduction How internal representations can be learned in "connectionist" networks?? First: Resemble the brain more closely than conventional methods Second: Are massively parallel

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Connectionist Models Units: simple, neuron-like processing elements called "units" that interact using weighted connections State (Activity Level): is determined by the input received from other units in the network Knowledge Long-Term Short-Term Changing Weights Add/Remove Connections Temporary weights Threshold

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Connectionist Research Issues Search Representation Local Distributed Supervised Learning Reinforcement Unsupervised

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Associative Memories without Hidden Units No hidden unit The aim is simply to store a set of associations between input & output vectors by modifying the weights. Linear ANs The state of an output unit is a linear function of the total input that it receives from the input units Perfect recall If the input vectors are orthogonal and have length 1 Nonlinear ANs Associations which have nonorthogonal input vectors Deficiencies Most tasks are nonlinear and complex

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Simple Supervised Learning Procedures Input units are directly connected to output units States are a continuous smooth function of their total input Error Surface

Simple Supervised Learning Procedures Nets with linear output units and no hidden unit: always find the minimum Nets with nonlinear output units and monotonic input-output function: aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaalways find the minimum Batch Version Online Version

Simple Supervised Learning Procedures Perceptron Convergance Procedure Initial weights Final weights Ignores magnitude of error. Does not settle down when there is no perfect set of weights ^ Does not work when the idea of ideal region breaks down (Multilayer nets)

Simple Supervised Learning Procedures Deficiencies Most "interesting" cannot be captured by any combination of weights in simple nets Gradient descent may be very slow if the elliptical cross-section of the error surface is very elongated

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Back Propagation Is a generalization of the least squares procedure that works for networks with hidden layers The central idea: derivatives can be computed efficiently by starting with the output layer and working backwards through the layers. Error surface in networks with hidden layers:

Back Propagation Mapping text to speech I am Presenting Discovering Semantic Features Phoneme Recognition

Back Propagation Reinforcement Version First, the mental model learns to predict expected reinforcement Second, derivative of the expected reinforcement can be backpropagated As a Maximum Likelihood Procedure Interpret each output vector as a specification of a conditional probability distribution Minimizing the squared error is equivalent to maximum likelihood estimation if output vectors are treated as the centers of Gaussian pdf. Deficiency Is not adequate for large tasks because learning time scales poorly

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Boltzman Machines Is a generalization of Hopfield network The units update their state according to a stochastic decision rule The units have state 0 or 1 according to the following probability Temp. Total Input If this rule is applied repeatedly, the network will reach thermal equilibrium The simplicity of the Boltzmann distribution leads to a very simple learning procedure which adjusts the weights so as to use the hidden units in an optimal way.

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Maximizing Mutual Information One semisupervised method is to provide it with information about what category the input vector came from Its incoming weights are modified so as to maximize the information that the state of the unit provides about the category of the input vector The derivative of the mutual information is relatively easy to compute and so it can be maximized by gradient ascent

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Unsupervised Hebbian Learning The weight modification depends on both presynaptic and postsynaptic activity It is shown that an unsupervised Hebbian learning procedure in which the weight change depends on the correlation of presynaptic and postsynaptic activity can produce a surprising number of the known properties of the receptive fields of neurons in visual cortex

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Competitive Learning Unsupervised learning Clusters inputs There is a set of hidden units which compete with one another to become active. When an input vector is presented to the network, the hidden unit which receives the greatest total input wins the competition and turns on with an activity level of I. A constraint on each weight vector should be imposed to keep the sum of the weights (or the sum of their squares) constant

Competitive Learning A constraint on each weight vector should be imposed to keep the sum of the weights (or the sum of their squares) constant x 2 2 2 + y + z = 1 Simple Geometric Model

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Reinforcement Learning Procedures We can assign credit to a local decision by measuring how it correlates with the global reinforcement signal. Advantage It is easy to implement because does not require any special apparatus for computing derivatives. Disadvantage It is very inefficient when there are more than a few local variables. A second disadvantage is that gradient ascent may get stuck in local optima.

Reinforcement Learning Procedures Delayed reinforcement In many real systems, there is a delay between an action and the resultant reinforcement Temporal credit assignment is performed by explicitly computing the effect of each activity level on the eventual outcome. Genetic Algorithm Genetic algorithms operate on a population of individuals to produce a better adapted population. There is a fitness function which assigns a real-valued fitness to each individual and the aim of the "learning" is to raise the average fitness of the population.

Overview 1. Introduction 2. Connectionist Models 3. Connectionist Research Issues 4. Associative Memories without Hidden Units 5. Simple Supervised Learning Procedures 6. Back propagation 7. Boltzman Machines 8. Maximizing Mutual Information 9. Unsupervised Hebbian Learning 10. Competitive Learning 11. Reinforcement Learning Procedures 12. Generalization

Generalization A major goal of connectionist learning is to produce networks that generalize correctly to new cases after training on a sufficiently large set of typical cases from some domain. Improve Generalization To introduce an extra term into the error function. This term penalizes large weights and it can be viewed as a way of building in an a priori bias is favor of simple models To impose equality constraints between weights that encode symmetries in the task.

Conclusion There are now many different connectionist learning procedures and many more variations will be discovered in the next few years. Major new advances can be expected on Making the learning time scale better To apply connectionist procedures to difficult tasks like speech recognition Simulating much larger networks Interpreting the behavior of real neural networks.