Psychology 452 Week 1: Connectionism and Association

Similar documents
OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Evolution of Symbolisation in Chimpanzees and Neural Nets

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Artificial Neural Networks

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Evolutive Neural Net Fuzzy Filtering: Basic Description

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

MTH 215: Introduction to Linear Algebra

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Abstractions and the Brain

How People Learn Physics

Circuit Simulators: A Revolutionary E-Learning Platform

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Python Machine Learning

Knowledge Transfer in Deep Convolutional Neural Nets

Learning to Schedule Straight-Line Code

Mathematics process categories

Learning Methods for Fuzzy Systems

Innovative Teaching in Science, Technology, Engineering, and Math

Mathematics Scoring Guide for Sample Test 2005

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Artificial Neural Networks written examination

Foothill College Summer 2016

INPE São José dos Campos

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION

Physics 270: Experimental Physics

Syllabus ENGR 190 Introductory Calculus (QR)

Human Emotion Recognition From Speech

Calculators in a Middle School Mathematics Classroom: Helpful or Harmful?

Speaker Identification by Comparison of Smart Methods. Abstract

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Mathematics. Mathematics

An Empirical and Computational Test of Linguistic Relativity

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

Lecture 1: Basic Concepts of Machine Learning

ENEE 302h: Digital Electronics, Fall 2005 Prof. Bruce Jacob

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

EECS 700: Computer Modeling, Simulation, and Visualization Fall 2014

SARDNET: A Self-Organizing Feature Map for Sequences

An empirical study of learning speed in backpropagation

Computer Science 141: Computing Hardware Course Information Fall 2012

Axiom 2013 Team Description Paper

A study of speaker adaptation for DNN-based speech synthesis

Reinforcement Learning by Comparing Immediate Reward

ME 443/643 Design Techniques in Mechanical Engineering. Lecture 1: Introduction

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

Neural Representation and Neural Computation. Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990),

Statewide Framework Document for:

Lecture 1: Machine Learning Basics

Introducing the New Iowa Assessments Mathematics Levels 12 14

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Lecturing Module

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

On the Formation of Phoneme Categories in DNN Acoustic Models

Alignment of Australian Curriculum Year Levels to the Scope and Sequence of Math-U-See Program

While you are waiting... socrative.com, room number SIMLANG2016

Accelerated Learning Course Outline

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

A Reinforcement Learning Variant for Control Scheduling

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Connectionism, Artificial Life, and Dynamical Systems: New approaches to old questions

Communication and Cybernetics 17

Accelerated Learning Online. Course Outline

Detailed course syllabus

Networks in Cognitive Science

Developing a concrete-pictorial-abstract model for negative number arithmetic

DIDACTIC MODEL BRIDGING A CONCEPT WITH PHENOMENA

Test Effort Estimation Using Neural Network

UNIT ONE Tools of Algebra

Speech Recognition at ICSI: Broadcast News and beyond

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Grade 6: Correlated to AGS Basic Math Skills

Undergraduate Program Guide. Bachelor of Science. Computer Science DEPARTMENT OF COMPUTER SCIENCE and ENGINEERING

Rule Learning With Negation: Issues Regarding Effectiveness

Lecture 10: Reinforcement Learning


AP Chemistry

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Student Perceptions of Reflective Learning Activities

Backwards Numbers: A Study of Place Value. Catherine Perez

Using the Artificial Neural Networks for Identification Unknown Person

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Radius STEM Readiness TM

An Introduction to Simio for Beginners

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Major Milestones, Team Activities, and Individual Deliverables

TOPICS LEARNING OUTCOMES ACTIVITES ASSESSMENT Numbers and the number system

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Speeding Up Reinforcement Learning with Behavior Transfer

The Algebra in the Arithmetic Finding analogous tasks and structures in arithmetic that can be used throughout algebra

An OO Framework for building Intelligence and Learning properties in Software Agents

Forget catastrophic forgetting: AI that learns after deployment

MYCIN. The MYCIN Task

Predicting Students Performance with SimStudent: Learning Cognitive Skills from Observation

D Road Maps 6. A Guide to Learning System Dynamics. System Dynamics in Education Project

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Transcription:

Psychology 452 Week 1: Connectionism and Association Course Overview Properties Of Connectionism Building Associations Into Networks The Hebb Rule The Delta Rule Michael R.W. Dawson PhD from University of Western Ontario Research interests in foundations of cognitive science, artificial neural networks, embodied cognitive science Research methods include computer simulation and LEGO robot fabrication For details about my research, go to my home web page Credentials Dawson has published several books on cognitive science He posts cognitive science material at Twitter.com/mrwdawson Course Objectives Introduce the foundations of connectionist cognitive science Provide hands on experience with various artificial neural networks 1

Course Evaluation Course WWW Support Midterm Exam (35%) Final Exam (40%) Same format as PSYCO 354 exams Text Assignments (25%) Done in class, handed in weekly Texts: Dawson, M.R.W. (2004). Minds and Machines: Connectionism and psychological modeling. Blackwell: Oxford, UK. Dawson, M.R.W. (2005). Connectionism: A Hands-On Approach. Blackwell: Oxford, UK Lots of WWW support Lectures, additional readings, information about assignments, pointers to other sites of relevance Software, training sets etc for assignments only available on the web http://www.bcp.psych.ualberta.ca/~mike/pearl_st reet/psyco452/ We will be exposed to this website in more detail during our hands-on activity later this evening Course Structure The Classical Approach When Weeks 1, 2, 3 Weeks 4, 5, 6 Week 7 Weeks 8, 9, 10 Weeks 11, 12 Week 13 What Connectionist Building Blocks Case Studies of Connectionism Midterm Exam Interpreting Connectionist Networks Deep Learning Basics Final Exam The Classical approach adopts a strict structure/rule distinction in its view of information processing. It views cognition as the rule-governed manipulation of symbols. 2

Connectionism Neuronal Inspiration Since the 1980s there has been an explosion of interest in parallel distributed processing (PDP) or connectionist architectures These architectures have been developed to solve a number of possible problems with classical accounts of cognitive science PDP modelers pay more attention to the brain than do Classical researchers A PDP processor can be viewed as an abstract, simplified description of a neuron Parallel Processing PDP models are networks of simple processors that operate simultaneously This causes fast computation, even if components are slow This is intended to fix the speed limitation of Classical models Distributed Representations A PDP network s knowledge is stored as a pattern of weighted connections between processors These connections are analogous to a Classical program This knowledge is very distributed, providing damage resistance and graceful degradation 3

Networks Learn What Can Networks Do? Artificial neural networks are rarely programmed Instead, they learn from experience Most of the networks that we will encounter learn from their mistakes The root of this learning is a basic law of association, the law of contiguity What kinds of tasks can modern networks perform? Networks are often used to classify patterns Networks are also capable of approximating functions We will see examples of both of these abilities throughout the course Let us consider how they might learn to do this! First Building Block: Association Hebb And Association James law of contiguity for associating two ideas together, where each idea is represented as a pattern of neural activity: When two elementary brainprocesses have been active together or in immediate succession, one of them, on reoccurring, tends to propagate its excitement into the other (James, 1890) William James When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes place in firing it, some growth process or metabolic change takes place in one or both cells such that A s efficiency, as one of the cells firing B, is increased (Hebb, 1949) Principle of contiguity! Let us use this principle to create a basic learning rule for a simple connectionist network! Donald Hebb 4

Distributed Associative Memory Hebb Rule 1 Simple connectionist architecture: Distributed associative memory Standard pattern associator One set of input units One set of output units Modifiable connections Task Learn associations between input and output patterns Later, present one pattern, and have the system recall the other associated pattern Let the memory start as a blank slate or tabula rasa Let all its connection weights be equal to zero Hebb Rule 2 Hebb Rule 3 Present two patterns of activity Associate the patterns because of their temporal contiguity Later, one pattern will cue the other Make more excitatory the connections between same-state processors Make more inhibitory the connections between opposite-state processors This is a version of Hebb s stated principle 5

Hebb Rule 4 Hebb Rule 5 To recall, activate processors with the cue Their activity sends a signal through existing connections This signal will cause activity in the output units The network signal should reconstruct the other pattern in the second set of processing units The system has learned the association between the two patterns, and is using it to permit one pattern to recall the other Desired Weight Changes Algebra of Learning Qualitatively, we can look at the activity of an input unit and an output unit to see how the weight between them should change to learn the association The relationship between them is multiplicative Activity Of Input Unit Activity Of Output Unit Direction Of Desired Weight Change Positive Positive Positive Negative Negative Positive Negative Positive Negative Positive Negative Negative Mathematically Hebb learning can be described as using the outer product of two vectors to create a matrix of weight changes These are added to existing connection weights This is associative learning! Trial (t+1) Operation Equation Describing Weight Values 0 Start with the 0 matrix 1 Associate a with b 2 Associate c with d W 0 =0 W 1 =W 0 + 1 =0+(b a T ) =(b a T ) W 2 =W 1 + 2 =(b a T )+ 2 =(b a T )+ (d c T ) 6

Algebra Of Recall Limitations Of Hebb Rule Recall from the memory involves sending an input signal through existing weights to produce output unit activity Algebraically this involves multiplying the matrix of existing connection weights through the vector of input unit activities r = Wc Hebb rule has many problems, which you will observe first hand from your exercises: Only learns orthogonal patterns Produces error when overtraining Unable to deal with linear dependence We would like to develop a new kind of Hebb learning rule This rule would permit the network to correctly recall correlated patterns This rule would also allow the network to improve its performance with repeated presentations of patterns Error And Weight Change The Delta Rule What would happen if we computed output unit error before we used the Hebb rule to learn associations? How would weights change if error was taken into account? Activity Of T - O Implication Operation To Direction Of Desired Input Unit Reduce Error Weight Change Positive Positive T > O O Positive Positive Negative T < O O Negative Positive Zero T = O None Zero Negative Positive T > O O Negative Negative Negative T < O O Positive Negative Zero T = O None Zero The delta rule can be viewed as a Hebb-style association between an input vector and an (output) error vector Repeated applications will reduce error The amount of learning depends on the amount of error The delta rule can be written as: t+1 = ((t - o) c T ) 7

The delta rule is a minor variation on the Hebb rule The delta rule is based on Hebb learning, and is the basis for other learning rules that we will encounter Comparing The Rules Two More Building Blocks Association is the first of three key building blocks for connectionist networks We still need to add nonlinearities into the processing units, letting them make decisions We still need to add some methods by which layers of these nonlinearities can be coordinated together These will be our topics in later lectures Chapter 1 Discussion Questions? Important Terms Analytic approach Synthetic approach Synthetic psychology 8