Artificial Learning in Artificial Memories Introduction

Similar documents
OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Artificial Neural Networks

LEGO MINDSTORMS Education EV3 Coding Activities

ENEE 302h: Digital Electronics, Fall 2005 Prof. Bruce Jacob

Beyond Classroom Solutions: New Design Perspectives for Online Learning Excellence

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Designing a Computer to Play Nim: A Mini-Capstone Project in Digital Design I

Circuit Simulators: A Revolutionary E-Learning Platform

Abstractions and the Brain

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Breaking the Habit of Being Yourself Workshop for Quantum University

Seminar - Organic Computing

Multidisciplinary Engineering Systems 2 nd and 3rd Year College-Wide Courses

Physics 270: Experimental Physics

Lecture 10: Reinforcement Learning

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Python Machine Learning

(Sub)Gradient Descent

Computer Science. Embedded systems today. Microcontroller MCR

An Introduction to Simio for Beginners

A Process-Model Account of Task Interruption and Resumption: When Does Encoding of the Problem State Occur?

A Context-Driven Use Case Creation Process for Specifying Automotive Driver Assistance Systems

Learning Methods for Fuzzy Systems

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Knowledge Transfer in Deep Convolutional Neural Nets

M55205-Mastering Microsoft Project 2016

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Strategy and Design of ICT Services

Software Maintenance

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Visual CP Representation of Knowledge

Evolutive Neural Net Fuzzy Filtering: Basic Description

A Pipelined Approach for Iterative Software Process Model

Infrared Paper Dryer Control Scheme

Commanding Officer Decision Superiority: The Role of Technology and the Decision Maker

XXII BrainStorming Day

Mastering Team Skills and Interpersonal Communication. Copyright 2012 Pearson Education, Inc. publishing as Prentice Hall.

Reinforcement Learning by Comparing Immediate Reward

Saliency in Human-Computer Interaction *

Axiom 2013 Team Description Paper

Phys4051: Methods of Experimental Physics I

Pod Assignment Guide

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Bluetooth mlearning Applications for the Classroom of the Future

Virtual Teams: The Design of Architecture and Coordination for Realistic Performance and Shared Awareness

Major Milestones, Team Activities, and Individual Deliverables

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

SAM - Sensors, Actuators and Microcontrollers in Mobile Robots

Learning to Schedule Straight-Line Code

Bluetooth mlearning Applications for the Classroom of the Future

EDEXCEL NATIONALS UNIT 25 PROGRAMMABLE LOGIC CONTROLLERS. ASSIGNMENT No.1 SELECTION CRITERIA

Evolution of Symbolisation in Chimpanzees and Neural Nets

Human Emotion Recognition From Speech

Contents. Foreword... 5

Backwards Numbers: A Study of Place Value. Catherine Perez

What is PDE? Research Report. Paul Nichols

USER ADAPTATION IN E-LEARNING ENVIRONMENTS

LISTENING STRATEGIES AWARENESS: A DIARY STUDY IN A LISTENING COMPREHENSION CLASSROOM

Computer Organization I (Tietokoneen toiminta)

Disciplinary Literacy in Science

Computer Architecture CSC

Types of curriculum. Definitions of the different types of curriculum

A Case-Based Approach To Imitation Learning in Robotic Agents

Rover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes

Artificial Neural Networks written examination

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Probability estimates in a scenario tree

University of Toronto Physics Practicals. University of Toronto Physics Practicals. University of Toronto Physics Practicals

Initial English Language Training for Controllers and Pilots. Mr. John Kennedy École Nationale de L Aviation Civile (ENAC) Toulouse, France.

Role of Blackboard Platform in Undergraduate Education A case study on physiology learning in nurse major

A General Class of Noncontext Free Grammars Generating Context Free Languages

Session H1B Teaching Introductory Electrical Engineering: Project-Based Learning Experience

BOOK INFORMATION SHEET. For all industries including Versions 4 to x 196 x 20 mm 300 x 209 x 20 mm 0.7 kg 1.1kg

Australian Journal of Basic and Applied Sciences

CEFR Overall Illustrative English Proficiency Scales

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Understanding and Changing Habits

Types of curriculum. Definitions of the different types of curriculum

An empirical study of learning speed in backpropagation

Linking Task: Identifying authors and book titles in verbose queries

Computer Science 141: Computing Hardware Course Information Fall 2012

Modeling user preferences and norms in context-aware systems

Data Fusion Models in WSNs: Comparison and Analysis

UNDERSTANDING DECISION-MAKING IN RUGBY By. Dave Hadfield Sport Psychologist & Coaching Consultant Wellington and Hurricanes Rugby.

similar to the majority ofcomputer science courses in colleges and universities today. Classroom time consisted of lectures, albeit, with considerable

Utilizing Soft System Methodology to Increase Productivity of Shell Fabrication Sushant Sudheer Takekar 1 Dr. D.N. Raut 2

BMBF Project ROBUKOM: Robust Communication Networks

Introduction and Motivation

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

The Strong Minimalist Thesis and Bounded Optimality

On the Combined Behavior of Autonomous Resource Management Agents

Getting Started with Deliberate Practice

Neuroscience I. BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6. Fall credit hours

Using the CU*BASE Member Survey

Transcription:

Artificial Learning in Artificial Memories John Robert Burger Professor Emeritus Department of Electrical and Computer Engineering 25686 Dahlin Road Veneta, OR 97487 (jrburger1@gmail.com) Abstract Memory refinements are designed below to detect those sequences of actions that have been repeated a given number n. Subsequently such sequences are permitted to run without CPU involvement. This mimics human learning. Actions are rehearsed and once learned, they are performed automatically without conscious involvement. Introduction This paper is written from the perspective of designing artificial brains that are modeled after human brains. This architectural work pays attention to human psychology including the interplay between short and long term memory, the realization of which is assumed constrained by circuit and system theory. This paper is a humble attempt, but at least it has an original aspect. My fellow engineering stereotypes tend to ignore human psychology; worse yet, scientists in life sciences, seem unaware of the constraints imposed by working circuits and systems. Although their products such as long term potentiation are impressive at first sight and sometimes work in cyberspace, they are most unpromising for practical humanoid realizations. Learning is much more than random or mathematical optimization leading to a successful sequence of actions, although no doubt successful actions need to be discovered. The discovery of successful actions ought to be termed searching or investigating, anything but learning. Learning is a specific attribute beyond mere memory that implies a long term physical modification of underlying circuitry such that, in the big picture, human efficacy increases. Computers today obviously cannot be modified in any fundamental way by their own programming. Consequently, the envisioned learning, as below, is unrelated to the field of machine learning. Machine learning is traditionally limited to the programming of structured, fundamentally unalterable computers (Sun & Giles, 2001; Alpaydin 2010). There are times when we might benefit from artificial learning in everyday life. For example, a person might perform certain keystrokes over and over; one opens a mail server, the inbox is opened, email is selected in chronological order and it is properly filed, each action hundreds of times per year. Yet today s best personal computers will not learn even this! This paper is not about the programming of artificial neural networks (Haykin 1994). The training of artificial neural networks using, for example, gradient descent, has been construed as learning, but properly this is not learning at all. It is an exercise in mathematical optimization without biological underpinning. 1

Learning Modeled After Biology Learning may be viewed in the biological sense as an internalization process that detects, remembers and reproduces successful sequences. Learning results in performing an action, or possibly several actions concurrently without cognitive effort or thinking, meaning without conscious involvement. Memory hardware is designed below to learn repeated actions, analogous to rehearsal when humans learn. Much of what we learn is a sequence of actions, for examples, proper components of walking, reciting a poem or reproducing a musical tune. Such things are memory-mapped actions and may be performed without CPU effort (analogous to thinking) once they have been learned, and this is a key point. Properties of artificial learning for memory-mapped sequences are summarized in Table 1. Table 1 Artificial Learning 1 A result of rehearsal 2 Occurs within long term memory 3 Enables long term modification of underlying circuitry 4 Permits action without CPU effort 5 Permits action without CPU delays 6 Permits action without CPU memory usage 7 Independent sequences may run concurrently Overview of a Memory System Artificial human memory has been modeled as employing cues in a routine to find matches within associative memory (Burger 2009). One may envision an anthropomorphic system of dynamic (short term) working memory as in Fig. 1. Working memory has been modeled as orchestrating long term memory (Burger Aug 30, 2010). Long term memory in humans is analogous to PROM (programmable read only memory that may be programmed exactly once). In humans, the contents of working memory are known (via consciousness) but all else, including editing processes are hidden. An important albeit neglected attribute of any memory system is its ability to learn. Learning is definitely not memorization. Learning here means an ability to run a sequence of memory words without controls from working memory (or CPU). This ability occurs only if the sequence is sufficiently rehearsed and only if a word of the sequence has been addressed by working memory. This type of learning has been termed state machine learning. Artificial Leaning for Artificial Memories A word of long term memory is assumed to hold components for images or actions. Each word in a practical system can be diagramed as in Fig. 2. The signals and commands held by the word are released by activating the enable. Once the signals and commands have all been discharged, there is a done signal. 2

To learn something that is being practiced, a timing filter is required to detect a repeated sequence. For example the filter must detect whenever a certain word is enabled immediately after another given word is done. A suitable digital filter to detect a sequence exploits the delay between active words, specified to be Delay 1 as in Fig. 3. Part (a) of the figure indicates that a done signal from Word 1 is held for a time Delay 1. Implementation of Delay 1 is not shown but is straightforward using elements of short term memory. When a second done signal arrives in a timely way from some other word, say from Word 2, the AND gate is activated to make X 1 true. X 1 results in the generation of a single spike. The spike operates a shift register of set-only D- latches identified as D 1, D 2... D n-1, D n. All D-latches initially are set to Boolean zero (cleared). Each time the 1-2 sequence occurs, a true signal shifts to the left one place in the shift register. Latch details appear in part (b) of the Fig. 3. Note that the D-latches can be set true only once in this plan, since learning, once it occurs, is assumed long term. The spike generator in part (c) of the figure uses a standard XOR (exclusive OR) to produce a brief spike on the leading edge of signal X 1. Spike width is determined by DELAY 2. Spike width is just enough to move the Boolean signal X 1 only one step along the shift register. AND, OR and XOR gates are readily available both in brain neurons and in hardware (Burger Apr 2010). The design in Fig. 4 supports possible state transitions from any given word to any other given word in a block of n words within long term memory. The maximum number of possible sequences is n! Fig. 4 shows only three words, although many more can be used. In an attempt to gain focus, consider the sequence 1-3-2 that has been used, say, ten times so that it is well learned. Then, whenever Word 1 is addressed (associatively) by short term memory a minor miracle occurs. After a time equal to Delay 1, Word 3 is addressed directly by Word 1. Then after another time equal to Delay 1, Word 2 is addressed directly by Word 3. This will accomplish a three-step procedure without CPU involvement. Note that switches S ij (1 i K, 1 j K, i j) enable the state machine. For example, a switch S 13 applies the done signal from Word 1 to the delayed enable of Word 3. Switches are viewed by neuroscientists as synaptic connections promoted by interneurons, but computer engineers create contacts artificially with a FET (Field Effect Transistor). In general for K words and an arbitrary sequence that may go forward or backward, K (K-1) timing filters are needed. This implies a like number of bus lines for the filter outputs and a like number of switches S ij (1 i K, 1 j K, i j). Also there must be K lines for memory word outputs and K lines for memory word inputs. The overhead is reasonable for systems with complexity such as humanoid memory. Limitations of Artificial Learning A pitfall is that once a learned action is activated, it executes asynchronously. In practical memory, as in human memory, it is always possible to override learning. In this case, one 3

simply opens the S ij circuit using another series switch (not shown). This returns control to working memory same as it does in humans. Learning cannot be erased easily, so if the learning is wrong, a correct version must be relearned. This is also true of human learning. A limitation of any state machine is that states (words) must be distinguishable. This means that a learned sequence cannot repeat a given word in a sequence. For example, one cannot learn the sequence 1-3-1-2 because state 1 is repeated. However, different versions of a repeated word are readily possible, permitting for example, 1a-2-1b-3. Looping is not automatic since it would involve repeating a word, for example, 1-2-3-1-2-3. In this system a given sequence may be repeated by only a call from working memory. Conclusions Learning is largely important both to humans and to intelligent robots. Survival matters, which hinges on reflexive actions concurrent with quick thinking, especially when there is no time to ponder the pros and cons of a response. From another view, learning enables multitasking in the sense that while learned actions are being performed one may think about something else. One may read music and also play the violin, for instance. Learning as above depends on timing filters that anyone can build and verify. These filters detect and remember a sequence that is rehearsed a given number of times (n). Learned sequences, once called, run automatically within long term memory. CPU effort is not required. Obviously, this frees central processing for more important work while a sequence is running. More importantly, long term memory is freed from the constraints of CPU control. Consequently artificial learning creates the possibility of parallel processing since many learned sequences may run simultaneously assuming they involve independent memory words. 4

Fig. 1 Short term (working) memory (consciousness) provides cues that are edited subliminally while recalls from associative long term memory are tested unconsciously before refreshing working memory Word Fig. 2 Symbol for a word of long term memory with controls 5

Learning (Digital) Filter (a) Q Dn Dn-1 D2 D1 X 1 Delay1 Word1 Word 2 Ck Bus X 2 X 1 (b) Example Data Latch D1 Ck X 1 Ck (c) Delay2 < Delay1 Fig. 3 Learning filter showing how rehearsal technically results in permanent learning 6

3-2 3-1 2-3 2-1 1-3 1-2 Q 12 DIG FLTS 1-2 Flt 1-3 Flt 3 2 1 S 12 S 13 3 2 1 D D=Delay1 Wd 1 2-1 Flt S 21 D Wd 2 2-3 Flt 3-1 Flt 3-2 Flt S 23 S 31 S 32 D Wd 3... Wd K Fig. 4 Embedded state machine to learn an arbitrary sequence of words References Alpaydin E, Introduction to Machine Learning, second edition, MIT Press 2010 Burger J R, Human Memory Modeled with Standard Analog and Digital Circuits Inspiration for Man-made Computers, Wiley, 2009 Burger J R, XOR at a Single Vertex -- Artificial Dendrites, arxiv:1004.2280v1 [cs.ne], Apr 2010; [Online]: Available: http://arxiv.org/abs/1004.2280 Burger J R, Artificial Brain Based on Credible Neural Circuits in a Human Brain, arxiv:1008.5161v1 [cs.ai], Aug 30, 2010; [Online}: Available: http://arxiv.org/abs/1008.5161 Haykin S, Neural Networks: A Comprehensive Foundation, Macmillan, 1994. Sun R, Giles CL, Sequence Learning: From Recognition and Prediction to Sequential Decision Making, July/August 2001 (vol. 16 no. 4) pp. 67-70; [Online]: Available: http://scholar.google.com/scholar?q=sequence+learning:+from+recognition+and+predicti on+to+sequential+decision+making&hl=en&as_sdt=0&as_vis=1&oi=scholar 7