Introduction to Neural Networks and Their History

Similar documents
Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Artificial Neural Networks

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

Artificial Neural Networks written examination

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

Axiom 2013 Team Description Paper

Seminar - Organic Computing

Lecture 1: Basic Concepts of Machine Learning

Python Machine Learning

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Learning Methods for Fuzzy Systems

Knowledge-Based - Systems

Abstractions and the Brain

1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION

Evolutive Neural Net Fuzzy Filtering: Basic Description

An OO Framework for building Intelligence and Learning properties in Software Agents

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

Dinesh K. Sharma, Ph.D. Department of Management School of Business and Economics Fayetteville State University

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Issues in the Mining of Heart Failure Datasets

TD(λ) and Q-Learning Based Ludo Players

Knowledge Transfer in Deep Convolutional Neural Nets

Rule Learning With Negation: Issues Regarding Effectiveness

SARDNET: A Self-Organizing Feature Map for Sequences

Speaker Identification by Comparison of Smart Methods. Abstract

An Introduction to Simio for Beginners

Time series prediction

Master s Programme in Computer, Communication and Information Sciences, Study guide , ELEC Majors

Lecture 10: Reinforcement Learning

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Soft Computing based Learning for Cognitive Radio

MYCIN. The MYCIN Task

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Evolution of Symbolisation in Chimpanzees and Neural Nets

Laboratorio di Intelligenza Artificiale e Robotica

Emergent Narrative As A Novel Framework For Massively Collaborative Authoring

Learning to Schedule Straight-Line Code

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Using focal point learning to improve human machine tacit coordination

Human Emotion Recognition From Speech

2017 Florence, Italty Conference Abstract

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

CSL465/603 - Machine Learning

Laboratorio di Intelligenza Artificiale e Robotica

INPE São José dos Campos

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

Softprop: Softmax Neural Network Backpropagation Learning

A GENERIC SPLIT PROCESS MODEL FOR ASSET MANAGEMENT DECISION-MAKING

Intelligent Agents. Chapter 2. Chapter 2 1

No Parent Left Behind

Henry Tirri* Petri Myllymgki

Forget catastrophic forgetting: AI that learns after deployment

Rule Learning with Negation: Issues Regarding Effectiveness

Module Title: Managing and Leading Change. Lesson 4 THE SIX SIGMA

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Neuroscience I. BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6. Fall credit hours

Accelerated Learning Online. Course Outline

Probabilistic Latent Semantic Analysis

Speech Recognition at ICSI: Broadcast News and beyond

ONE YEAR IN BARCELONA, PART I+II

Circuit Simulators: A Revolutionary E-Learning Platform

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

DEVELOPMENT OF AN INTELLIGENT MAINTENANCE SYSTEM FOR ELECTRONIC VALVES

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach

(Sub)Gradient Descent

General Microbiology (BIOL ) Course Syllabus

Syntactic systematicity in sentence processing with a recurrent self-organizing network

Undergraduate Program Guide. Bachelor of Science. Computer Science DEPARTMENT OF COMPUTER SCIENCE and ENGINEERING

On-Line Data Analytics

Classification Using ANN: A Review

MINISTRY OF EDUCATION

Kamaldeep Kaur University School of Information Technology GGS Indraprastha University Delhi

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

An Embodied Model for Sensorimotor Grounding and Grounding Transfer: Experiments With Epigenetic Robots

LEGO MINDSTORMS Education EV3 Coding Activities

Test Effort Estimation Using Neural Network

Innovative Methods for Teaching Engineering Courses

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Computerized Adaptive Psychological Testing A Personalisation Perspective

Multiagent Simulation of Learning Environments

Top US Tech Talent for the Top China Tech Company

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Accelerated Learning Course Outline

University of Groningen. Systemen, planning, netwerken Bosman, Aart

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

A Pipelined Approach for Iterative Software Process Model

Reinforcement Learning by Comparing Immediate Reward

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

A Context-Driven Use Case Creation Process for Specifying Automotive Driver Assistance Systems

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

***** Article in press in Neural Networks ***** BOTTOM-UP LEARNING OF EXPLICIT KNOWLEDGE USING A BAYESIAN ALGORITHM AND A NEW HEBBIAN LEARNING RULE

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

IT Students Workshop within Strategic Partnership of Leibniz University and Peter the Great St. Petersburg Polytechnic University

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Transcription:

Introduction to Neural Networks and Their History Introduction to Neural Networks : Lecture 1 (part 2) John A. Bullinaria, 2004 1. What are Neural Networks? 2. Why are Artificial Neural Networks Worth Studying? 3. Learning in Neural Networks 4. A Brief History of the Field 5. Artificial Neural Networks compared with Classical Symbolic AI 6. Some Current Artificial Neural Network Applications L1-8

What are Neural Networks? 1. Neural Networks (NNs) are networks of neurons, for example, as found in real (i.e. biological) brains. 2. Artificial Neurons are crude approximations of the neurons found in brains. They may be physical devices, or purely mathematical constructs. 3. Artificial Neural Networks (ANNs) are networks of Artificial Neurons, and hence constitute crude approximations to parts of real brains. They may be physical devices, or simulated on conventional computers. 4. From a practical point of view, an ANN is just a parallel computational system consisting of many simple processing elements connected together in a specific way in order to perform a particular task. 5. One should never lose sight of how crude the approximations are, and how over-simplified our ANNs are compared to real brains. L1-9

Why are Artificial Neural Networks worth studying? 1. They are extremely powerful computational devices (Turing equivalent, universal computers). 2. Massive parallelism makes them very efficient. 3. They can learn and generalize from training data so there is no need for enormous feats of programming. 4. They are particularly fault tolerant this is equivalent to the graceful degradation found in biological systems. 5. They are very noise tolerant so they can cope with situations where normal symbolic systems would have difficulty. 6. In principle, they can do anything a symbolic/logic system can do, and more. (In practice, getting them to do it can be rather difficult ) L1-10

What are Artificial Neural Networks used for? As with the field of AI in general, there are two basic goals for neural network research: Brain modelling : The scientific goal of building models of how real brains work. This can potentially help us understand the nature of human intelligence, formulate better teaching strategies, or better remedial actions for brain damaged patients. Artificial System Building : The engineering goal of building efficient systems for real world applications. This may make machines more powerful, relieve humans of tedious tasks, and may even improve upon human performance. These should not be thought of as competing goals. We often use exactly the same networks and techniques for both. Frequently progress is made when the two approaches are allowed to feed into each other. There are fundamental differences though, e.g. the need for biological plausibility in brain modelling, and the need for computational efficiency in artificial system building. L1-11

Learning in Neural Networks There are many forms of neural networks. Most operate by passing neural activations through a network of connected neurons. One of the most powerful features of neural networks is their ability to learn and generalize from a set of training data. They adapt the strengths/weights of the connections between neurons so that the final output activations are correct. There are three broad types of learning: 1. Supervised Learning (i.e. learning with a teacher) 2. Reinforcement learning (i.e. learning with limited feedback) 3. Unsupervised learning (i.e. learning with no help) This module will study in some detail the most common learning algorithms for the most common types of neural network. L1-12

A Brief History of the Field 1943 McCulloch and Pitts proposed the McCulloch-Pitts neuron model 1949 Hebb published his book The Organization of Behavior, in which the Hebbian learning rule was proposed. 1958 Rosenblatt introduced the simple single layer networks now called Perceptrons. 1969 Minsky and Papert s book Perceptrons demonstrated the limitation of single layer perceptrons, and almost the whole field went into hibernation. 1982 Hopfield published a series of papers on Hopfield networks. 1982 Kohonen developed the Self-Organising Maps that now bear his name. 1986 The Back-Propagation learning algorithm for Multi-Layer Perceptrons was rediscovered and the whole field took off again. 1990s 2000s The sub-field of Radial Basis Function Networks was developed. The power of Ensembles of Neural Networks and Support Vector Machines becomes apparent. L1-13

ANNs compared with Classical Symbolic AI The distinctions can put under three headings: 1. Level of Explanation 2. Processing Style 3. Representational Structure These lead to a traditional set of dichotomies: 1. Sub-symbolic vs. Symbolic 2. Non-modular vs. Modular 3. Distributed representation vs. Localist representation 4. Bottom up vs. Top Down 5. Parallel processing vs. Sequential processing In practice, the distinctions are becoming increasingly blurred. L1-14

Some Current Artificial Neural Network Applications Brain modelling Models of human development help children with developmental problems Simulations of adult performance aid our understanding of how the brain works Neuropsychological models suggest remedial actions for brain damaged patients Real world applications Financial modelling predicting stocks, shares, currency exchange rates Other time series prediction climate, weather, airline marketing tactician Computer games intelligent agents, backgammon, first person shooters Control systems autonomous adaptable robots, microwave controllers Pattern recognition speech recognition, hand-writing recognition, sonar signals Data analysis data compression, data mining, PCA, GTM Noise reduction function approximation, ECG noise reduction Bioinformatics protein secondary structure, DNA sequencing L1-15

Overview and Reading 1. Artificial Neural Networks are powerful computational systems consisting of many simple processing elements connected together to perform tasks analogously to biological brains. 2. They are massively parallel, which makes them efficient, robust, fault tolerant and noise tolerant. 3. They can learn from training data and generalize to new situations. 4. They are useful for brain modelling and real world applications involving pattern recognition, function approximation, prediction, Reading 1. Haykin: Sections 1.1, 1.8, 1.9 2. Gurney: Sections 1.1, 1.2, 1.3 3. Beale & Jackson: Sections 1.1, 1.2, 1.3, 1.4 4. Ham & Kostanic: Sections 1.1, 1.2 L1-16