Neural Networks. CSC 4504 : Langages formels et applications. J Paul Gibson, D311.

Similar documents
Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Artificial Neural Networks written examination

Artificial Neural Networks

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Python Machine Learning

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Learning Methods for Fuzzy Systems

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

Evolution of Symbolisation in Chimpanzees and Neural Nets

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Softprop: Softmax Neural Network Backpropagation Learning

Second Exam: Natural Language Parsing with Neural Networks

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION

Knowledge Transfer in Deep Convolutional Neural Nets

Lecture 1: Machine Learning Basics

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

An OO Framework for building Intelligence and Learning properties in Software Agents

Seminar - Organic Computing

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Lecture 1: Basic Concepts of Machine Learning

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Test Effort Estimation Using Neural Network

INPE São José dos Campos

Learning to Schedule Straight-Line Code

Evolutive Neural Net Fuzzy Filtering: Basic Description

(Sub)Gradient Descent

Classification Using ANN: A Review

Early Model of Student's Graduation Prediction Based on Neural Network

Lecture 10: Reinforcement Learning

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Using focal point learning to improve human machine tacit coordination

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Deep Neural Network Language Models

Mathematics process categories

Are You Ready? Simplify Fractions

CSL465/603 - Machine Learning

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

Dialog-based Language Learning

CS177 Python Programming

Students Understanding of Graphical Vector Addition in One and Two Dimensions

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

An Empirical and Computational Test of Linguistic Relativity

Mathematics Scoring Guide for Sample Test 2005

Predicting Students Performance with SimStudent: Learning Cognitive Skills from Observation

Human Emotion Recognition From Speech

Connectionism, Artificial Life, and Dynamical Systems: New approaches to old questions

Issues in the Mining of Heart Failure Datasets

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

Breaking the Habit of Being Yourself Workshop for Quantum University

Knowledge-Based - Systems

Speaker Identification by Comparison of Smart Methods. Abstract

Moderator: Gary Weckman Ohio University USA

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Neural Representation and Neural Computation. Philosophical Perspectives, Vol. 4, Action Theory and Philosophy of Mind (1990),

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Probability estimates in a scenario tree

THE REFLECTIVE SUPERVISION TOOLKIT

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Statistical Analysis of Climate Change, Renewable Energies, and Sustainability An Independent Investigation for Introduction to Statistics

LEGO MINDSTORMS Education EV3 Coding Activities

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

SARDNET: A Self-Organizing Feature Map for Sequences

phone hidden time phone

Modeling function word errors in DNN-HMM based LVCSR systems

Automatic Pronunciation Checker

On the Formation of Phoneme Categories in DNN Acoustic Models

2017 Florence, Italty Conference Abstract

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Shockwheat. Statistics 1, Activity 1

Multi-genre Writing Assignment

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Grade 6: Correlated to AGS Basic Math Skills

Introduction and Motivation

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Modeling function word errors in DNN-HMM based LVCSR systems

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Forget catastrophic forgetting: AI that learns after deployment

Similar Triangles. Developed by: M. Fahy, J. O Keeffe, J. Cooper

Laboratorio di Intelligenza Artificiale e Robotica

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

TD(λ) and Q-Learning Based Ludo Players

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Henry Tirri* Petri Myllymgki

Abstractions and the Brain

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

CS Machine Learning

A study of speaker adaptation for DNN-based speech synthesis

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Time series prediction

Laboratorio di Intelligenza Artificiale e Robotica

Secret Code for Mazes

Transcription:

CSC 4504 : Langages formels et applications J Paul Gibson, D311 paul.gibson@telecom-sudparis.eu /~gibson/teaching/csc4504/problem11-neuralnetworks.pdf Neural Networks 1

2

The following slides are a summary of material (by DANIEL SHIFFMAN) found at: http://natureofcode.com/book/chapter-10-neural-networks/ 1943 - Warren S. McCulloch (neuroscientist) and Walter Pitts (logician) developed the first conceptual model of an artificial neural network "A logical calculus of the ideas imminent in nervous activity, describes the concept of a neuron, a single cell living in a network of cells that computes - it receives inputs, processes those inputs, and generates an output. Their work is not meant to accurately describe how the biological brain works. Rather, an (artificial) neural network was designed as a computational model based on the brain to solve certain kinds of problems. These problems are usually easy-for-a-human, difficult-for-a-machine: e.g., pattern recognition. 3

A neural network is a connectionist computational system. Information is processed collectively, in parallel throughout a network of nodes (the nodes, in this case, being neurons) 4

Weights in a perceptron A neural network is not just a complex system, but a complex adaptive system, meaning it can change its internal structure based on the information flowing through it. Typically, this is achieved through the adjusting of weights. This adaption can be thought of as learning 5

Learning strategies: Supervised Learning Essentially, a strategy that involves a teacher that is smarter than the network itself. For example, let s take the facial recognition example. The teacher shows the network a bunch of faces, and the teacher already knows the name associated with each face. The network makes its guesses, then the teacher provides the network with the answers. The network can then compare its answers to the known correct ones and make adjustments according to its errors. Unsupervised Learning Required when there isn t an example data set with known answers. Imagine searching for a hidden pattern in a data set. An application of this is clustering, i.e. dividing a set of elements into groups according to some unknown pattern. Reinforcement Learning A strategy built on observation. Think of a little mouse running through a maze. If it turns left, it gets a piece of cheese; if it turns right, it receives a little shock. (Don t worry, this is just a pretend mouse.) Presumably, the mouse will learn over time to turn left. Its neural network makes a decision with an outcome (turn left or right) and observes its environment (yum or ouch). If the observation is negative, the network can adjust its weights in order to make a different decision the next time. Reinforcement learning is common in robotics and autonomous vehicles. 6

The Perceptron Invented in 1957 by Frank Rosenblatt at the Cornell Aeronautical Laboratory, a perceptron is the simplest neural network possible: a computational model of a single neuron. A perceptron consists of one or more inputs, a processor, and a single output. 7

A perceptron follows the feed-forward model, meaning inputs are sent into the neuron, are processed, and result in an output. 8

The Perceptron Algorithm: 1 For every input, multiply that input by its weight. 2 Sum all of the weighted inputs. 3 Compute the output of the perceptron based on that sum passed through an activation function (e.g., the sign of the sum). 9

Simple Pattern Recognition Using a Perceptron Now that we understand the computational process of a perceptron, we can look at an example of one used for pattern recognition: Can we train a perceptron to decide if a given point is above or below a specified line? 10

Initial design is flawed because the point 0,0 will always give a 0 output (for any weights) A bias input always has the value of 1 and is also weighted 11

Did the perceptron get it right? At this point, the perceptron has no better than a 50/50 chance of arriving at the right answer. Remember, when we created it, we gave each weight a random value. A neural network isn t magic. It s not going to be able to guess anything correctly unless we teach it how to! 12

Supervised learning With this method, the network is provided with inputs for which there is a known answer. This way the network can find out if it has made a correct guess. If it s incorrect, the network can learn from its mistake and adjust its weights. The process is as follows: 1 Provide the perceptron with inputs for which there is a known answer. 2 Ask the perceptron to guess an answer. 3 Compute the error. (Did it get the answer right or wrong?) 4 Adjust all the weights according to the error. 5 Return to Step 1 and repeat! 13

Proposed Training Approach Neural Networks 14

Problem: Neural Networks 1) implement the simple perceptron (as at the web site, 10.4 Coding the Perceptron- http://natureofcode.com/book/chapter-10-neural-networks/) in the language of your choice; and test that it learns as required 2) design and implement a perceptron based NN to learn to recognise whether a given point is inside a specified triangle: X1,Y1 X2,Y2 Hint - you may need a different network topology and may need a back propagation technique. 15

Further Reading Neural Networks Rumelhart, David E., Geoffrey E. Hinton, and Ronald J. Williams. "Learning representations by back-propagating errors." Cognitive modeling 5.3 (1988): 1. Widrow, Bernard, and Michael A. Lehr. "30 years of adaptive neural networks: perceptron, madaline, and backpropagation." Proceedings of the IEEE 78.9 (1990): 1415-1442. Basheer, I. A., and M. Hajmeer. "Artificial neural networks: fundamentals, computing, design, and application." Journal of microbiological methods 43.1 (2000): 3-31. Hinton, Geoffrey E., and Ruslan R. Salakhutdinov. "Reducing the dimensionality of data with neural networks." Science 313.5786 (2006): 504-507. Bengio, Yoshua. "Learning deep architectures for AI." Foundations and trends in Machine Learning 2.1 (2009): 1-127. 16