Introduction to Neural Networks. Terrance DeVries

Similar documents
Lecture 1: Machine Learning Basics

Python Machine Learning

(Sub)Gradient Descent

Artificial Neural Networks written examination

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Evolutive Neural Net Fuzzy Filtering: Basic Description

Artificial Neural Networks

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

arxiv: v1 [cs.lg] 15 Jun 2015

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Generative models and adversarial training

Axiom 2013 Team Description Paper

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

INPE São José dos Campos

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Human Emotion Recognition From Speech

Test Effort Estimation Using Neural Network

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Assignment 1: Predicting Amazon Review Ratings

Softprop: Softmax Neural Network Backpropagation Learning

Model Ensemble for Click Prediction in Bing Search Ads

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

An empirical study of learning speed in backpropagation

Learning to Schedule Straight-Line Code

CSL465/603 - Machine Learning

Knowledge Transfer in Deep Convolutional Neural Nets

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Calibration of Confidence Measures in Speech Recognition

School of Innovative Technologies and Engineering

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

Evolution of Symbolisation in Chimpanzees and Neural Nets

Speaker Identification by Comparison of Smart Methods. Abstract

arxiv: v1 [cs.lg] 7 Apr 2015

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Learning Methods for Fuzzy Systems

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

A study of speaker adaptation for DNN-based speech synthesis

Application of Virtual Instruments (VIs) for an enhanced learning environment

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

Software Maintenance

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Lecture 10: Reinforcement Learning

SARDNET: A Self-Organizing Feature Map for Sequences

ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering

Spinal Cord. Student Pages. Classroom Ac tivities

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

CS Machine Learning

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Modeling function word errors in DNN-HMM based LVCSR systems

Networks in Cognitive Science

Accelerated Learning Course Outline

Soft Computing based Learning for Cognitive Radio

A Comparison of Annealing Techniques for Academic Course Scheduling

Mathematics process categories

Neuroscience I. BIOS/PHIL/PSCH 484 MWF 1:00-1:50 Lecture Center F6. Fall credit hours

A Review: Speech Recognition with Deep Learning Methods

Classification Using ANN: A Review

Word Segmentation of Off-line Handwritten Documents

Using focal point learning to improve human machine tacit coordination

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Discriminative Learning of Beam-Search Heuristics for Planning

Probability and Statistics Curriculum Pacing Guide

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

arxiv: v1 [cs.cv] 10 May 2017

12- A whirlwind tour of statistics

Truth Inference in Crowdsourcing: Is the Problem Solved?

FF+FPG: Guiding a Policy-Gradient Planner

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Reinforcement Learning by Comparing Immediate Reward

Early Model of Student's Graduation Prediction Based on Neural Network

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

On the Combined Behavior of Autonomous Resource Management Agents

Radius STEM Readiness TM

Time series prediction

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Data Fusion Through Statistical Matching

Modeling function word errors in DNN-HMM based LVCSR systems

Lecture 1: Basic Concepts of Machine Learning

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Probability and Game Theory Course Syllabus

Focused on Understanding and Fluency

Accelerated Learning Online. Course Outline

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

Introduction to Simulation

Chapters 1-5 Cumulative Assessment AP Statistics November 2008 Gillespie, Block 4

MTH 141 Calculus 1 Syllabus Spring 2017

Math 181, Calculus I

A Reinforcement Learning Variant for Control Scheduling

Semantic and Context-aware Linguistic Model for Bias Detection

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach

Transcription:

Introduction to Neural Networks Terrance DeVries

Contents 1. Brief overview of neural networks 2. Introduction to PyTorch (Jupyter notebook) 3. Implementation of simple neural network (Jupyter notebook)

What is an Artificial Neural Network? Predictive model that can learn to map given inputs to desired outputs Mathematical function designed to mimic the brain Artificial Neural Network Biological Neural Network

The Biological Neuron The brain contains billions of interconnected neurons. 1. Dendrites take in inputs 2. Cell does some electrochemical processing 3. If resulting voltage is greater than some threshold, the neuron fires 4. Signal is sent down axon to other neurons

The Artificial Neuron Artificial neural networks are composed of many artificial neurons. 1. Neuron receives inputs 2. Each input is multiplied by some weight and then summed together 3. Pass response through an activation function 4. Output signal is sent to other neurons

The Artificial Neuron An artificial neuron without an activation function is simply linear regression x = input value y = predicted value m = slope of the line b = bias y = mx + b

Activation Function Simulates the firing of a biological neuron Allows the neural network to model non-linear problems (only if the activation function is also non-linear) (Equivalent to having no activation function)

Interactive Demo https://playground.tensorflow.org

Universal Approximation Theorem A neural network with at least one hidden layer can approximate any continuous function. This is very powerful: for any set of input-output pairs, there exists a neural network that can almost perfectly model them Some limitations: Number of neurons may be impractically large Generalization to new samples is not guaranteed It may be difficult to find the correct weights

How Do We Find the Correct Weights? Stochastic Gradient Descent (SGD): Iterative method for optimizing differentiable functions. 1. Randomly initialize weights and select learning rate 2. Repeat until convergence: To calculate we need a loss function, and to calculate we use error backpropogation.

Loss Function Loss function measures how far away the prediction is from the desired output (i.e. error) Use gradient descent to minimize the loss Regression loss function: Mean squared error (MSE): Classification loss function: Cross entropy:

Error Backpropogation In order to calculate the error attributed to each weight we use the backpropogation algorithm: 1. Propagate forward through the network to generate an output 2. Calculate the loss (i.e. error) 3. Use chain rule to calculate the error associated with each neuron

Training Loop 1. Load batch of training inputs 2. Perform forward pass 3. Calculate loss 4. Backpropogate errors 5. Update weights 6. Repeat until convergence Epoch One pass through the training loop is called an iteration. One pass through the dataset is called an epoch. Multiple epochs are usually required before the model converges.

Why Neural Networks? Automatic feature extraction No need to hand-craft features Extremely versatile Can be adapted to a wide variety of non-standard problems Performance scales with the amount of data

Deep Learning Libraries Provides optimized implementations of common neural network building blocks Automatic differentiation - no need to manually calculate derivatives! Some libraries provide tools for deploying trained models

Jupyter Notebook https://jupyter.co60.ca