Machine Learning: Neural Networks. Junbeom Park Radiation Imaging Laboratory, Pusan National University

Similar documents
Python Machine Learning

(Sub)Gradient Descent

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Artificial Neural Networks written examination

Lecture 1: Machine Learning Basics

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Softprop: Softmax Neural Network Backpropagation Learning

arxiv: v1 [cs.lg] 15 Jun 2015

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Artificial Neural Networks

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Generative models and adversarial training

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Lecture 1: Basic Concepts of Machine Learning

CSL465/603 - Machine Learning

Using focal point learning to improve human machine tacit coordination

An empirical study of learning speed in backpropagation

STA 225: Introductory Statistics (CT)

Axiom 2013 Team Description Paper

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Learning to Schedule Straight-Line Code

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

A Review: Speech Recognition with Deep Learning Methods

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Dropout improves Recurrent Neural Networks for Handwriting Recognition

*** * * * COUNCIL * * CONSEIL OFEUROPE * * * DE L'EUROPE. Proceedings of the 9th Symposium on Legal Data Processing in Europe

A study of speaker adaptation for DNN-based speech synthesis

arxiv: v1 [cs.lg] 7 Apr 2015

Speaker Identification by Comparison of Smart Methods. Abstract

arxiv: v2 [cs.ro] 3 Mar 2017

CS Machine Learning

Model Ensemble for Click Prediction in Bing Search Ads

A Deep Bag-of-Features Model for Music Auto-Tagging

School of Innovative Technologies and Engineering

Soft Computing based Learning for Cognitive Radio

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Human Emotion Recognition From Speech

Learning Methods for Fuzzy Systems

Abstractions and the Brain

On the Formation of Phoneme Categories in DNN Acoustic Models

arxiv: v1 [cs.cv] 10 May 2017

Learning From the Past with Experiment Databases

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Evolutive Neural Net Fuzzy Filtering: Basic Description

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

WHEN THERE IS A mismatch between the acoustic

Second Exam: Natural Language Parsing with Neural Networks

Sociology 521: Social Statistics and Quantitative Methods I Spring Wed. 2 5, Kap 305 Computer Lab. Course Website

Probability and Statistics Curriculum Pacing Guide

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

INPE São José dos Campos

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

EGRHS Course Fair. Science & Math AP & IB Courses

arxiv: v1 [cs.cl] 27 Apr 2016

Knowledge Transfer in Deep Convolutional Neural Nets

1 NETWORKS VERSUS SYMBOL SYSTEMS: TWO APPROACHES TO MODELING COGNITION

Mathematics Program Assessment Plan

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

arxiv: v4 [cs.cl] 28 Mar 2016

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Guide to the Uniform mark scale (UMS) Uniform marks in A-level and GCSE exams

SARDNET: A Self-Organizing Feature Map for Sequences

Rover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

Dinesh K. Sharma, Ph.D. Department of Management School of Business and Economics Fayetteville State University

Detailed course syllabus

Multivariate k-nearest Neighbor Regression for Time Series data -

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Knowledge-Based - Systems

Early Model of Student's Graduation Prediction Based on Neural Network

A method to teach or reinforce concepts of restriction enzymes, RFLPs, and gel electrophoresis. By: Heidi Hisrich of The Dork Side

Deep Neural Network Language Models

Speech Emotion Recognition Using Support Vector Machine

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH AND LANGUAGE PROCESSING, VOL XXX, NO. XXX,

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Cultivating DNN Diversity for Large Scale Video Labelling

Assignment 1: Predicting Amazon Review Ratings

Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition

Laboratorio di Intelligenza Artificiale e Robotica

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

THE enormous growth of unstructured data, including

Test Effort Estimation Using Neural Network

Adaptive learning based on cognitive load using artificial intelligence and electroencephalography

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Mathematics. Mathematics

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

Transcription:

Machine Learning: Neural Networks Junbeom Park (pjb385@gmail.com) Radiation Imaging Laboratory, Pusan National University

1 Contents 1. Introduction 2. Machine Learning Definition and Types Supervised Learning Regression Gradient Descent Algorithm Classification 3. Neural Networks Structure History and Theoretical background Application in Research Signal Computation 4. Homework

2 Introduction

3

4 Machine Learning Field of study that gives computers the ability to learn without being explicitly programmed - 1959, Arthur Samuel (1901~1990) Supervised Types of Machine Learning Unsupervised

5 Supervised Learning 1. Regression Test score based on time spent for study Hours 2 4 8 12 Score 30 50 80 90 2. Classification Pass/Fail based on time spent for study Hours 2 4 8 12 Pass/Fail F F P P

6 Regression 1. Hypothesis Learning regression model is equal to formulating a hypothesis for a given data set. The task is to find a better hypothesis. HH xx = WWWW + bb 2. Cost function Introduced to fit the line to a given data set. Using squared cost function Distance can be expressed regardless of sign. Points of the farther distance have a greater effect. mm CCCCCCCC WW, bb = 1 mm ii=1 HH xx ii yy ii 2

7 Gradient Descent Algorithm This algorithm is used in many minimization problems. Formal definition of cost function to minimize cost. CCCCCCCC WW = 1 mm 2mm ii=1 WWxx ii yy ii 2 1. Start with initial guesses 2. Change some parameters a little bit to reduce cost. 3. After modifications, select the gradient which reduces cost the most possible. 4. Repeat the above process until you converge to a local minimum.

8 Gradient Descent Algorithm The gradient can be calculated by differentiating the cost function. WW WW ηη WW CCCCCCCC(WW) WW WW ηη WW 1 mm 2mm ii=1 WWxx ii yy ii 2 WW WW ηη 1 mm 2mm 2 WWxx ii yy ii xx ii ii=1 In case of the original cost function CCCCCCCC WW, bb = 1 mm HH xx ii yy ii 2 ii=1 CCCCCCCC mm mm WW WW ηη 1 mm ii=1 WWxx ii yy ii xx ii WW bb

9 Expend to Multi-Variables Regression using multi-inputs HH xx = WWWW + bb HH xx 1, xx 2,, xx nn = WW 1 xx 1 + WW 2 xx 2 + + WW nn xx nn + bb Using matrix (Implementation) HH XX = XXXX xx 1 xx nn WW 1 WW nn = xx 1 WW 1 + + xx 3 WW 3 Cost function mm CCCCCCCC WW, bb = 1 mm ii=1 HH xx 1 ii, xx 2 ii,, xx nn ii yy ii 2 CCCCCCCC WW = 1 2 XXWW yy mm

10 Classification The linear hypothesis has several disadvantages in the classification. Other appropriate type of hypothesis Since derivative is not possible, gradient descent algorithm can not be applied.

11 Logistic Hypothesis Linear hypothesis HH XX = XXXX Logistic hypothesis (Sigmoid) HH XX = 1 1 + ee XXXX The sigmoid function is differentiable. HHH XX = HH XX 1 HH XX These regression & classification functions are used as the activation function of neural networks.

12 Structure of Neural Networks

13 History of Neural Networks 1943 McCulloch: logical computation model based on simple neural networks A Logical Calculus of The Ideas Immanent in Nervous Activity 1949 Hebb: presentation of learning laws based on synapses. The Organization of Behavior 1957 Rosenblatt: development of perceptron terminology and algorithm. The Perceptron, A Perceiving and Recognizing Automaton Project Para

14 History of Neural Networks 1969 Minsky: XOR problems A long recession Perceptrons?? OR AND XOR 1. To solve XOR problem, we need to use multi-layer perceptrons. 2. No one on earth had found a viable way to train multi-layer perceptrons.

15 History of Neural Networks 1986 Rumelhart: development of error back-propagation algorithm. Learning Representations by Back-Propagating Errors For training networks, performing a cost function minimization via gradient descent algorithm.

16 Application in research Training Result Processing Result Original Radiograph Processing

17 Signal Computation: Feed-Forward Input signal vector : xx 1 NN

18 Signal Computation: Feed-Forward dd = xx UU 1 NN + 1 NN + 1 MM = 1 MM

19 Signal Computation: Feed-Forward aa = gg dd = 1 1 ee dd 1 MM

20 Signal Computation: Feed-Forward zz = aa VV 1 MM + 1 MM + 1 LL = 1 LL

21 Signal Computation: Feed-Forward yy = h zz = ααzz + ββ 1 LL

22 Signal Computation: Feed-Forward ee = 1 (yy tt)2 2 1 LL

23 Signal Computation: Error Back-propagation LL LL LL 1 1 (MM + 1) = LL MM + 1 R ηη Diagonal Matrix of the Derivatives of h The Learning Rate

24 Signal Computation: Error Back-propagation MM MM MM LL LL LL LL 1 1 (NN + 1) = MM NN + 1 Q ηη Diagonal Matrix of the Derivatives of g The Learning Rate

25 Training: Epoch

26 Training: Overfitting

27 Homework Training Result Processing Result Reference Processing

28 Homework Simulation of Material Composition/Decomposition Training Data Set Testing Data Set

29 Homework Error Plot Training Results References Testing Results + REVIEW

30 Into the Deep Learning Vanishing Gradient Problem The deeper? The harder to train! Rectified Linear Unit (ReLU) Prevent overfitting A burning issue Data Augmentation Drop out Regularization Deep Learning Algorithms Convolutional Neural Networks (CNN): Pattern recognition, Classification Recurrent Neural Networks (RNN): Sequence data processing, Translation