Deep (Structured) Learning

Similar documents
Python Machine Learning

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Knowledge Transfer in Deep Convolutional Neural Nets

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

(Sub)Gradient Descent

A Review: Speech Recognition with Deep Learning Methods

arxiv: v1 [cs.lg] 15 Jun 2015

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Artificial Neural Networks written examination

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

A study of speaker adaptation for DNN-based speech synthesis

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Speech Emotion Recognition Using Support Vector Machine

Calibration of Confidence Measures in Speech Recognition

Lecture 1: Machine Learning Basics

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

CSL465/603 - Machine Learning

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

INPE São José dos Campos

arxiv: v2 [cs.cv] 30 Mar 2017

Human Emotion Recognition From Speech

Attributed Social Network Embedding

Dropout improves Recurrent Neural Networks for Handwriting Recognition

Evolutive Neural Net Fuzzy Filtering: Basic Description

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

A Deep Bag-of-Features Model for Music Auto-Tagging

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Axiom 2013 Team Description Paper

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

Model Ensemble for Click Prediction in Bing Search Ads

A Latent Semantic Model with Convolutional-Pooling Structure for Information Retrieval

arxiv: v1 [cs.lg] 7 Apr 2015

Generative models and adversarial training

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

Artificial Neural Networks

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Test Effort Estimation Using Neural Network

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Learning Methods for Fuzzy Systems

Softprop: Softmax Neural Network Backpropagation Learning

A Compact DNN: Approaching GoogLeNet-Level Accuracy of Classification and Domain Adaptation

Knowledge-Based - Systems

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

Learning Methods in Multilingual Speech Recognition

Australian Journal of Basic and Applied Sciences

A deep architecture for non-projective dependency parsing

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Speaker Identification by Comparison of Smart Methods. Abstract

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Segmental Conditional Random Fields with Deep Neural Networks as Acoustic Models for First-Pass Word Recognition

CS Machine Learning

Deep Facial Action Unit Recognition from Partially Labeled Data

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Deep Neural Network Language Models

arxiv: v4 [cs.cl] 28 Mar 2016

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Lecture 10: Reinforcement Learning

arxiv: v1 [cs.cv] 10 May 2017

Dialog-based Language Learning

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

Classification Using ANN: A Review

Word Segmentation of Off-line Handwritten Documents

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

An empirical study of learning speed in backpropagation

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

THE world surrounding us involves multiple modalities

arxiv: v1 [cs.cl] 27 Apr 2016

How to read a Paper ISMLL. Dr. Josif Grabocka, Carlotta Schatten

Soft Computing based Learning for Cognitive Radio

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

arxiv: v2 [cs.ir] 22 Aug 2016

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

DNN ACOUSTIC MODELING WITH MODULAR MULTI-LINGUAL FEATURE EXTRACTION NETWORKS

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

Master s Programme in Computer, Communication and Information Sciences, Study guide , ELEC Majors

Kamaldeep Kaur University School of Information Technology GGS Indraprastha University Delhi

arxiv: v1 [cs.cl] 2 Apr 2017

Second Exam: Natural Language Parsing with Neural Networks

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

WHEN THERE IS A mismatch between the acoustic

Lip Reading in Profile

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

The A2iA Multi-lingual Text Recognition System at the second Maurdor Evaluation

Transcription:

Deep (Structured) Learning Yasmine Badr 06/23/2015 NanoCAD Lab UCLA

What is Deep Learning? [1] A wide class of machine learning techniques and architectures Using many layers of non-linear information processing that are hierarchical in nature. Inspired from Human information processing mechanisms Example: human speech production and perception systems are both equipped with clearly layered hierarchical structures in transforming the information from the waveform level to the linguistic level 2

General Challenges of Deep Learning Pervasive presence of local optima in non-convex objective function in deep networks Great computational demand due to the size of the deep network Traditional Back-Propagation used for training the nondeep ANN do not work well, beyond few hidden layers 3

3 Categories of Deep Learning Models 1. Deep supervised models 2. Deep unsupervised model Discriminative deep networks 3. Hybrid models Use unsupervised learning models to improve the training of a supervised learning models 4

Deep Learning today.. Google Microsoft Voice translator: english to chinese in realtime Facebook 5

Some Deep Learning Architectures Deep neural networks [2] An ANN with multiple hidden layers of units between the input and output layers. Extra layers enable modeling complex data with fewer units than a similarly performing shallow network 6

Some Deep Learning Architectures Deep neural networks (cont d) Can be trained using Back propagation or gradient descent with: Mini-batching to enhance computation time computing the gradient on several training examples at once rather than individual examples A lot of learning algorithms to find initial weights Because sweeping space is not practical Techniques to prevent overfitting L1 regularization in training to enforce sparsity Dropout regularization: randomly dropping units from hidden layers during training, to break rare dependencies that can occur in training data 7

Some Deep Learning Architectures Convolutional Neural Networks [3][4] Very good results in speech and image recognition Designed to take advantage of the 2D structure of an input image (or other 2D input such as a speech signal). Achieved with local connections and tied weights followed by some form of pooling which results in translation invariant features. 8

Some Deep Learning Architectures Convolutional Neural Networks [3][4] (cont d) There are at least 3 types of layers: convolution, pooling and fully connected layers. There can be more than one layer of each of these types. Convolution layer: Each neurons in this layer processes (as input) a certain region in the image. Result is a convolution operation. The convolution kernel is actually learnt as part of the weight learning process.» For e.g. some neurons will learn a certain edge detection convolution kernel. Also Multiple neurons process each region, which leads to several filters applied on each region of the image. Usually weights are shared between neurons which process different regions, for example the same "learnt" edge detection kernel will be applied to all the regions in the image. 9

Some Deep Learning Architectures Convolutional Neural Networks [3][4] (cont d) Pooling layer: it subsamples the result of the previous layer, for example it can perform a max operation on every 2 neighboring values in the results of the convolution (subsampling). Objective is to reduce the size of the representation and achieve translation invariance. This is a fixed layer and does not need training, because it performs a fixed function (e.g average or max) Fully connected layer: this is the normal layer that we know from traditional neural networks. 10

Some Deep Learning Architectures Convolution Neural Networks [3] (cont d) Easier to train and have many fewer parameters than fully connected networks with the same number of hidden units Image source: h?p://parse.ele.tue.nl/educaaon/cluster2 11

References [1] Li Deng and Dong Yu; Deep Learning Methods and Applications; Foundations and Trends in Signal Processing; Vol.7; 2014 [2] https://en.wikipedia.org/wiki/deep_learning [3] http://ufldl.stanford.edu/tutorial/supervised/ ConvolutionalNeuralNetwork/ [4] http://cs231n.github.io/convolutional-networks/ 12