Generative models and adversarial training

Similar documents
Lecture 1: Machine Learning Basics

Python Machine Learning

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Artificial Neural Networks written examination

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

arxiv: v1 [cs.lg] 15 Jun 2015

arxiv: v2 [stat.ml] 30 Apr 2016 ABSTRACT

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Speech Recognition at ICSI: Broadcast News and beyond

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

(Sub)Gradient Descent

Semi-Supervised Face Detection

arxiv: v2 [cs.cv] 30 Mar 2017

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

While you are waiting... socrative.com, room number SIMLANG2016

Lecture 10: Reinforcement Learning

Rule Learning With Negation: Issues Regarding Effectiveness

Human Emotion Recognition From Speech

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

INPE São José dos Campos

A study of speaker adaptation for DNN-based speech synthesis

CS Machine Learning

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

A Case Study: News Classification Based on Term Frequency

Taxonomy-Regularized Semantic Deep Convolutional Neural Networks

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Assignment 1: Predicting Amazon Review Ratings

Action Recognition and Video

Comment-based Multi-View Clustering of Web 2.0 Items

Multi-Dimensional, Multi-Level, and Multi-Timepoint Item Response Modeling.

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Rule Learning with Negation: Issues Regarding Effectiveness

Learning Methods in Multilingual Speech Recognition

WHEN THERE IS A mismatch between the acoustic

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

Automatic Speaker Recognition: Modelling, Feature Extraction and Effects of Clinical Environment

FF+FPG: Guiding a Policy-Gradient Planner

Georgetown University at TREC 2017 Dynamic Domain Track

Deep Facial Action Unit Recognition from Partially Labeled Data

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

Unsupervised Acoustic Model Training for Simultaneous Lecture Translation in Incremental and Batch Mode

Modeling function word errors in DNN-HMM based LVCSR systems

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

CSL465/603 - Machine Learning

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Support Vector Machines for Speaker and Language Recognition

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Digital Signal Processing: Speaker Recognition Final Report (Complete Version)

arxiv: v1 [cs.cv] 10 May 2017

Evolutive Neural Net Fuzzy Filtering: Basic Description

Introduction to Simulation

Software Maintenance

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Reinforcement Learning by Comparing Immediate Reward

Using focal point learning to improve human machine tacit coordination

Test Effort Estimation Using Neural Network

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Australian Journal of Basic and Applied Sciences

arxiv:submit/ [cs.cv] 2 Aug 2017

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Modeling function word errors in DNN-HMM based LVCSR systems

A Neural Network GUI Tested on Text-To-Phoneme Mapping

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Reducing Features to Improve Bug Prediction

Model Ensemble for Click Prediction in Bing Search Ads

Knowledge Transfer in Deep Convolutional Neural Nets

Switchboard Language Model Improvement with Conversational Data from Gigaword

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

Speech Emotion Recognition Using Support Vector Machine

Why Did My Detector Do That?!

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

Calibration of Confidence Measures in Speech Recognition

Softprop: Softmax Neural Network Backpropagation Learning

Computerized Adaptive Psychological Testing A Personalisation Perspective

arxiv: v4 [cs.cv] 13 Aug 2017

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Speaker recognition using universal background model on YOHO database

arxiv: v1 [cs.cv] 2 Jun 2017

Axiom 2013 Team Description Paper

Combining Proactive and Reactive Predictions for Data Streams

Acquiring Competence from Performance Data

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Truth Inference in Crowdsourcing: Is the Problem Solved?

Word Segmentation of Off-line Handwritten Documents

Active Learning. Yingyu Liang Computer Sciences 760 Fall

A Pipelined Approach for Iterative Software Process Model

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Transcription:

Day 4 Lecture 1 Generative models and adversarial training Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University

What is a generative model? A model P(X; ϴ) that we can draw samples from. P(X = x) E.g. A Gaussian Mixture Model Fitting: EM algorithm Drawing samples: x Draw sample from categorical distribution to select Gaussian Draw sample from Gaussian GMMs are not generally complex enough to draw samples of images from. x

Why are generative models important? Model the probability density of images Understanding P(X) may help us understand P(Y X) Generate novel content Generate training data for discriminative networks Artistic applications Image completion Monte-carlo estimators

Generative adversarial networks New method of training deep generative models Idea: pit a generator and a discriminator against each other Generator tries to draw samples from P(X) Discriminator tries to tell if sample came from the generator or the real world Both discriminator and generator are deep networks (differentiable functions) Can train with backprop: train discriminator for a while, then train generator, then discriminator,

Generative adversarial networks (conceptual) Sample Real Loss Latent random variable Real world images Discriminator Fake Generator Sample

The generator Deterministic mapping from a latent random vector to sample from q(x) ~ p(x) Usually a deep neural network. E.g. DCGAN:

The discriminator Parameterised function that tries to distinguish between samples from real images p(x) and generated ones q(x).... conv conv Usually a deep convolutional neural network. F F

Training GANs Alternate between training the discriminator and generator Differentiable module Sample Real Loss Latent random variable Real world images Discriminator Fake Generator Sample Differentiable module

1. 2. Fix generator weights, draw samples from both real world and generated images Train discriminator to distinguish between real world and generated images Sample Real Loss Latent random variable Real world images Discriminator Fake Generator Sample Backprop error to update discriminator weights

1. 2. 3. Fix discriminator weights Sample from generator Backprop error through discriminator to update generator weights Sample Real Loss Latent random variable Real world images Discriminator Fake Generator Sample Backprop error to update generator weights

Training GANs Iterate these two steps until convergence (which may not happen) Updating the discriminator should make it better at discriminating between real images and generated ones (discriminator improves) Updating the generator makes it better at fooling the current discriminator (generator improves) Eventually (we hope) that the generator gets so good that it is impossible for the discriminator to tell the difference between real and generated images. Discriminator accuracy = 0.5

Some examples

ImageNet Source: https://openai.com/blog/generativemodels/

CIFAR-10 Source: https://openai.com/blog/generativemodels/

Credit: Alec Radford Code on GitHub

Credit: Alec Radford Code on GitHub

Issues Known to be very difficult to train: Formulated as a game between two networks Unstable dynamics: hard to keep generator and discriminator in balance Optimization can oscillate between solutions Generator can collapse Possible to use supervised labels to help prevent this: https://arxiv.org/abs/1606. 03498

Predicting the future with adversarial training Want to train a classifier to predict the pixels in frame (t+k) from pixels in frame t. Many possible futures for same frame Using supervised classification results in blurry solutions: loss if minimized if classifier averages over possibilities when predicting. We really want a sample, not the mean Adversarial training can solve this: easy for an adversary to detect blurry frames Mathieu et al. Deep multi-scale video prediction beyond mean square error, ICLR 2016 (https://arxiv.org/abs/1511.05440)

Mathieu et al. Deep multi-scale video prediction beyond mean square error, ICLR 2016 (https://arxiv.org/abs/1511.05440)

Summary Adversarial networks pit a generator network against a discriminator (adversary) Can be trained to draw realistic sharp samples Training can be difficult: can oscillate or generator can collapse