CSC321 Lecture 19: Generative Adversarial Networks

Similar documents
Generative models and adversarial training

Lecture 1: Machine Learning Basics

(Sub)Gradient Descent

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

A Case Study: News Classification Based on Term Frequency

Speech Recognition at ICSI: Broadcast News and beyond

Artificial Neural Networks written examination

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

YMCA SCHOOL AGE CHILD CARE PROGRAM PLAN

Calibration of Confidence Measures in Speech Recognition

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Python Machine Learning

WHEN THERE IS A mismatch between the acoustic

Enduring Understandings: Students will understand that

Feature-oriented vs. Needs-oriented Product Access for Non-Expert Online Shoppers

Mathematics process categories

Truth Inference in Crowdsourcing: Is the Problem Solved?

Case study Norway case 1

Functional Skills Mathematics Level 2 assessment

What is a Mental Model?

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

The Strong Minimalist Thesis and Bounded Optimality

12- A whirlwind tour of statistics

Knowledge Transfer in Deep Convolutional Neural Nets

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Software Maintenance

Individual Differences & Item Effects: How to test them, & how to test them well

Short vs. Extended Answer Questions in Computer Science Exams

METHODS FOR EXTRACTING AND CLASSIFYING PAIRS OF COGNATES AND FALSE FRIENDS

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

Evolutive Neural Net Fuzzy Filtering: Basic Description

Taxonomy of the cognitive domain: An example of architectural education program

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Speech Emotion Recognition Using Support Vector Machine

End-of-Module Assessment Task K 2

Session 2B From understanding perspectives to informing public policy the potential and challenges for Q findings to inform survey design

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

INPE São José dos Campos

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Training Pack. Kaizen Focused Improvement Teams (F.I.T.)

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

new research in learning and working

An Empirical and Computational Test of Linguistic Relativity

Deep Facial Action Unit Recognition from Partially Labeled Data

How People Learn Physics

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

arxiv: v2 [stat.ml] 30 Apr 2016 ABSTRACT

The Effect of Income on Educational Attainment: Evidence from State Earned Income Tax Credit Expansions

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

CONSTRUCTION OF AN ACHIEVEMENT TEST Introduction One of the important duties of a teacher is to observe the student in the classroom, laboratory and

Corpus Linguistics (L615)

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Seminar - Organic Computing

Modeling function word errors in DNN-HMM based LVCSR systems

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Model Ensemble for Click Prediction in Bing Search Ads

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

Geo Risk Scan Getting grips on geotechnical risks

SARDNET: A Self-Organizing Feature Map for Sequences

A Game-based Assessment of Children s Choices to Seek Feedback and to Revise

Human Emotion Recognition From Speech

arxiv: v1 [math.at] 10 Jan 2016

Essentials of Ability Testing. Joni Lakin Assistant Professor Educational Foundations, Leadership, and Technology

Word Segmentation of Off-line Handwritten Documents

Benjamin Pohl, Yves Richard, Manon Kohler, Justin Emery, Thierry Castel, Benjamin De Lapparent, Denis Thévenin, Thomas Thévenin, Julien Pergaud

arxiv: v1 [cs.lg] 15 Jun 2015

Global Television Manufacturing Industry : Trend, Profit, and Forecast Analysis Published September 2012

Modeling function word errors in DNN-HMM based LVCSR systems

Probability and Game Theory Course Syllabus

Cooperative Game Theoretic Models for Decision-Making in Contexts of Library Cooperation 1

Analysis: Evaluation: Knowledge: Comprehension: Synthesis: Application:

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Why Did My Detector Do That?!

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Semi-Supervised Face Detection

Detailed course syllabus

Extending Place Value with Whole Numbers to 1,000,000

Revisiting the role of prosody in early language acquisition. Megha Sundara UCLA Phonetics Lab

Finding, Hiring, and Directing e-learning Voices Harlan Hogan, E-learningvoices.com

Analysis of Enzyme Kinetic Data

Unit: Human Impact Differentiated (Tiered) Task How Does Human Activity Impact Soil Erosion?

MULTIMEDIA Motion Graphics for Multimedia

LEGO MINDSTORMS Education EV3 Coding Activities

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Assignment 1: Predicting Amazon Review Ratings

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

A Latent Semantic Model with Convolutional-Pooling Structure for Information Retrieval

Secondary English-Language Arts

Extending Learning Across Time & Space: The Power of Generalization

Rover Races Grades: 3-5 Prep Time: ~45 Minutes Lesson Time: ~105 minutes

SEMAFOR: Frame Argument Resolution with Log-Linear Models

Operational Knowledge Management: a way to manage competence

SOCIAL STUDIES GRADE 1. Clear Learning Targets Office of Teaching and Learning Curriculum Division FAMILIES NOW AND LONG AGO, NEAR AND FAR

What to Do When Conflict Happens

Transcription:

CSC321 Lecture 19: Generative Adversarial Networks Roger Grosse Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 1 / 25

Overview In generative modeling, we d like to train a network that models a distribution, such as a distribution over images. One way to judge the quality of the model is to sample from it. This field has seen rapid progress: 2009 Roger Grosse 2015 CSC321 Lecture 19: Generative Adversarial Networks 2018 2 / 25

Overview Four modern approaches to generative modeling: Generative adversarial networks (today) Reversible architectures (next lecture) Autoregressive models (Lecture 7, and next lecture) Variational autoencoders (CSC412) All four approaches have different pros and cons. Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 3 / 25

Implicit Generative Models Implicit generative models implicitly define a probability distribution Start by sampling the code vector z from a fixed, simple distribution (e.g. spherical Gaussian) The generator network computes a differentiable function G mapping z to an x in data space Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 4 / 25

Implicit Generative Models A 1-dimensional example: Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 5 / 25

Implicit Generative Models https://blog.openai.com/generative-models/ Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 6 / 25

Implicit Generative Models This sort of architecture sounded preposterous to many of us, but amazingly, it works. Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 7 / 25

Generative Adversarial Networks The advantage of implicit generative models: if you have some criterion for evaluating the quality of samples, then you can compute its gradient with respect to the network parameters, and update the network s parameters to make the sample a little better The idea behind Generative Adversarial Networks (GANs): train two different networks The generator network tries to produce realistic-looking samples The discriminator network tries to figure out whether an image came from the training set or the generator network The generator network tries to fool the discriminator network Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 8 / 25

Generative Adversarial Networks Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 9 / 25

Generative Adversarial Networks Let D denote the discriminator s predicted probability of being data Discriminator s cost function: cross-entropy loss for task of classifying real vs. fake images J D = E x D [ log D(x)] + E z [ log(1 D(G(z)))] One possible cost function for the generator: the opposite of the discriminator s J G = J D = const + E z [log(1 D(G(z)))] This is called the minimax formulation, since the generator and discriminator are playing a zero-sum game against each other: max G min J D D Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 10 / 25

Generative Adversarial Networks Updating the discriminator: Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 11 / 25

Generative Adversarial Networks Updating the generator: Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 12 / 25

Generative Adversarial Networks Alternating training of the generator and discriminator: Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 13 / 25

A Better Cost Function We introduced the minimax cost function for the generator: J G = E z [log(1 D(G(z)))] One problem with this is saturation. Recall from our lecture on classification: when the prediction is really wrong, Logistic + squared error gets a weak gradient signal Logistic + cross-entropy gets a strong gradient signal Here, if the generated sample is really bad, the discriminator s prediction is close to 0, and the generator s cost is flat. Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 14 / 25

A Better Cost Function Original minimax cost: J G = E z [log(1 D(G(z)))] Modified generator cost: J G = E z [ log D(G(z))] This fixes the saturation problem. Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 15 / 25

Generative Adversarial Networks Since GANs were introduced in 2014, there have been hundreds of papers introducing various architectures and training methods. Most modern architectures are based on the Deep Convolutional GAN (DC-GAN), where the generator and discriminator are both conv nets. GAN Zoo: https://github.com/hindupuravinash/the-gan-zoo Good source of horrible puns (VEEGAN, Checkhov GAN, etc.) Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 16 / 25

GAN Samples Celebrities: Karras et al., 2017. Progressive growing of GANs for improved quality, stability, and variation Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 17 / 25

GAN Samples Bedrooms: Karras et al., 2017. Progressive growing of GANs for improved quality, stability, and variation Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 18 / 25

GAN Samples Objects: Karras et al., 2017. Progressive growing of GANs for improved quality, stability, and variation Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 19 / 25

GAN Samples GANs revolutionized generative modeling by producing crisp, high-resolution images. The catch: we don t know how well they re modeling the distribution. Can t measure the log-likelihood they assign to held-out data. Could they be memorizing training examples? (E.g., maybe they sometimes produce photos of real celebrities?) We have no way to tell if they are dropping important modes from the distribution. See Wu et al., On the quantitative analysis of decoder-based generative models for partial answers to these questions. Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 20 / 25

CycleGAN Style transfer problem: change the style of an image while preserving the content. Data: Two unrelated collections of images, one for each style Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 21 / 25

CycleGAN If we had paired data (same content in both styles), this would be a supervised learning problem. But this is hard to find. The CycleGAN architecture learns to do it from unpaired data. Train two different generator nets to go from style 1 to style 2, and vice versa. Make sure the generated samples of style 2 are indistinguishable from real images by a discriminator net. Make sure the generators are cycle-consistent: mapping from style 1 to style 2 and back again should give you almost the original image. Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 22 / 25

CycleGAN Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 23 / 25

CycleGAN Style transfer between aerial photos and maps: Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 24 / 25

CycleGAN Style transfer between road scenes and semantic segmentations (labels of every pixel in an image by object category): Roger Grosse CSC321 Lecture 19: Generative Adversarial Networks 25 / 25