Probabilistic Graphical Models and Their Applications

Similar documents
Lecture 1: Machine Learning Basics

Python Machine Learning

(Sub)Gradient Descent

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Generative models and adversarial training

Lecture 10: Reinforcement Learning

CSL465/603 - Machine Learning

Exploration. CS : Deep Reinforcement Learning Sergey Levine

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

arxiv: v2 [cs.cv] 30 Mar 2017

Semi-Supervised Face Detection

Lecture 1: Basic Concepts of Machine Learning

Probability and Statistics Curriculum Pacing Guide

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Learning From the Past with Experiment Databases

CS Machine Learning

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Calibration of Confidence Measures in Speech Recognition

Artificial Neural Networks written examination

Rule Learning With Negation: Issues Regarding Effectiveness

Probabilistic Latent Semantic Analysis

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Truth Inference in Crowdsourcing: Is the Problem Solved?

Axiom 2013 Team Description Paper

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Switchboard Language Model Improvement with Conversational Data from Gigaword

Australian Journal of Basic and Applied Sciences

CS 446: Machine Learning

Speech Recognition at ICSI: Broadcast News and beyond

STA 225: Introductory Statistics (CT)

Shockwheat. Statistics 1, Activity 1

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Evolutive Neural Net Fuzzy Filtering: Basic Description

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Rule Learning with Negation: Issues Regarding Effectiveness

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Speech Emotion Recognition Using Support Vector Machine

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Assignment 1: Predicting Amazon Review Ratings

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Human Emotion Recognition From Speech

Math 96: Intermediate Algebra in Context

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

BMBF Project ROBUKOM: Robust Communication Networks

Comparison of network inference packages and methods for multiple networks inference

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Model Ensemble for Click Prediction in Bing Search Ads

The Evolution of Random Phenomena

Learning to Rank with Selection Bias in Personal Search

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Learning Methods in Multilingual Speech Recognition

WHEN THERE IS A mismatch between the acoustic

A NEW ALGORITHM FOR GENERATION OF DECISION TREES

Word learning as Bayesian inference

Statewide Framework Document for:

Machine Learning and Development Policy

Word Segmentation of Off-line Handwritten Documents

A survey of multi-view machine learning

Applications of data mining algorithms to analysis of medical data

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Interactive Whiteboard

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

INPE São José dos Campos

Deep Facial Action Unit Recognition from Partially Labeled Data

A study of speaker adaptation for DNN-based speech synthesis

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Radius STEM Readiness TM

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

MGT/MGP/MGB 261: Investment Analysis

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

arxiv: v1 [cs.lg] 15 Jun 2015

An OO Framework for building Intelligence and Learning properties in Software Agents

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Software Maintenance

stateorvalue to each variable in a given set. We use p(x = xjy = y) (or p(xjy) as a shorthand) to denote the probability that X = x given Y = y. We al

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Discriminative Learning of Beam-Search Heuristics for Planning

Introduction to Causal Inference. Problem Set 1. Required Problems

Learning Methods for Fuzzy Systems

Edexcel GCSE. Statistics 1389 Paper 1H. June Mark Scheme. Statistics Edexcel GCSE

Issues in the Mining of Heart Failure Datasets

Indian Institute of Technology, Kanpur

IT Students Workshop within Strategic Partnership of Leibniz University and Peter the Great St. Petersburg Polytechnic University

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Reinforcement Learning by Comparing Immediate Reward

Welcome to. ECML/PKDD 2004 Community meeting

arxiv: v1 [cs.cv] 10 May 2017

On-the-Fly Customization of Automated Essay Scoring

Laboratorio di Intelligenza Artificiale e Robotica

Transcription:

Probabilistic Graphical Models and Their Applications Bjoern Andres and Bernt Schiele Max Planck Institute for Informatics slides adapted from Peter Gehler October 26, 2016 Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 1 / 69

Organization Intro Lecture 2 hours/week Wed: 14:00 16:00, Room: E1.4 024 Exercises 2 hours/week Thu: 10:00 12:00, Room E1.4 024 Starts next Thursday Course web page: http://www.d2.mpi-inf.mpg.de/gm Slides Pointers to Books and Papers Homework assignments Semesterapparat in library Mailing list: see webpage how to subscribe Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 2 / 69

Exercises & Exam Intro Exercises: Typically one assignment per week Typically from Wednesday Wednesday Theoretical and Practical Exercises Starts this Thursday with Matlab primer Final Grade: 50% exercises, 50% oral exam (oral exam has to be passed obviously!) Exam Oral exam at the end of the semester Can be taken in English or German Tutors Eldar Insafutdinov (eldar@mpi-inf.mpg.de) Evgeny Levinkov (levinkov@mpi-inf.mpg.de) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 3 / 69

Intro Related Classes @UdS High-Level Computer Vision (SS), Fritz & Schiele Machine Learning (WS), Hein Statistical Learning I+II (SS,WS), Lengauer Optimization I+II, Convex Optimization (SS,WS),... Pattern and Speech Recognition (WS), Klakow Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 4 / 69

Intro Offers in our Research Group Master- and Bachelor Theses HiWi-positions, etc. in Topics in machine learning Topics in computer vision Topics in machine learning applied to computer vision Come, talk to us Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 5 / 69

Literature Intro All books in a Semesterapparat Main book for the graphical model part Barber, Bayesian Reasoning and Machine Learning, Cambridge University Press, 2011, ISBN-13: 978-0521518147, http://tinyurl.com/3flppuo Extra References Bishop, Pattern Recognition and Machine Learning, Springer New York, 2006, ISBN-13: 978-0387310732 Koller, Friedman, Probabilistic Graphical Models: Principles and Techniques, The MIT Press, 2009, ISBN-13: 978-0262013192 MacKay, Information Theory, Inference and Learning Algorithms, Cambridge Universsity Press, 2003, ISBN-13: 978-0521642989 Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 6 / 69

Literature Intro Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 7 / 69

Topic overview 2016/17 Intro Recap: Probability and Decision theory (today) Graphical Models Basics (Directed, Undirected, Factor Graphs) Inference Learning Inference Deterministic Inference (Sum-Prodcut, Junction Tree) Approximate Inference (Loopy BP, Sampling, Variational) Application to Computer Vision Problems Body Pose Estimation Object Detection Semantic Segmentation Image Denoising... Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 9 / 69

Today s topics Intro Overview: Machine Learning What is machine learning? Different problem settings and examples Probability theory Decision theory, inference and decision Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 10 / 69

Machine Learning Machine Learning Overview Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 11 / 69

Machine Learning Machine learning what s that? Do you use machine learning systems already? Can you think of an application? Can you define the term machine learning? Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 12 / 69

Machine Learning Goal of machine learning: Machines that learn to perform a task from experience We can formalize this as y = f(x; w) (1) y is called output variable, x the input variable and w the model parameters (typically learned) Classification vs regression: regression: y continuous classification: y discrete (e.g. class membership) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 13 / 69

Machine Learning Goal of machine learning: Machines that learn to perform a task from experience We can formalize this as y is called output variable, x the input variable and w the model parameters (typically learned) learn... adjust the parameter w... a task... the function f y = f(x; w) (2)... from experience using a training dataset D, where of either D = {x 1,..., x n } or D = {(x 1, y 1 ),..., (x n, y n )} Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 14 / 69

Machine Learning Different Scenarios Unsupervised Learning Supervised Learning Reinforcement Learning Let s discuss Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 15 / 69

Machine Learning Supervised Learning Given are pairs of training examples from X Y D = {(x 1, y 1 ), (x 2, y 2 ),..., (x n, y n )} (3) Goal is to learn the relationship between x and y Given a new example point x predict y y = f(x; w) (4) We want to generalize to unseen data Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 16 / 69

Machine Learning Supervised Learning Examples Face Detection Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 17 / 69

Machine Learning Supervised Learning Examples Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 18 / 69

Machine Learning Supervised Learning Examples Semantic Image Segmentation Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 19 / 69

Machine Learning Supervised Learning Examples Body Part Estimation (in Kinect) Figure from Decision Tree Fields, Nowozin et al., ICCV11 Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 20 / 69

Machine Learning Supervised Learning Examples Person identification Credit card fraud detection Industrial inspection Speech recognition Action classification in videos Human body pose estimation Visual object detection Prediction survival rate of a patient... Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 21 / 69

Machine Learning Supervised Learning - Models Flashing more keywords Multilayer Perceptron (Backpropagation) (Deep) Convolutional Neural Networks (Backpropagation) Linear Regression, Logistic Regression Support Vector Machine (SVM) Boosting Graphical models Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 22 / 69

Unsupervised Learning Machine Learning We are given some input data points D = {x 1, x 2,..., x n } (5) Goals: Determine the data distribution p(x) density estimation Visualize the data by projections dimensionality reduction Find groupings of the data clustering Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 23 / 69

Machine Learning Unsupervised Learning Examples Image Priors for Denoising Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 24 / 69

Machine Learning Unsupervised Learning Examples Image Priors for Inpainting Image from A generative perspective on MRFs in low-level vision, Schmidt et al., CVPR2010 black line: statistics form original images, blue and red: statistics after applying two different algorithms Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 25 / 69

Machine Learning Unsupervised Learning Examples Human Shape Model SCAPE: Shape Completion and Animation of People, Anguelov et al. Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 26 / 69

Machine Learning Unsupervised Learning Examples Clustering scientific publications according to topics A generative model for human motion Generating training data for Microsoft Kinect xbox controller Clustering flickr images Novelty detection, predicting outliers Anomality detection in visual inspection Video surveillance Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 27 / 69

Machine Learning Unsupervised Learning Models Just flashing some keywords ( Machine Learning) Mixture Models Neural Networks K-Means Kernel Density Estimation Principal Component Analysis (PCA) Graphical Models (here) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 28 / 69

Reinforcement Learning Machine Learning Setting: given a situation, find an action to maximize a reward function Feedback: we only get feedback of how well we are doing we do not get feedback what the best action would be ( indirect teaching ) Feedback given as reward: each action yields reward, or a reward is given at the end (e.g. robot has found his goal, computer has won game in Backgammon) Exploration: try out new actions Exploitation: use known actions that yield high rewards Find a good trade-off between exploration and exploitation Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 29 / 69

Machine Learning Variations of the general theme All problems fall in these broad categories But your problem will surely have some extra twists Many different variations of the aforementioned problems are studied separately Let s look at some... Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 30 / 69

Machine Learning Semi-Supervised Learning We are given a dataset of l labeled examples D l = {(x 1, y 1 ),..., (x l, y l )} as in supervised learning Additionally we are given a set of u unlabeled examples D u = {x l+1,..., x l+u } as in unsupervised learning Goal is y = f(x; w) Question: how can we utilize the extra information in D u? Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 31 / 69

Machine Learning Semi-Supervised Learning: Two Moons Two labeled examples (red and blue) and additional unlabeled black dots Two moons Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 32 / 69

Machine Learning Transductive Learning We are given a set of labeled examples D = {(x 1, y 1 ),..., (x n, y n )} (6) Additionally we know the test data points {x te 1,..., xte m} (not their labels!) Can we do better, including this knowledge? This should be easier than making predictions for the entire set X Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 33 / 69

On-line Learning Machine Learning The training data is presented step-by-step and is never available entirely At each time-step t we are given a new datapoint x t (or (x t, y t )) When is online learning a sensible scenario? We want to continuously update the model we can train a model with little data, but the model should become better over time when more data is available (similar to how humans learn) We have limited storage for data and the model a viable setting for large-scale datasets (e.g. the size of the internet) How do we learn in this scenario? Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 34 / 69

Machine Learning Large-Scale Learning Learning with millions of examples Study fast learning algorithms (e.g. parallelizable, special hardware) Problems of storing the data, computing the features, etc. There is no strict definition for large-scale Small-scale learning: limiting factor is number of examples Large-scale learning: limited by maximal time for computation (and/or maximal storage capacity) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 35 / 69

Machine Learning Active Learning We are given a set of examples Goal is to learn y = f(x; w) Each label y i costs something, e.g. C i R + Question: How to learn well while paying little? D = {x 1,..., x n } (7) This is almost always the case, labeling is expensive Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 36 / 69

Machine Learning Structured Output Learning We are given a set of training examples D = {(x 1, y 1 ),..., (x n, y n )}, but y Y contains more structure than y R or y { 1, 1} Consider binary image segmentation y is entire image labeling Y is the set of all labelings 2 #pixels Other examples: y could be a graph, a tree, a ranking,... Goal is to learn a function f(x, y; w) and predict y = argmax f(x, ȳ; w) ȳ Y Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 37 / 69

Machine Learning Some final comments All topics are under active development and research Supervised classification: basically understood Broad range of applications, many exciting developments Adopting a ML view has far reaching consequences, it touches problems of empirical sciences in general Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 38 / 69

Probability Theory Probability Theory Brief Review Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 39 / 69

Probability Theory Brief Review A random variable (RV) X can take values from some discrete set of outcomes X. We usually use the short-hand notation for the probability that X takes value x With p(x) for p(x = x) [0, 1] (8) we denote the probability distribution over X p(x), (9) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 40 / 69

Brief Review Probability Theory Two random variables (RVs) are called independent if p(x = x, Y = y) = p(x = x)p(y = y) (10) Joint probability (of X and Y ) p(x, y) instead p(x = x, Y = y) (11) Conditional probability p(x y) instead p(x = x Y = y) (12) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 41 / 69

Probability Theory The Rules of Probability Sum rule p(x) = y Y p(x, Y = y) (13) we marginalize out y. p(x = x) is also called a marginal probability Product Rule p(x, Y ) = p(y X)p(X) (14) And as a consequence: Bayes Theorem or Bayes Rule p(y X) = p(x Y )p(y ) p(x) (15) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 42 / 69

Vocabulary Probability Theory Joint Probability p(x i, y j ) = n ij N Marginal Probability p(x i ) = c i N Conditional Probability y j c i = n ij j }{{} n ij x i p(y j x i ) = n ij c i N = ij n ij Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 43 / 69

Probability Theory Probability Densities Now X is a continuous random variable, eg taking values in R Probability that X takes a value in the interval (a, b) is p(x (a, b)) = b and we call p(x) the probability density over x a p(x)dx (16) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 44 / 69

Probability Densities Probability Theory p(x) must satisfy the following conditions p(x) 0 (17) p(x)dx = 1 (18) The probability that x lies in (, z) is given by the cumulative distribution function P (z) = z p(x)dx (19) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 45 / 69

Probability Densities Probability Theory Figure : Probability density of a continuous variable Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 46 / 69

Probability Theory Expectation and Variances Expectation E[f] = x X E[f] = x X p(x)f(x) (20) p(x)f(x)dx (21) Sometimes we denote the distribution that we take the expectation over as a subscript, eg E p( y) [f] = x X p(x y)f(x) (22) Variance [ var[f] = E (f(x) E [f(x)]) 2] (23) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 47 / 69

Decision Theory Decision Theory Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 48 / 69

Digit Classification Decision Theory Classify digits a versus b Figure : The digits a and b Goal: classify new digits such that the error probability is minimized Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 49 / 69

Decision Theory Digit Classification - Priors Prior Distribution How often do the letters a and b occur? Let us assume C 1 = a p(c 1 ) = 0.75 (24) C 2 = b p(c 2 ) = 0.25 (25) The prior has to be a distribution, in particular p(c k ) = 1 (26) k=1,2 Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 50 / 69

Decision Theory Digit Classification - Class Conditionals We describe every digit using some feature vector the number of black pixels in each box relation between width and height Likelihood: How likely has x been generated from p( a), resp. p( b)? Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 51 / 69

Decision Theory Digit Classification Which class should we assign x to? The answer Class a Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 52 / 69

Decision Theory Digit Classification Which class should we assign x to? The answer Class b Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 53 / 69

Decision Theory Digit Classification Which class should we assign x to? The answer Class a, since p(a)=0.75 Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 54 / 69

Decision Theory Bayes Theorem How do we formalize this? We already mentioned Bayes Theorem p(y X) = p(x Y )p(y ) p(x) (27) Now we apply it p(c k x) = p(x C k)p(c k ) p(x) = p(x C k)p(c k ) j p(x C j)p(c j ) (28) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 55 / 69

Decision Theory Bayes Theorem Some terminology! Repeated from last slide: p(c k x) = p(x C k)p(c k ) p(x) We use the following names = p(x C k)p(c k ) j p(x C j)p(c j ) (29) Posterior = Likelihood Prior Normalization Factor (30) Here the normalization factor is easy to compute. Keep an eye out for it, it will haunt us until the end of this class (and longer :) ) It is also called the Partition Function, common symbol Z Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 56 / 69

Bayes Theorem Decision Theory Likelihood Likelihood Prior Posterior = Likelihood Prior Normalization Factor Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 57 / 69

How to Decide? Decision Theory Two class problem C 1, C 2, plotting Likelihood Prior Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 58 / 69

Minmizing the Error Decision Theory p(error) = p(x R 2, C 1 ) + p(x R 1, C 2 ) (31) = p(x R 2 C 1 )p(c 1 ) + p(x R 1 C 2 )p(c 2 ) (32) = p(x C 1 )p(c 1 )dx + R 2 p(x C 2 )p(c 2 )dx R 1 (33) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 59 / 69

Decision Theory General Loss Functions So far we considered misclassification error only This is also referred to as 0/1 loss Now suppose we are given a more general loss function : Y Y R + (34) (y, ŷ) (y, ŷ) (35) How do we read this? (y, ŷ) is the cost we have to pay if y is the true class but we predict ŷ instead Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 60 / 69

Decision Theory Example: Predicting Cancer : Y Y R + (36) (y, ŷ) (y, ŷ) (37) Given: X-Ray image, Question: Cancer yes or no? Should we have another medical check of the patient? diagnosis : cancer normal truth : cancer 0 1000 normal 1 0 For discrete sets Y this is a loss matrix Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 61 / 69

Decision Theory Digit Classification Which class should we assign x to? (p(a) = p(b) = 0.5) The answer It depends on the loss Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 62 / 69

Decision Theory Minmizing Expected Loss (or Error) The expected loss for x (averaged over all decisions) E[ ] = k=1,...,k j=1,...,k R j (C k, C j )p(x, C k )dx (38) And how do we predict? Decide on one y! y = argmin y Y (C k, y)p(c k x) (39) k=1,...,k = argmin E p( x) [ (, y)] (40) y Y Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 63 / 69

Inference and Decision Decision Theory We broke down the process into two steps Inference: obtaining the probabilities p(ck x) Decision: Obtain optimal class assignment Two steps!! The probabilites p( x) represent our belief of the world The loss tells us what to do with it! 0/1 loss implies deciding for max probability (exercise) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 64 / 69

Decision Theory Three Approaches to Solve Decision Problems 1. Generative models: infer the class conditionals p(x C k ), k = 1,..., K (41) then combine using Bayes Theorem p(c k x) = p(x C k)p(c k ) p(x) 2. Discriminative models: infer posterior probabilities directly p(c k x) (42) 3. Find a discriminative function minimizing Expected Loss f : X {1,..., K} (43) Let s discuss these options Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 65 / 69

Decision Theory Generative Models Pros: The name generative is because we can generate samples from the learnt distribution We can infer p(x C k ) (or p(x) for short) Cons: With high dimensionality of x X we need a large training set to determine the class-conditionals We may not be interested in all quantities Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 66 / 69

Decision Theory Discriminative Models Pros: No need to model p(x C k ) (i.e. in general easier) Cons: No access to model p(x C k ) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 67 / 69

Decision Theory Discriminative Functions When solving a problem of interest, do not solve a harder / more general problem as an intermediate step. Vladimir Vapnik Pros: One integrated system, we directly estimate the quantity of interest Cons: Need during training time revision requires re-learning No access to probabilities or uncertainty, thus difficult to reject decision? Prominent example: Support Vector Machines (SVMs) Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 68 / 69

Next Time... Decision Theory... we will meet our new friends: Andres & Schiele (MPII) Probabilistic Graphical Models October 26, 2016 69 / 69