CS 2750: Machine Learning. Other Topics. Prof. Adriana Kovashka University of Pittsburgh April 13, 2017

Similar documents
Active Learning. Yingyu Liang Computer Sciences 760 Fall

(Sub)Gradient Descent

Lecture 1: Machine Learning Basics

Laboratorio di Intelligenza Artificiale e Robotica

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Laboratorio di Intelligenza Artificiale e Robotica

Python Machine Learning

arxiv: v2 [cs.cv] 30 Mar 2017

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Speech Recognition at ICSI: Broadcast News and beyond

CS Machine Learning

Georgetown University at TREC 2017 Dynamic Domain Track

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Linking Task: Identifying authors and book titles in verbose queries

Switchboard Language Model Improvement with Conversational Data from Gigaword

Semi-Supervised Face Detection

Human Emotion Recognition From Speech

Going to School: Measuring Schooling Behaviors in GloFish

Generative models and adversarial training

CS 446: Machine Learning

Learning From the Past with Experiment Databases

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Artificial Neural Networks written examination

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Welcome to. ECML/PKDD 2004 Community meeting

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Exploration. CS : Deep Reinforcement Learning Sergey Levine

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Lecture 10: Reinforcement Learning

A Case Study: News Classification Based on Term Frequency

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

A Vector Space Approach for Aspect-Based Sentiment Analysis

A Compact DNN: Approaching GoogLeNet-Level Accuracy of Classification and Domain Adaptation

Rule Learning With Negation: Issues Regarding Effectiveness

Dialog-based Language Learning

Copyright by Sung Ju Hwang 2013

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Reinforcement Learning by Comparing Immediate Reward

arxiv:submit/ [cs.cv] 2 Aug 2017

Learning to Rank with Selection Bias in Personal Search

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

CSL465/603 - Machine Learning

The stages of event extraction

arxiv: v1 [cs.lg] 15 Jun 2015

arxiv: v1 [cs.cl] 2 Apr 2017

On the Combined Behavior of Autonomous Resource Management Agents

Attributed Social Network Embedding

Calibration of Confidence Measures in Speech Recognition

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

Webly Supervised Learning of Convolutional Networks

Probabilistic Latent Semantic Analysis

arxiv: v1 [cs.cl] 20 Jul 2015

Regret-based Reward Elicitation for Markov Decision Processes

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Deep Facial Action Unit Recognition from Partially Labeled Data

Word Segmentation of Off-line Handwritten Documents

An Effective Framework for Fast Expert Mining in Collaboration Networks: A Group-Oriented and Cost-Based Method

Diverse Concept-Level Features for Multi-Object Classification

Model Ensemble for Click Prediction in Bing Search Ads

arxiv: v4 [cs.cv] 13 Aug 2017

Assignment 1: Predicting Amazon Review Ratings

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Generating Natural-Language Video Descriptions Using Text-Mined Knowledge

Rule Learning with Negation: Issues Regarding Effectiveness

Top US Tech Talent for the Top China Tech Company

THE world surrounding us involves multiple modalities

Characteristics of Collaborative Network Models. ed. by Line Gry Knudsen

Summarizing Answers in Non-Factoid Community Question-Answering

Truth Inference in Crowdsourcing: Is the Problem Solved?

Speech Emotion Recognition Using Support Vector Machine

Exposé for a Master s Thesis

Taxonomy-Regularized Semantic Deep Convolutional Neural Networks

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Case Acquisition Strategies for Case-Based Reasoning in Real-Time Strategy Games

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

Conference Presentation

AQUA: An Ontology-Driven Question Answering System

Knowledge Transfer in Deep Convolutional Neural Nets

Axiom 2013 Team Description Paper

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Go fishing! Responsibility judgments when cooperation breaks down

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

A Comparison of Two Text Representations for Sentiment Analysis

Disambiguation of Thai Personal Name from Online News Articles

Mining Student Evolution Using Associative Classification and Clustering

Radius STEM Readiness TM

Combining Proactive and Reactive Predictions for Data Streams

Transductive Multi-view Embedding for Zero-Shot Recognition and Annotation

Paper 2. Mathematics test. Calculator allowed. First name. Last name. School KEY STAGE TIER

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Transcription:

CS 2750: Machine Learning Other Topics Prof. Adriana Kovashka University of Pittsburgh April 13, 2017

Plan for last lecture Overview of other topics and applications Reinforcement learning Active learning Domain adaptation Unsupervised feature learning using context Ranking

Reinforcement learning

Reinforcement learning So far we ve considered offline learning where we learn a model first and make predictions Reinforcement learning is a type of online learning Lies in between supervised and unsupervised

Reinforcement learning You have an agent acting in an environment, exploring possible behaviors with the intent of maximizing some reward For example, the agent wants to learn how to play some game so that it wins frequently

Reinforcement learning States Actions Rewards https://www.nervanasys.com/demystifying-deep-reinforcement-learning/

Reinforcement learning States e.g. image of board Actions up/down Rewards if won, +1, if lost, -1 http://karpathy.github.io/2016/05/31/rl/

Q-Learning https://www.nervanasys.com/demystifying-deep-reinforcement-learning/

Policy gradients Wait before updating model parameters until end of game when we know if we won or lost, use outcome as gradient to backprop Credit assignment: Which actions should I reward in case I won? Reward is given for a certain action taken in a certain state If I won, reward all actions that led to this Penalize all actions that led to a loss http://karpathy.github.io/2016/05/31/rl/

Learning to play Atari games w/ RL Games Total reward collected Mnih et al., Playing Atari with Deep Reinforcement Learning, 2013

Learning to localize objects w/ RL Caicedo and Lazebnik, Active Object Localization with Deep Reinforcement Learning, ICCV 2015

Active learning

Pool-based sampling Settles, Active Learning (Synthesis Lectures on AI and ML), 2012

Selective sampling (stream-based) Settles, Active Learning (Synthesis Lectures on AI and ML), 2012

Query synthesis Settles, Active Learning (Synthesis Lectures on AI and ML), 2012

Uncertainty sampling Settles, Active Learning (Synthesis Lectures on AI and ML), 2012

Measures of uncertainty Least confident Smallest margin highest-probability label 2 nd -highestprobability label Highest entropy Settles, Active Learning (Synthesis Lectures on AI and ML), 2012

Actively choosing sample and annotation type Kovashka et al., Actively Selecting Annotations Among Objects and Attributes, ICCV 2011

Expected entropy reduction on all data Our entropy-based selection function seeks to maximize the expected object label entropy reduction. We measure object class entropy on the labeled and unlabeled image sets: We seek maximal expected entropy reduction, which is equivalent to minimum entropy after the label addition: By predicting entropy change over all data, selection accounts for the impact of all desired interactions between labels and data. Kovashka et al., Actively Selecting Annotations Among Objects and Attributes, ICCV 2011

Object label depends on attribute labels Kovashka et al., Actively Selecting Annotations Among Objects and Attributes, ICCV 2011

Choose object or attribute label The expected entropy scores for object label and attribute label additions can be expressed as follows. Note that these two formulations are comparable since they both measure entropy of the object class. Then the best (image, label) choice can be made as: where x ranges over unlabeled images and q ranges over possible label types. Kovashka et al., Actively Selecting Annotations Among Objects and Attributes, ICCV 2011

Query by committee Settles, Active Learning (Synthesis Lectures on AI and ML), 2012

Cluster-based Settles, Active Learning (Synthesis Lectures on AI and ML), 2012

Domain adaptation

The same class looks different in different domains

Adaptive SVM Target domain: Auxiliary (source) domain: Standard SVM: Yang et al., Adapting SVM Classifiers to Data with Shifted Distributions, ICDM Workshops 2007

Adaptive SVM Adaptive SVM objective: learned on auxiliary domain with standard SVM Adaptive SVM dual problem: Adaptive SVM prediction: prediction from auxiliary Yang et al., Adapting SVM Classifiers to Data with Shifted Distributions, ICDM Workshops 2007

Personalized image search Like this but with curlier hair Allow user to whittle away irrelevant images via comparative feedback on attributes of results But different users might perceive attributes differently Kovashka et al., WhittleSearch: Image Search with Relative Attribute Feedback, CVPR 2012

Semantic visual attributes High-level descriptive properties shared by objects Human-understandable and machine-detectable Middle ground between user and system smiling large-lips metallic high heel long-hair ornaments red perspective open natural

Users perceive attributes differently Formal? User labels: 50% yes 50% no More ornamented? or User labels: 50% first 20% second 30% equally Binary attribute Relative attribute There may be valid perceptual differences within an attribute, yet existing methods assume monolithic attribute sufficient Kovashka and Grauman, Attribute Adaptation for Personalized Image Search, ICCV 2013

Learning user-specific attributes Standard approach: Crowd Vote on labels formal not formal Our idea: User formal not formal formal not formal Treat as a domain adaptation problem Adapt generic attribute model with minimal userspecific labeled examples Kovashka and Grauman, Attribute Adaptation for Personalized Image Search, ICCV 2013

Learning adapted attributes Adapting binary attribute classifiers: Given user-labeled data and generic model, learn adapted model, Yang et al., Adapting SVM Classifiers to Data with Shifted Distributions, ICDM Workshops 2007

Learning adapted attributes formal Adapted boundary Generic boundary not formal

Adapted attribute accuracy Result over all 3 datasets, 32 attributes, and 75 users Generic learns a model from the crowd (no personalization) Our method most accurately captures perceived attributes Kovashka and Grauman, Attribute Adaptation for Personalized Image Search, ICCV 2013

Domain adaptation w/ metric learning Colors = domains, shapes = classes Saenko et al., Adapting visual category models to new domains, ECCV 2010

Domain adaptation with metric learning Want to learn to relate two domains, x is from one domain, y is from the other Constraints in learned space: Use nearest neighbor classifier in learned space Saenko et al., Adapting visual category models to new domains, ECCV 2010

Invariant representations w/ deep nets q d is the probability that a sample belongs to the d-th domain Tzeng et al., Simultaneous Deep Transfer Across Domains and Tasks, ICCV 2015

Invariant representations w/ deep nets Bousmalis et al., Domain Separation Networks, NIPS 2016

Unsupervised feature learning using context

Skip-gram model (word embeddings) WE(king) WE(man) + WE(woman) = WE(queen) Mikolov et al., Distributed Representations of Words and Phrases, NIPS 2013

Mikolov et al., Distributed Representations of Words and Phrases, NIPS 2013

Context prediction for images 1 2 3 4 5 A 6 7 8 B Doersch et al., Unsupervised Visual Representation Learning by Context Prediction, ICCV 2015

Relative position task 8 possible locations Classifier CNN CNN Randomly Sample Patch Sample Second Patch Doersch et al., Unsupervised Visual Representation Learning by Context Prediction, ICCV 2015

Classifier Patch Embedding Input Nearest Neighbors CNN CNN Note: connects across instances! Doersch et al., Unsupervised Visual Representation Learning by Context Prediction, ICCV 2015

Ranking

Relative attributes We need to compare images by attribute strength bright smiling natural Parikh and Grauman, Relative attributes, ICCV 2011

Learning relative attributes We want to learn a spectrum (ranking model) for an attribute, e.g. brightness. Supervision consists of: Ordered pairs Similar pairs Parikh and Grauman, Relative attributes, ICCV 2011

Learning relative attributes Learn a ranking function Image features Learned parameters that best satisfies the constraints: Parikh and Grauman, Relative attributes, ICCV 2011

Learning relative attributes Max-margin learning to rank formulation Rank margin w m Image Relative attribute score Joachims, Optimizing search engines using clickthrough data, KDD 2002