Web and Internet Economics

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Georgetown University at TREC 2017 Dynamic Domain Track

Lecture 10: Reinforcement Learning

Laboratorio di Intelligenza Artificiale e Robotica

CSL465/603 - Machine Learning

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Axiom 2013 Team Description Paper

Python Machine Learning

(Sub)Gradient Descent

Laboratorio di Intelligenza Artificiale e Robotica

An investigation of imitation learning algorithms for structured prediction

Learning Methods for Fuzzy Systems

Automatic Discretization of Actions and States in Monte-Carlo Tree Search

Lecture 1: Machine Learning Basics

Lecture 1: Basic Concepts of Machine Learning

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Speeding Up Reinforcement Learning with Behavior Transfer

Human-like Natural Language Generation Using Monte Carlo Tree Search

Probabilistic Latent Semantic Analysis

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

TD(λ) and Q-Learning Based Ludo Players

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

Reinforcement Learning by Comparing Immediate Reward

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

CS Machine Learning

AI Agent for Ice Hockey Atari 2600

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

High-level Reinforcement Learning in Strategy Games

Artificial Neural Networks written examination

Robot Learning Simultaneously a Task and How to Interpret Human Instructions

Improving Action Selection in MDP s via Knowledge Transfer

Human Emotion Recognition From Speech

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

A Reinforcement Learning Approach for Adaptive Single- and Multi-Document Summarization

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Speech Recognition at ICSI: Broadcast News and beyond

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

Regret-based Reward Elicitation for Markov Decision Processes

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Task Completion Transfer Learning for Reward Inference

Reducing Features to Improve Bug Prediction

Learning Prospective Robot Behavior

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Speech Emotion Recognition Using Support Vector Machine

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Learning and Transferring Relational Instance-Based Policies

Learning Methods in Multilingual Speech Recognition

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

AMULTIAGENT system [1] can be defined as a group of

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Assignment 1: Predicting Amazon Review Ratings

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Welcome to. ECML/PKDD 2004 Community meeting

A Reinforcement Learning Variant for Control Scheduling

Task Completion Transfer Learning for Reward Inference

Interaction Design Considerations for an Aircraft Carrier Deck Agent-based Simulation

Evolutive Neural Net Fuzzy Filtering: Basic Description

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Switchboard Language Model Improvement with Conversational Data from Gigaword

Natural Language Processing: Interpretation, Reasoning and Machine Learning

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

INPE São José dos Campos

FF+FPG: Guiding a Policy-Gradient Planner

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

An OO Framework for building Intelligence and Learning properties in Software Agents

Mining Student Evolution Using Associative Classification and Clustering

Softprop: Softmax Neural Network Backpropagation Learning

Speech Recognition using Acoustic Landmarks and Binary Phonetic Feature Classifiers

The Use of Statistical, Computational and Modelling Tools in Higher Learning Institutions: A Case Study of the University of Dodoma

A study of speaker adaptation for DNN-based speech synthesis

arxiv: v2 [cs.ro] 3 Mar 2017

Adaptive Generation in Dialogue Systems Using Dynamic User Modeling

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Modeling function word errors in DNN-HMM based LVCSR systems

Applications of data mining algorithms to analysis of medical data

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

Mining Association Rules in Student s Assessment Data

Discriminative Learning of Beam-Search Heuristics for Planning

Courses in English. Application Development Technology. Artificial Intelligence. 2017/18 Spring Semester. Database access

Rule Learning With Negation: Issues Regarding Effectiveness

CS 598 Natural Language Processing

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Summarizing Answers in Non-Factoid Community Question-Answering

Soft Computing based Learning for Cognitive Radio

Modeling function word errors in DNN-HMM based LVCSR systems

Computerized Adaptive Psychological Testing A Personalisation Perspective

BYLINE [Heng Ji, Computer Science Department, New York University,

Learning to Schedule Straight-Line Code

Massachusetts Institute of Technology Tel: Massachusetts Avenue Room 32-D558 MA 02139

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language

Transcription:

Web and Internet Economics Introduction to Machine Learning Matteo Papini a.a. 2017/2018

Internet Commerce vs Regular Commerce Efficiency Pull driven marketing and advertising Trust and reputation Personalization New opportunities

Internet Commerce vs Regular Commerce Efficiency Pull driven marketing and advertising Trust and reputation Personalization New opportunities

Internet Commerce vs Regular Commerce Efficiency Pull driven marketing and advertising Trust and reputation Personalization New opportunities

Internet Commerce vs Regular Commerce Efficiency Pull driven marketing and advertising Trust and reputation Personalization New opportunities

Internet Commerce vs Regular Commerce Efficiency Pull driven marketing and advertising Trust and reputation Personalization New opportunities

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

Problem Problem: Models are based on several unknown parameters Examples: Online advertising Website optimization Pricing Sponsored search Targeting Recommendations Solution: Learn from data!

What is Machine Learning? The real question is: what is learning? Mitchell (1997): A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E

What is Machine Learning? The real question is: what is learning? Mitchell (1997): A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E

What is Machine Learning? The real question is: what is learning? Mitchell (1997): A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E

What is ML useful for? Computer vision and robotics Speech recognition Biology and medicine Financial industry Information retrieval, Web search,... Video gaming Many application and many jobs...

Why Machine Learning? We need computers to make informed decisions on new, unseen data Often it is too difficult to design a set of rules by hand Machine learning allows to automatically extract relevant information from data applying it to analyze new data

Why Machine Learning? We need computers to make informed decisions on new, unseen data Often it is too difficult to design a set of rules by hand Machine learning allows to automatically extract relevant information from data applying it to analyze new data

Why Machine Learning? We need computers to make informed decisions on new, unseen data Often it is too difficult to design a set of rules by hand Machine learning allows to automatically extract relevant information from data applying it to analyze new data

Machine Learning

Machine Learning Models Supervised Learning Learn the model Unsupervised Learning Learn the representation Reinforcement Learning Learn to control

Machine Learning Models Supervised Learning Learn the model Unsupervised Learning Learn the representation Reinforcement Learning Learn to control

Machine Learning Models Supervised Learning Learn the model Unsupervised Learning Learn the representation Reinforcement Learning Learn to control

Supervised Learning Goal Estimating the unknown model that maps known inputs to known outputs Training set: D = { x, y } y = f(x) Problems Classification Regression Techniques Artificial Neural Networks Support Vector Machines Decision trees

Supervised Learning Goal Estimating the unknown model that maps known inputs to known outputs Training set: D = { x, y } y = f(x) Problems Classification Regression Techniques Artificial Neural Networks Support Vector Machines Decision trees

Supervised Learning Goal Estimating the unknown model that maps known inputs to known outputs Training set: D = { x, y } y = f(x) Problems Classification Regression Techniques Artificial Neural Networks Support Vector Machines Decision trees

Supervised Learning Goal Estimating the unknown model that maps known inputs to known outputs Training set: D = { x, y } y = f(x) Problems Classification Regression Techniques Artificial Neural Networks Support Vector Machines Decision trees

Supervised Learning Goal Estimating the unknown model that maps known inputs to known outputs Training set: D = { x, y } y = f(x) Problems Classification Regression Techniques Artificial Neural Networks Support Vector Machines Decision trees

Supervised Learning Goal Estimating the unknown model that maps known inputs to known outputs Training set: D = { x, y } y = f(x) Problems Classification Regression Techniques Artificial Neural Networks Support Vector Machines Decision trees

Supervised Learning Goal Estimating the unknown model that maps known inputs to known outputs Training set: D = { x, y } y = f(x) Problems Classification Regression Techniques Artificial Neural Networks Support Vector Machines Decision trees

Supervised Learning: Example

Unsupervised Learning Goal Learning a more efficient representation of a set of unknown inputs Training set: D = {x}? = f(x) Problems Compression Clustering Techniques K-means Self-organizing maps Principal Component Analysis

Unsupervised Learning Goal Learning a more efficient representation of a set of unknown inputs Training set: D = {x}? = f(x) Problems Compression Clustering Techniques K-means Self-organizing maps Principal Component Analysis

Unsupervised Learning Goal Learning a more efficient representation of a set of unknown inputs Training set: D = {x}? = f(x) Problems Compression Clustering Techniques K-means Self-organizing maps Principal Component Analysis

Unsupervised Learning Goal Learning a more efficient representation of a set of unknown inputs Training set: D = {x}? = f(x) Problems Compression Clustering Techniques K-means Self-organizing maps Principal Component Analysis

Unsupervised Learning Goal Learning a more efficient representation of a set of unknown inputs Training set: D = {x}? = f(x) Problems Compression Clustering Techniques K-means Self-organizing maps Principal Component Analysis

Unsupervised Learning Goal Learning a more efficient representation of a set of unknown inputs Training set: D = {x}? = f(x) Problems Compression Clustering Techniques K-means Self-organizing maps Principal Component Analysis

Unsupervised Learning Goal Learning a more efficient representation of a set of unknown inputs Training set: D = {x}? = f(x) Problems Compression Clustering Techniques K-means Self-organizing maps Principal Component Analysis

Unsupervised Learning: Example

Unsupervised Learning: Example

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

Reinforcement Learning Goal Learning the optimal policy Training set: D = { x, u, x, r } π (x) = arg max u {Q (x, u)}, where Q (x, u) must be estimated. Problems Markov Decision Process (MDP) Partially Observable MDP (POMDP) Stochastic Games (SG) Techniques Q-learning SARSA Fitted Q-iteration

But Who s Counting? But Who s Counting?

Successful Stories: AlphaGo

Successful Stories: Parkour Parkour

Successful Stories: Atari Atari

Teaching Material Sutton and Barto, Reinforcement Learning: an Introduction, MIT Press, 1998. http://incompleteideas.net/book/the-book-2nd.html Look for the second edition Chapters 1-6 S. Bubeck and N. Cesa-Bianchi, Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. In Foundations and Trends in Machine Learning, Vol 5: No 1, 1-122, 2012. http://www.princeton.edu/ sbubeck/surveybcb12.pdf Sections 1, 2, and 3.

Teaching Material Sutton and Barto, Reinforcement Learning: an Introduction, MIT Press, 1998. http://incompleteideas.net/book/the-book-2nd.html Look for the second edition Chapters 1-6 S. Bubeck and N. Cesa-Bianchi, Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. In Foundations and Trends in Machine Learning, Vol 5: No 1, 1-122, 2012. http://www.princeton.edu/ sbubeck/surveybcb12.pdf Sections 1, 2, and 3.

Teaching Material Sutton and Barto, Reinforcement Learning: an Introduction, MIT Press, 1998. http://incompleteideas.net/book/the-book-2nd.html Look for the second edition Chapters 1-6 S. Bubeck and N. Cesa-Bianchi, Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. In Foundations and Trends in Machine Learning, Vol 5: No 1, 1-122, 2012. http://www.princeton.edu/ sbubeck/surveybcb12.pdf Sections 1, 2, and 3.

Teaching Material Sutton and Barto, Reinforcement Learning: an Introduction, MIT Press, 1998. http://incompleteideas.net/book/the-book-2nd.html Look for the second edition Chapters 1-6 S. Bubeck and N. Cesa-Bianchi, Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. In Foundations and Trends in Machine Learning, Vol 5: No 1, 1-122, 2012. http://www.princeton.edu/ sbubeck/surveybcb12.pdf Sections 1, 2, and 3.

Teaching Material Sutton and Barto, Reinforcement Learning: an Introduction, MIT Press, 1998. http://incompleteideas.net/book/the-book-2nd.html Look for the second edition Chapters 1-6 S. Bubeck and N. Cesa-Bianchi, Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. In Foundations and Trends in Machine Learning, Vol 5: No 1, 1-122, 2012. http://www.princeton.edu/ sbubeck/surveybcb12.pdf Sections 1, 2, and 3.

Teaching Material Sutton and Barto, Reinforcement Learning: an Introduction, MIT Press, 1998. http://incompleteideas.net/book/the-book-2nd.html Look for the second edition Chapters 1-6 S. Bubeck and N. Cesa-Bianchi, Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. In Foundations and Trends in Machine Learning, Vol 5: No 1, 1-122, 2012. http://www.princeton.edu/ sbubeck/surveybcb12.pdf Sections 1, 2, and 3.

Teaching Material Sutton and Barto, Reinforcement Learning: an Introduction, MIT Press, 1998. http://incompleteideas.net/book/the-book-2nd.html Look for the second edition Chapters 1-6 S. Bubeck and N. Cesa-Bianchi, Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. In Foundations and Trends in Machine Learning, Vol 5: No 1, 1-122, 2012. http://www.princeton.edu/ sbubeck/surveybcb12.pdf Sections 1, 2, and 3.