Machine Learning. Professor Sridhar Mahadevan

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Lecture 1: Machine Learning Basics

CSL465/603 - Machine Learning

Python Machine Learning

Lecture 1: Basic Concepts of Machine Learning

Generative models and adversarial training

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Spring 2014 SYLLABUS Michigan State University STT 430: Probability and Statistics for Engineering

Lecture 10: Reinforcement Learning

CS Machine Learning

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Axiom 2013 Team Description Paper

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Reinforcement Learning by Comparing Immediate Reward

Evolution of Symbolisation in Chimpanzees and Neural Nets

Evolutive Neural Net Fuzzy Filtering: Basic Description

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

GRADUATE STUDENT HANDBOOK Master of Science Programs in Biostatistics

Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Laboratorio di Intelligenza Artificiale e Robotica

Semi-Supervised Face Detection

Using Web Searches on Important Words to Create Background Sets for LSI Classification

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Active Learning. Yingyu Liang Computer Sciences 760 Fall

CS/SE 3341 Spring 2012

Abstractions and the Brain

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

(Sub)Gradient Descent

MGT/MGP/MGB 261: Investment Analysis

COMPUTER-AIDED DESIGN TOOLS THAT ADAPT

Seminar - Organic Computing

A Case-Based Approach To Imitation Learning in Robotic Agents

Artificial Neural Networks written examination

Learning From the Past with Experiment Databases

Firms and Markets Saturdays Summer I 2014

KLI: Infer KCs from repeated assessment events. Do you know what you know? Ken Koedinger HCI & Psychology CMU Director of LearnLab

A Case Study: News Classification Based on Term Frequency

Human Emotion Recognition From Speech

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Probabilistic Latent Semantic Analysis

Georgetown University at TREC 2017 Dynamic Domain Track

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Laboratorio di Intelligenza Artificiale e Robotica


Assignment 1: Predicting Amazon Review Ratings

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

EECS 700: Computer Modeling, Simulation, and Visualization Fall 2014

Learning Methods for Fuzzy Systems

School of Innovative Technologies and Engineering

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Penn State University - University Park MATH 140 Instructor Syllabus, Calculus with Analytic Geometry I Fall 2010

STA 225: Introductory Statistics (CT)

Speeding Up Reinforcement Learning with Behavior Transfer

S T A T 251 C o u r s e S y l l a b u s I n t r o d u c t i o n t o p r o b a b i l i t y

Mathematics. Mathematics

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Indian Institute of Technology, Kanpur

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

A Neural Network GUI Tested on Text-To-Phoneme Mapping

MTH 215: Introduction to Linear Algebra

Reducing Features to Improve Bug Prediction

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

An OO Framework for building Intelligence and Learning properties in Software Agents

An investigation of imitation learning algorithms for structured prediction

EGRHS Course Fair. Science & Math AP & IB Courses

Learning Disability Functional Capacity Evaluation. Dear Doctor,

Discriminative Learning of Beam-Search Heuristics for Planning

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Stochastic Calculus for Finance I (46-944) Spring 2008 Syllabus

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Learning Methods in Multilingual Speech Recognition

Introduction to Causal Inference. Problem Set 1. Required Problems

arxiv: v1 [cs.lg] 3 May 2013

Attributed Social Network Embedding

Grade Band: High School Unit 1 Unit Target: Government Unit Topic: The Constitution and Me. What Is the Constitution? The United States Government

Corrective Feedback and Persistent Learning for Information Extraction

arxiv: v1 [cs.lg] 15 Jun 2015

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Top US Tech Talent for the Top China Tech Company

Math 181, Calculus I

Word Segmentation of Off-line Handwritten Documents

A study of speaker adaptation for DNN-based speech synthesis

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Calibration of Confidence Measures in Speech Recognition

Modeling function word errors in DNN-HMM based LVCSR systems

Massachusetts Institute of Technology Tel: Massachusetts Avenue Room 32-D558 MA 02139

Robot Learning Simultaneously a Task and How to Interpret Human Instructions

Counseling 150. EOPS Student Readiness and Success

State University of New York at Buffalo INTRODUCTION TO STATISTICS PSC 408 Fall 2015 M,W,F 1-1:50 NSC 210

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Undergraduate Program Guide. Bachelor of Science. Computer Science DEPARTMENT OF COMPUTER SCIENCE and ENGINEERING

Transcription:

Machine Learning Professor Sridhar Mahadevan mahadeva@cs.umass.edu Lecture 1 Home page:www-edlab.cs.umass.edu/cs689 Quizzes, mini-projects: moodle.umass.edu Discussion forum:piazza.com CMPSCI 689 p. 1/35

What is "Learning"? Motor skills: walk, drive a bicycle, drive, play tennis or golf, play the piano. Language: Speech recognition, read and write natural languages Spatial knowledge: Navigate between spatial locations, physical layout of a room. Symbolic knowledge: algebra, arithmetic, calculus. Social rules: how to interact with people, animals, machines... CMPSCI 689 p. 2/35

What Activity is Shown Here? 32 CMPSCI 689 p. 3/35

The Challenge of Learning How is it possible that animals and humans are able to learn so much knowledge from a relatively small number of examples? Most of what is learned is already built-in (The Blank Slate, Steve Pinker). The brain is hardwired to learn specific classes of functions (e.g., language, faces, motor control). Evolution has equipped the brain with some amazingly clever algorithms. The brain is massively parallel with a 100 billion slow" unreliable computing units (neurons). CMPSCI 689 p. 4/35

Abstract Definition of "Learning" "Learning" denotes changes in a system that are adaptive in that they enable the system to perform the same task or similar tasks drawn from the same population better over time (Herbert Simon, 1980). Learning denotes knowledge acquisition in the absense of explicit programming (Valiant, 1986). CMPSCI 689 p. 5/35

Why should Machines "Learn"? "Learning" can be viewed as a form of implicit programming. Imagine a robot that learns to play tennis by observing people play, and by trial and error. If the task changes over time, learning can make a machine adaptive. Learning may enable a machine to outperform human programming. CMPSCI 689 p. 6/35

Why Study Machine Learning? If you invent a breakthrough in artificial intelligence, so machines can learn, that is worth 10 Microsofts. Bill Gates quoted in NY Times, Monday March 3, 2004. CMPSCI 689 p. 7/35

IBM Jeopardy Quiz Program CMPSCI 689 p. 8/35

Speech Recognition on Smart Phones CMPSCI 689 p. 9/35

Imagenet Vision Challenge CMPSCI 689 p. 10/35

Mapping Images to Text CMPSCI 689 p. 11/35

Autonomous Driving CMPSCI 689 p. 12/35

Machine Learning on Mars CMPSCI 689 p. 13/35

First Machine Learning Program CMPSCI 689 p. 14/35

Work done at the ALL Lab CMPSCI 689 p. 15/35

Google Deep Mind CMPSCI 689 p. 16/35

Reinforcement Learning in the Brain CMPSCI 689 p. 17/35

Related Fields Biology: Brain, Development, Evolution, Genetics, Neuroscience. Information Theory: Coding Theory, Entropy. Linguistics: Grammars, Language acquisition Mathematics: Calculus, Linear Algebra, Optimization. Psychology: Analogy, Concept Learning, Curiosity, Discovery, Memory, Reinforcement Philosophy: Causality, Induction, Theory Formation Statistics: Probability Distributions, Estimation, Hypothesis Testing. CMPSCI 689 p. 18/35

Learning as Search The process of learning can be viewed as one of searching through a space of hypotheses H for one that best fits the data. The data can be viewed as samples from a (known, unknown) probability distribution The data can be discrete (e.g., rooms in a building, words, web pages), or continuous (sensor measurements). The data may be labeled (category or reward signal) or unlabeled CMPSCI 689 p. 19/35

Data Modeling Data from a known distribution: Assumes that the data is coming from a specific class of distributions P(x θ) (e.g., Multinomial, Normal, Poisson) Models: Logistic regression, Mixure model, Hidden Markov Model, Dynamic Bayes Nets. Distribution-free learning: Examples: Deep learning, Decision trees, Nearest Neighbor, Support Vector Machines, Manifold learning. CMPSCI 689 p. 20/35

Problem Formulations Density estimation: Unsupervised learning Estimate (joint) distribution of the data P(X) Classification: Supervised learning Estimate conditional distribution P(Y X) Regression: Function approximation Estimate conditional mean E(Y X) Reinforcement Learning: Control learning Learn a policy π mapping states (S) to actions (A) that maximize long-term rewards (R). CMPSCI 689 p. 21/35

The Indus Script Fig. 1. An example of an Indus seal, showing the three 4000 year old undeciphered language CMPSCI 689 p. 22/35

Deciphering the Indus Script CMPSCI 689 p. 23/35

Markov Model CMPSCI 689 p. 24/35

Is the Indus Script a Language? CMPSCI 689 p. 25/35

Limitations of Learning Computational learning theory (Gold, 1960s; Valiant, 1986; Vapnik and Chervonenkis, 1974) A "complexity"-theory distribution-free model of learning. This theory identifies conditions under which reliable learning is possible. Makes rich connections to algorithmic hardness results (complexity classes). Led to some of the best machine learning algorithms (support vector machines). CMPSCI 689 p. 26/35

PAC Learning Given a class H of functions on a space of instances X, a fixed but unknown distribution P on X, how many examples are needed to "learn" any f H? Learner outputs an approximation h whose true error w.r.t. P is ǫ,0 < ǫ < 1. Learner converges to a good approximation with probability 1 δ,0 < δ < 1. Finite H: Learner needs m 1 ǫ examples. ( log( 1 δ )+log( H )) General H: m 1 ǫ ( log( 1 δ )+VC(H)) CMPSCI 689 p. 27/35

Administrivia Class lectures: M/Wed 2:30-3:45, Room 142 My office hours: M/Wed 1:30:-2:30, Room 204 TAs: Clemens Rosenbaum, Francisco Garcia Get a class account onpiazza.com Ed lab account onelnux*.cs.umass.edu (MATLAB) CMPSCI 689 p. 28/35

Recommended Texts Kevin Murphy, Machine Learning: A Probabilistic Approach, MIT Press, 2012. Richard Sutton and Andrew Barto, Reinforcement Learning: An Introduction, MIT Press, 2009. Hastie, Tibshirani, and Friedman, Statistical Learning, Springer-Verlag (2nd edition). (available online) David Mackay, Information Theory, Inference, and Machine Learning (Cambridge Univ. Press). (available online) CMPSCI 689 p. 29/35

Background Material Linear algebra (e.g., Strang) Statistics (e.g., Casella and Berger) Optimization (e.g., Boyd and Vanderberghe (available online)) Multivariate calculus (e.g., Lagrange multipliers) CMPSCI 689 p. 30/35

Many Software Resources MATLAB (available on edlab machines) Python ML packages R and RStudio statistics package Weka Java based ML package Theano, Torch, Caffe, Mocha: deep learning packages CMPSCI 689 p. 31/35

Course Outline September October November Unsupervised Learning Supervised Learning Reinforcement Learning CMPSCI 689 p. 32/35

Weekly Readings and Course Project Readings: See class web page Final project: Oct 19th: Preliminary project proposal Dec 7th, 9th: Final project presentations. CMPSCI 689 p. 33/35

Course Grading Section Weight Mini projects 30% Quizzes 30% Final Project 30% Independent Activities 10% CMPSCI 689 p. 34/35

Reading for Next Week Read the survey article "A Few Useful Things to Know About Machine Learning" by Pedro Domingos (see Moodle for paper or class web page). Read Chapter 1 in Murphy textbook Review basic concepts from linear algebra: matrices, vector spaces, subspaces, eigenvalues/eigenvectors, orthogonality. Review basic probability/statistics: random variables, distributions, moments (means, variances). CMPSCI 689 p. 35/35