Machine Learning Kaushik Sinha Fall 2013

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

(Sub)Gradient Descent

Python Machine Learning

Lecture 1: Machine Learning Basics

Class Meeting Time and Place: Section 3: MTWF10:00-10:50 TILT 221

Computers Change the World

Lecture 1: Basic Concepts of Machine Learning

Algebra Nation and Computer Science for MS Initiatives. Marla Davis, Ph.D. NBCT Office of Secondary Education

CSL465/603 - Machine Learning

Lecture 6: Applications

A study of speaker adaptation for DNN-based speech synthesis

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

LEARNER VARIABILITY AND UNIVERSAL DESIGN FOR LEARNING

Human Emotion Recognition From Speech

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Spring 2014 SYLLABUS Michigan State University STT 430: Probability and Statistics for Engineering

MTH 215: Introduction to Linear Algebra

CS 446: Machine Learning

CS 3516: Computer Networks

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Syllabus: CS 377 Communication and Ethical Issues in Computing 3 Credit Hours Prerequisite: CS 251, Data Structures Fall 2015

TD(λ) and Q-Learning Based Ludo Players

B. How to write a research paper

EECS 700: Computer Modeling, Simulation, and Visualization Fall 2014

Probabilistic Latent Semantic Analysis

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

UNIT ONE Tools of Algebra

Speak Up 2012 Grades 9 12

Artificial Neural Networks written examination

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

A Neural Network GUI Tested on Text-To-Phoneme Mapping

CS177 Python Programming

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

CS Machine Learning

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Data Structures and Algorithms

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

CS Course Missive

WEBSITES TO ENHANCE LEARNING

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Texas A&M University - Central Texas PSYK PRINCIPLES OF RESEARCH FOR THE BEHAVIORAL SCIENCES. Professor: Elizabeth K.

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Generative models and adversarial training

AST Introduction to Solar Systems Astronomy

Android App Development for Beginners

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Welcome to SAT Brain Boot Camp (AJH, HJH, FJH)

Aerospace Engineering

Syllabus Foundations of Finance Summer 2014 FINC-UB

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

E-Commerce & Social Networking BADM 364 Fall 2014

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

MATH 108 Intermediate Algebra (online) 4 Credits Fall 2008

K 1 2 K 1 2. Iron Mountain Public Schools Standards (modified METS) Checklist by Grade Level Page 1 of 11

arxiv: v2 [cs.cv] 30 Mar 2017

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Page 1 of 8 REQUIRED MATERIALS:

Cross Language Information Retrieval

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

SCT Banner Student Fee Assessment Training Workbook October 2005 Release 7.2

AP Chemistry

Focused on Understanding and Fluency

preassessment was administered)

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

SCHOOL WITHOUT CLASSROOMS BERLIN ARCHITECTURE COMPETITION TO

CHANCERY SMS 5.0 STUDENT SCHEDULING

Teaching a Laboratory Section

Beginning to Flip/Enhance Your Classroom with Screencasting. Check out screencasting tools from (21 Things project)

Learning, Communication, and 21 st Century Skills: Students Speak Up For use with NetDay Speak Up Survey Grades 3-5

Modeling user preferences and norms in context-aware systems

BA 130 Introduction to International Business

Axiom 2013 Team Description Paper

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Course Syllabus for Calculus I (Summer 2017)

Northern Kentucky University Department of Accounting, Finance and Business Law Financial Statement Analysis ACC 308

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Postprint.

CPMT 1347 Computer System Peripherals COURSE SYLLABUS

Consequences of Your Good Behavior Free & Frequent Praise

MOODLE 2.0 GLOSSARY TUTORIALS

Machine Learning and Development Policy

Syllabus ENGR 190 Introductory Calculus (QR)

Math 181, Calculus I

EDCI 699 Statistics: Content, Process, Application COURSE SYLLABUS: SPRING 2016

CS 100: Principles of Computing

CENTRAL MAINE COMMUNITY COLLEGE Introduction to Computer Applications BCA ; FALL 2011

Evaluating Statements About Probability

ECE-492 SENIOR ADVANCED DESIGN PROJECT

B.S/M.A in Mathematics

COSI Meet the Majors Fall 17. Prof. Mitch Cherniack Undergraduate Advising Head (UAH), COSI Fall '17: Instructor COSI 29a

IAT 888: Metacreation Machines endowed with creative behavior. Philippe Pasquier Office 565 (floor 14)

Circuit Simulators: A Revolutionary E-Learning Platform

Speech Recognition at ICSI: Broadcast News and beyond

Please read this entire syllabus, keep it as reference and is subject to change by the instructor.

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

STUDENTS' RATINGS ON TEACHER

Capital Budgeting 1. Syllabus

Teaching Architecture Metamodel-First

Transcription:

Machine Learning Kaushik Sinha Fall 2013

Administrative Stuff

Introduction Instructor: Asst. Prof. Kaushik Sinha 2 lectures per week MW 12:30-1:45 pm Office Hours MW 11:30-12:30 Jabara Hall 243

Study Groups (2-3 people) This course will cover non-trivial material, learning in a group makes it less hard and more fun! It is recommended (but not required)

Prerequisites Three pillars of ML: Statistics / Probability Linear Algebra Multivariate Calculus Should be confident in at least 1/ 3, ideally 2/ 3.

Grades... Your grade is a composite of: (Homework) Exams (Mid-term + Final) Final Project

Homework You can discuss homework with your peers but your submitted answer should be your own! Make honest attempt on all questions

Exams Exams will be (to some degree) based on homework assignments Best preparation: Make sure you really really understand the homework assignments 2 Exams: Final + Midterm Will be 40% of your grade.

Final Project 40% of your grade. 4 page writeup. Joint effort of 2-3 people. Come up with your own ideas. Application: Some interesting application of machine learning In-depth study: Reproduce results from a high-acclaimed paper Research: Incorporate ML into a research project Extra credit is given for working systems (e.g. iphone-, web app) Details will be posted on course website later

Cheating Don t cheat! Use your common sense. I won t be your friend anymore!

MACHINE LEARNING!!!

What is Machine Learning? Formally: (Mitchell 1997): A computer program A is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E. Informally: Algorithms that improve on some task with experience.

When should we use ML? Not ML problems: Traveling Salesman, 3-Sat, etc. ML Problems: Hard to formalize, but human expert can provide examples / feedback. Computer needs to learn from feedback. Is there a sign of cancer in this fmri scan? What will the Dow Jones be tomorrow? Teach a robot to ride a unicycle.

Sometimes easy for humans, hard for computers Male or Female? Even 1 year old children can identify gender pretty reliably Easy to come up with examples. But impossible to formalize as a CS problem. You need machine learning!

Example: Problem: Given an image of a handwritten digit, what digit is it? Input: Problem: You have absolutely no idea how to do this! Clever Algorithm Output: 2

Example: Problem: Given an image of a handwritten digit, what digit is it? 0 1 2 3 4 5 6 7 8 9 Input: Output: Clever Algorithm 2 Problem: You have absolutely no idea how to do this! Good news: You have examples

Example: Problem: Given an image of a handwritten digit, what digit is it? The Machine Learning Approach: 0 1 2 3 4 5 6 7 8 9 Machine Learning Algorithm Input: Output: Clever Algorithm 2

Example: Problem: Given an image of a handwritten digit, what digit is it? 0 1 2 3 4 5 6 7 8 9 Training Machine Learning Algorithm Testing Learned Algorithm 2

Handwritten Digits Recognition (1990-1995) Pretty much solved in the mid nine-tees. (Lecun et al) Convolutional Neural Networks Now used by USPS for zip-codes, ATMs for automatic check cashing etc.

TD-Gammon (1994) Gerry Tesauro (IBM) teaches a neural network to play Backgammon. The net plays 100K+ games against itself and beats world champion [Neurocomputation 1994] Algorithm teaches itself how to play so well!!!

Deep Blue (1997) IBM s Deep Blue wins against Kasparov in chess. Crucial winning move is made due to Machine Learning (G. Tesauro).

Watson (2011) IBM s Watson wins the game show jeopardy against former winners Brad Rutters and Ken Jennings. Extensive Machine Learning techniques were used.

Face Detection (2001) Viola Jone s solves face detection Previously very hard problem in computer vision Now commodity in off-the-shelf cellphones / cameras

Grand Challenge (2005) Darpa Grand Challenge: The vehicle must drive autonomously 150 Miles through the dessert along a difficult route. 2004 Darpa Grand Challenge huge disappointment, best team makes 11.78 / 150 miles 2005 Darpa Grand Challenge 2 is completed by several ML powered teams.

Speech, Netflix,... iphone ships with built-in speech recognition Google mobile search speech based (very reliable) Automatic translation...

ML is the engine for many fields... Natural Language Processing Computer Vision Machine Learning Computatio nal Biology Robotics

Internet companies Collecting massive amounts of data Hoping that some smart Machine Learning person makes money out of it. Your future job!

Example: Webmail Spam filtering Given Email, predict if it is spam or not. Ad - matching Given user info predict which ad will be clicked on.

Example: Websearch Ad Matching Given query, predict which ad will be clicked on. Web-search ranking Given query, predict which document will be clicked on.

Example: Google News Document clustering Given news articles, automatically identify and sort them by topic.

When will it stop? The human brain is one big learning machine We know that we can still do a lot better! However, it is hard. Very few people can design new ML algorithms. But many people can use them!

What types of ML are there? As far as this course is concerned: supervised learning: Given labeled examples, find the right prediction of an unlabeled example. (e.g. Given annotated images learn to detect faces.) unsupervised learning: Given data try to discover similar patterns, structure, low dimensional (e.g. automatically cluster news articles by topic)

Basic Setup Pre-processing Clean up the data. Boring but necessary. Feature Extraction Use expert knowledge to get representation of data. Learning Focus of this course. (Post-processing) Whatever you do when you are done.

Feature Extraction

Feature Extraction Represent data in terms of vectors. Features are statistics that describe the data. Real World Data Vector Space Each dimension is one feature.

Handwritten digits Features are statistics that describe the data Feature: width/height Pretty good for 1 vs. 2 Not so good for 2 vs. 3 Feature: raw pixels 16x16 Works for digits (to some degree) Does not work for trickier stuff 256x1

Bag of Words for Images Sparse Vector Image: Interest Points 0 1 0 0 0 3 Dictionary of possible interest points. 0 0 0 Extract interest points and represent the image as a bag of interest points. 0

Text (Bag of Words) Text documents: Bag of Words 0 1 in into... 0 0 0... 2 is 0... 0 0 Take dictionary with n words. Represent a text document as n dimensional vector, where the i-th dimension contains the number of times word i appears in the document. 0

Audio? Movies? QuickTime and a Photo - JPEG decompressor are needed to see this picture. Use a sliding window and Fast Fourier Transform Treat it as a sequence of images

Feature Space Everything that can be stored on a computer can stored as a vector Representation is critical for successful learning. [Not in this course, though.] Throughout this course we will assume data is just points in a Feature Space Important distinction: sparse / dense Most features are zero Every feature is present

Mini-Quiz T/F: Every traditional CS problem is also an ML problem. FALSE T/F: Image Features are always dense. FALSE T/F: The feature space can be very high dimensional. TRUE T/F: Bag of words features are sparse. TRUE

Mini-Quiz T/F: Every traditional CS problem is also an ML problem. FALSE T/F: Image Features are always dense. FALSE T/F: The feature space can be very high dimensional. TRUE T/F: Bag of words features are sparse. TRUE