Introduction to Machine Learning CMSC 422

Similar documents
(Sub)Gradient Descent

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Python Machine Learning

Lecture 1: Machine Learning Basics

Lecture 1: Basic Concepts of Machine Learning

CSL465/603 - Machine Learning

Data Structures and Algorithms

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

A Case Study: News Classification Based on Term Frequency

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

CS 446: Machine Learning

CS Machine Learning

EGRHS Course Fair. Science & Math AP & IB Courses

EECS 700: Computer Modeling, Simulation, and Visualization Fall 2014

Human Emotion Recognition From Speech

Version Space. Term 2012/2013 LSI - FIB. Javier Béjar cbea (LSI - FIB) Version Space Term 2012/ / 18

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Class Meeting Time and Place: Section 3: MTWF10:00-10:50 TILT 221

Rule Learning With Negation: Issues Regarding Effectiveness

Undergraduate Program Guide. Bachelor of Science. Computer Science DEPARTMENT OF COMPUTER SCIENCE and ENGINEERING

CS177 Python Programming

Axiom 2013 Team Description Paper

ECO 3101: Intermediate Microeconomics

Learning From the Past with Experiment Databases

Self Study Report Computer Science

Word Segmentation of Off-line Handwritten Documents

Spring 2014 SYLLABUS Michigan State University STT 430: Probability and Statistics for Engineering

Math 181, Calculus I

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

CALCULUS III MATH

PHY2048 Syllabus - Physics with Calculus 1 Fall 2014

Syllabus - ESET 369 Embedded Systems Software, Fall 2016

B.S/M.A in Mathematics

CS Course Missive

Syllabus: PHI 2010, Introduction to Philosophy

CS 100: Principles of Computing

CS 101 Computer Science I Fall Instructor Muller. Syllabus

Penn State University - University Park MATH 140 Instructor Syllabus, Calculus with Analytic Geometry I Fall 2010

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Laboratorio di Intelligenza Artificiale e Robotica

Active Learning. Yingyu Liang Computer Sciences 760 Fall

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

Firms and Markets Saturdays Summer I 2014

Intensive English Program Southwest College

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

SYLLABUS. EC 322 Intermediate Macroeconomics Fall 2012

Rule Learning with Negation: Issues Regarding Effectiveness

Artificial Neural Networks written examination

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

ASTRONOMY 2801A: Stars, Galaxies & Cosmology : Fall term

Design and Creation of Games GAME

Nutrition 10 Contemporary Nutrition WINTER 2016

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Learning Lesson Study Course

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Semi-Supervised Face Detection

Indicators Teacher understands the active nature of student learning and attains information about levels of development for groups of students.

Compositional Semantics

Softprop: Softmax Neural Network Backpropagation Learning

Syllabus Foundations of Finance Summer 2014 FINC-UB

Switchboard Language Model Improvement with Conversational Data from Gigaword

ITSC 1301 Introduction to Computers Course Syllabus

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Physics 270: Experimental Physics

MAR Environmental Problems & Solutions. Stony Brook University School of Marine & Atmospheric Sciences (SoMAS)

The lab is designed to remind you how to work with scientific data (including dealing with uncertainty) and to review experimental design.

arxiv: v1 [cs.lg] 15 Jun 2015

Indian Institute of Technology, Kanpur

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

We are strong in research and particularly noted in software engineering, information security and privacy, and humane gaming.

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

STA 225: Introductory Statistics (CT)

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Foothill College Summer 2016

EDINA SENIOR HIGH SCHOOL Registration Class of 2020

COSI Meet the Majors Fall 17. Prof. Mitch Cherniack Undergraduate Advising Head (UAH), COSI Fall '17: Instructor COSI 29a

Course Syllabus for Math

Speech Emotion Recognition Using Support Vector Machine

Introduction to CS 100 Overview of UK. CS September 2015

MATH 1A: Calculus I Sec 01 Winter 2017 Room E31 MTWThF 8:30-9:20AM

Lecture 10: Reinforcement Learning

Word learning as Bayesian inference

Generative models and adversarial training

Office Hours: Mon & Fri 10:00-12:00. Course Description

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

CS 3516: Computer Networks

Laboratorio di Intelligenza Artificiale e Robotica

Please read this entire syllabus, keep it as reference and is subject to change by the instructor.

Exploration. CS : Deep Reinforcement Learning Sergey Levine

A survey of multi-view machine learning

Beyond the Pipeline: Discrete Optimization in NLP

Math 96: Intermediate Algebra in Context

Knowledge Transfer in Deep Convolutional Neural Nets

Transcription:

Introduction to Machine Learning CMSC 422 Ramani Duraiswami Machine Learning studies representations and algorithms that allow machines to improve their performance on a task from experience. This is a broad overview of existing methods for machine learning and an introduction to adaptive systems in general

Prerequisites CMSC351 (Algorithms) and CMSC330 (Programming Languages) Recommended: STAT400. (Applied probability and statistics) and Linear Algebra. These previous courses require CMSC250 (Discrete Structures), CMSC216 (Computer Systems) Which in turn require CMSC131 (Object oriented programming) and MATH141 (Calculus) Course is about data, representations, mathematical modeling, and programming

Sections Two sections Prof. Marine Carpuat, 0101 This section, 0201 Cover the same material, but using somewhat different slides/notes Same textbook Common online homework Different exams/ exam dates

Topics Foundations of Supervised Learning Decision trees and inductive bias Geometry and nearest neighbors Perceptron Practical concerns: feature design, evaluation, debugging Beyond binary classification Advanced Supervised Learning Linear models and gradient descent Support Vector Machines Naive Bayes models and probabilistic modeling Neural networks and deep learning Kernels Ensemble learning

Topics Unsupervised learning K-means PCA Selected advanced topics (as time permits) Expectation maximization Online learning Markov decision processes Imitation learning

Homework Will try to have it at least every week Will not be excessive Essential for learning --- must do as in addition to read. Homework will be released on Canvas 20% No late homework

Homeworks

Textbook, and Class Preparation Textbook is free and online. Written by a colleague, Prof. Hal Daume III http://ciml.info Expect you to read material from the text, and other readings before the class Many other notes and books available and a few are listed in the syllabus

Projects Three projects in Python Project 1: Classification Project 2: Multiclass and Linear Models Project 3: PCAs and SVMs Remember you cannot publish or share project solutions cheating

Exams Exams Mid Term exam worth 20%, Date TBD Final exam worth 30 %, Saturday, May 13, 8:00-10:00am Closed book, Closed notes, in class Allowed a cheat sheet in your own handwriting

Where to find the readings: A Course in Machine Learning view and submit assignments: Canvas check your grades: Canvas ask and answer questions, participate in discussions and surveys, contact the instructors, and everything else: Piazza Please use piazza instead of email

What is Learning? Ability to use previous data to perform future actions Biological systems do it all the time H. Simon - Learning denotes changes in the system that are adaptive in the sense that they enable the system to do the task or tasks drawn from the same population more efficiently and more effectively the next time.

Machine Learning is Everywhere Slide adapted from Prof. Roth, UIUC 13

Learning is the future Learning techniques will be a basis for every application that involves a connection to the messy real world Basic learning algorithms are ready for use in applications today Prospects for broader future applications make for exciting fundamental research and development opportunities Many unresolved issues Theory and Systems While it s hot, there are many things we don t know how to do 14

Work in Machine Learning Artificial Intelligence; Theory; Experimental CS Makes Use of: Probability and Statistics; Linear Algebra; Theory of Computation; Related to: Philosophy, Psychology (cognitive, developmental), Neurobiology, Linguistics, Vision, Speech, Robotics,. Has applications in: AI (Natural Language; Vision; Speech & Audio; Planning; HCI) Engineering (Agriculture; Civil; ) Computer Science (Compilers; Architecture; Systems; data bases) Analytics 15

Today s topics What does it mean to learn by example? Classification tasks Inductive bias Formalizing learning

Classification tasks How would you write a program to distinguish a picture of me from a picture of someone else? Provide examples pictures of me and pictures of other people and let a classifier learn to distinguish the two.

Classification tasks How would you write a program to distinguish a sentence is grammatical or not? Provide examples of grammatical and ungrammatical sentences and let a classifier learn to distinguish the two.

Classification tasks How would you write a program to distinguish cancerous cells from normal cells? Provide examples of cancerous and normal cells and let a classifier learn to distinguish the two.

Classification tasks How would you write a program to distinguish cancerous cells from normal cells? Provide examples of cancerous and normal cells and let a classifier learn to distinguish the two.

Let s try it out Your task: learn a classifier to distinguish class A from class B from examples

Examples of class A:

Examples of class B

Let s try it out learn a classifier from examples Now: predict class on new examples using what you ve learned

What if my program came up with

Key ingredients needed for learning Training vs. test examples Memorizing the training examples is not enough! Need to generalize to make good predictions on test examples Inductive bias Many classifier hypotheses are plausible Need assumptions about the nature of the relation between examples and classes

Machine Learning as Function Approximation Problem setting Set of possible instances X Unknown target function f: X Y Set of function hypotheses H = h h: X Y} Input Training examples { x 1, y 1, x N, y N } of unknown target function f Output Hypothesis h H that best approximates target function f

Formalizing induction: Loss Function l(y, f(x)) where y is the truth and f x is the system s prediction e.g. l y, f(x) = 0 if y = f(x) 1 otherwise Captures our notion of what is important to learn

Formalizing induction: Data generating distribution Where does the data come from? Data generating distribution A probability distribution D over (x, y) pairs We don t know what D is! We only get a random sample from it: our training data

Formalizing induction: Expected loss f should make good predictions as measured by loss l on future examples that are also drawn from D Formally ε, the expected loss of f over D with respect to l should be small ε E x,y ~D l(y, f(x)) = D x, y l(y, f(x)) (x,y)

Formalizing induction: Training error We can t compute expected loss because we don t know what D is We only have a sample of D training examples { x 1, y 1, x N, y N } All we can compute is the training error ε N 1 n=1 N l(y n, f(x n ))

Formalizing Induction Given a loss function l a sample from some unknown data distribution D Our task is to compute a function f that has low expected error over D with respect to l. E x,y ~D l(y, f(x)) = D x, y l(y, f(x)) (x,y)

Recap: introducing machine learning What does it mean to learn by example? Classification tasks Learning requires examples + inductive bias Generalization vs. memorization Formalizing the learning problem Function approximation Learning as minimizing expected loss

Your tasks before next class Check out course webpage, Canvas, Piazza Do the readings Get started on HW01 due Thursday 10:59am