Machine Learning: CS 6375 Introduction. Instructor: Vibhav Gogate The University of Texas at Dallas

Similar documents
Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

(Sub)Gradient Descent

CSL465/603 - Machine Learning

Lecture 1: Machine Learning Basics

Lecture 1: Basic Concepts of Machine Learning

Python Machine Learning

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

CS Machine Learning

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Spring 2016 Stony Brook University Instructor: Dr. Paul Fodor

Laboratorio di Intelligenza Artificiale e Robotica

Data Structures and Algorithms

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Axiom 2013 Team Description Paper

Lecture 10: Reinforcement Learning

Agents and environments. Intelligent Agents. Reminders. Vacuum-cleaner world. Outline. A vacuum-cleaner agent. Chapter 2 Actuators

Self Study Report Computer Science

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Rule Learning With Negation: Issues Regarding Effectiveness

Firms and Markets Saturdays Summer I 2014

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Artificial Neural Networks written examination

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Corrective Feedback and Persistent Learning for Information Extraction

Human Emotion Recognition From Speech

Probabilistic Latent Semantic Analysis

MGT/MGP/MGB 261: Investment Analysis

A Neural Network GUI Tested on Text-To-Phoneme Mapping

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Speech Emotion Recognition Using Support Vector Machine

Model Ensemble for Click Prediction in Bing Search Ads

Time series prediction

Reducing Features to Improve Bug Prediction

Learning From the Past with Experiment Databases

Calibration of Confidence Measures in Speech Recognition

ASTR 102: Introduction to Astronomy: Stars, Galaxies, and Cosmology

Laboratorio di Intelligenza Artificiale e Robotica

Rule Learning with Negation: Issues Regarding Effectiveness

Intelligent Agents. Chapter 2. Chapter 2 1

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Australian Journal of Basic and Applied Sciences

Exploration. CS : Deep Reinforcement Learning Sergey Levine

A Case Study: News Classification Based on Term Frequency

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Math 96: Intermediate Algebra in Context

Evaluation of Teach For America:

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017


Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

SAM - Sensors, Actuators and Microcontrollers in Mobile Robots

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Measurement. When Smaller Is Better. Activity:

Welcome to. ECML/PKDD 2004 Community meeting

Spring 2014 SYLLABUS Michigan State University STT 430: Probability and Statistics for Engineering

Chapter 2. Intelligent Agents. Outline. Agents and environments. Rationality. PEAS (Performance measure, Environment, Actuators, Sensors)

Probability and Game Theory Course Syllabus

DOCTOR OF PHILOSOPHY HANDBOOK

CS/SE 3341 Spring 2012

Foothill College Summer 2016

Learning Methods for Fuzzy Systems

Syllabus Foundations of Finance Summer 2014 FINC-UB

Word Segmentation of Off-line Handwritten Documents

MTH 141 Calculus 1 Syllabus Spring 2017

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Massachusetts Institute of Technology Tel: Massachusetts Avenue Room 32-D558 MA 02139

Using dialogue context to improve parsing performance in dialogue systems

Students Understanding of Graphical Vector Addition in One and Two Dimensions

Active Learning. Yingyu Liang Computer Sciences 760 Fall

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Paper Reference. Edexcel GCSE Mathematics (Linear) 1380 Paper 1 (Non-Calculator) Foundation Tier. Monday 6 June 2011 Afternoon Time: 1 hour 30 minutes

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Modeling user preferences and norms in context-aware systems

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Large-Scale Web Page Classification. Sathi T Marath. Submitted in partial fulfilment of the requirements. for the degree of Doctor of Philosophy

EGRHS Course Fair. Science & Math AP & IB Courses

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Semi-Supervised Face Detection

WHEN THERE IS A mismatch between the acoustic

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

EECS 700: Computer Modeling, Simulation, and Visualization Fall 2014

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

CS 101 Computer Science I Fall Instructor Muller. Syllabus

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Mining Student Evolution Using Associative Classification and Clustering

Issues in the Mining of Heart Failure Datasets

*In Ancient Greek: *In English: micro = small macro = large economia = management of the household or family

Learning Methods in Multilingual Speech Recognition

Spring 2015 IET4451 Systems Simulation Course Syllabus for Traditional, Hybrid, and Online Classes

Georgetown University at TREC 2017 Dynamic Domain Track

A survey of multi-view machine learning

Transcription:

Machine Learning: CS 6375 Introduction Instructor: Vibhav Gogate The University of Texas at Dallas

Logistics Instructor: Vibhav Gogate Email: Vibhav.Gogate@utdallas.edu Office: ECSS 3.406 Office hours: Mondays 4 p.m. to 6 p.m. TA: To be announced Web: http://www.hlt.utdallas.edu/~vgogate/ml/index.html Discussion Board Discussion board on Piazza. This will be the main on-line forum for discussing assignments and course material, and interacting with other students, TA and me. We will also post course-wide announcements on Piazza.

Evaluation Five homeworks (25%) 5% each Due two weeks later Some programming, some exercises Assigned via elearning. One Project (25%) One Midterm (15%) March 28, in class One Final (35%) May 14, ECSS 2.303, 2:00 to 4:45 p.m. Exams are closed book. You will be allowed a cheat sheet, a doublesided 8.5 x 11 page. A (90 or above), A- (85 to 89), B+ (80 to 84), B (75 to 79), B- (70 to 74), Fail (69 and below).

Source Materials T. Mitchell, Machine Learning, McGraw-Hill (Required/Recommended) C. Bishop, Pattern Recognition and Machine Learning, Springer (Required/Recommended) R. Duda, P. Hart & D. Stork, Pattern Classification (2 nd ed.), Wiley (Recommended) Papers

Why Study Machine Learning: A Few Quotes A breakthrough in machine learning would be worth ten Microsofts (Bill Gates, Microsoft) Machine learning is the next Internet (Tony Tether, Former Director, DARPA) Machine learning is the hot new thing (John Hennessy, President, Stanford) Web rankings today are mostly a matter of machine learning (Prabhakar Raghavan, Dir. Research, Yahoo) Machine learning is going to result in a real revolution (Greg Papadopoulos, CTO, Sun)

So What Is Machine Learning? Automating automation Getting computers to program themselves Writing software is the bottleneck Let the data do the work instead!

Traditional Programming Data Program Computer Output Machine Learning Data Output Computer Program

Magic? No, more like gardening Seeds = Algorithms Nutrients = Data Gardener = You Plants = Programs

Definition: Machine Learning! T. Mitchell: Well posed machine learning Improving performance via experience Formally, A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, it its performance at tasks in T as measured by P, improves with experience. H. Simon Learning denotes changes in the system that are adaptive in the sense that they enable the system to do the task or tasks drawn from the same population more efficiently and more effectively the next time. The ability to perform a task in a situation which has never been encountered before (Learning = Generalization)

Example 1: A Chess learning problem Task T: playing chess Performance measure P: percent of games won against opponents Training Experience E: playing practice games against itself

Example 2: Autonomous Vehicle Problem Task T: driving on a public highway/roads using vision sensors Performance Measure P: percentage of time the vehicle is involved in an accident Training Experience E: a sequence of images and steering commands recorded while observing a human driver

ML in a Nutshell Tens of thousands of machine learning algorithms Hundreds new every year Every machine learning algorithm has three components: Representation Evaluation Optimization

Representation Decision trees Sets of rules / Logic programs Instances Graphical models (Bayes/Markov nets) Neural networks Support vector machines Model ensembles Etc.

Accuracy Precision and recall Squared error Likelihood Posterior probability Cost / Utility Margin Entropy K-L divergence Etc. Evaluation

Optimization Combinatorial optimization E.g.: Greedy search Convex optimization E.g.: Gradient descent Constrained optimization E.g.: Linear programming

Types of Learning Supervised (inductive) learning Training data includes desired outputs Unsupervised learning Training data does not include desired outputs Find hidden structure in data Semi-supervised learning Training data includes a few desired outputs Reinforcement learning the learner interacts with the world via actions and tries to find an optimal policy of behavior with respect to rewards it receives from the environment

Types of Supervised Learning Problems Classification: learning to predict a discrete value from a predefined set of values Regression: learning to predict a continuous/real value

Machine Learning: Applications Examples of what you will study in class in action!

Classification Example: Spam Filtering Classify as Spam or Not Spam

Classification Example: Weather Prediction

Regression example: Predicting Gold/Stock prices Good ML can make you rich (but there is still some risk involved). Given historical data on Gold prices, predict tomorrow s price!

Similarity Determination

Collaborative Filtering The problem of collaborative filtering is to predict how well a user will like an item that he has not rated given a set of historical preference judgments for a community of users.

Collaborative Filtering

Collaborative Filtering

Clustering: Discover Structure in data

Machine learning has grown in leaps and bounds The main approach for Speech Recognition Robotics Natural Language Processing Computational Biology Sensor networks Computer Vision Web And so on Alice/Bob says: I know machine learning very well! Potential Employer: You are hired!!!

What We ll Cover Supervised learning: Decision tree induction, Rule induction, Instance-based learning, Bayesian learning, Neural networks, Support vector machines, Linear Regression, Model ensembles, Graphical models, Learning theory, etc. Unsupervised learning: Clustering, Dimensionality reduction Reinforcement learning: Markov Decision Processes, Q- learning, etc. General machine learning concepts and techniques: Feature selection, cross-validation, maximum likelihood estimation, gradient descent, expectation-maximization Your responsibility: Brush up on some important background Linear algebra, Statistics 101, Vectors, Probability theory