Linear classifiers: Scaling up learning via SGD
|
|
- Veronica McDonald
- 5 years ago
- Views:
Transcription
1 This image cannot currently be displayed. Linear classifiers: Scaling up learning via SGD Emily Fox University of Washington January 27, 2017 Stochastic gradient descent: Learning, one data point at a time 1
2 Stochastic gradient ascent w (t) Update coefficients w (t+1) Update coefficients w (t+2) Update coefficients w (t+3) Update coefficients w (t+4) Data Use only small subsets of data Compute gradient Many updates for each pass over data 3 Stochastic gradient ascent for logistic regression init w (1) =0, t=1 until for converged i=1,,n for j=0,,d partial[j] = Each time, pick Sum over different data point i data points w j (t+1) w j (t) + η partial[j] t t
3 Why would stochastic gradient ever work??? Gradient is direction of steepest ascent Gradient is best direction, but any direction that goes up would be useful 6 3
4 In ML, steepest direction is sum of little directions from each data point Sum over data points 7 For most data points, contribution points up Stochastic gradient: Pick a data point and move in direction Most of the time, total likelihood will increase 8 4
5 Stochastic gradient ascent: Most iterations increase likelihood, but sometimes decrease it On average, make progress until converged for i=1,,n for j=0,,d w (t+1) j w (t) j + η t t Convergence path 5
6 Convergence paths Gradient Stochastic gradient 11 Stochastic gradient convergence is noisy Stochastic gradient makes noisy progress Stochastic gradient achieves higher likelihood sooner, but it s noisier Better Avg. log likelihood Gradient usually increases likelihood smoothly 12 Total time proportional to # passes over data 6
7 Eventually, gradient catches up Note: should only trust average quality of stochastic gradient Better Avg. log likelihood Stochastic gradient Gradient 13 The last coefficients may be really good or really bad!! 14 Stochastic gradient will eventually oscillate around a solution w (1005) was good How do we minimize risk of picking bad coefficients w (1000) was bad Minimize noise: don t return last learned coefficients Output average: = 1 w (t) T 7
8 Summary of why stochastic gradient works Gradient finds direction of steepest ascent Gradient is sum of contributions from each data point Stochastic gradient uses direction from 1 data point On average increases likelihood, sometimes decreases 15 Stochastic gradient has noisy convergence Online learning: Fitting models from streaming data 8
9 Batch vs online learning Batch learning All data is available at start of training time Online learning Data arrives (streams in) over time - Must train model as data arrives! t=1 t=2 t=3 t=4 time Data ML algorithm Data Data Data Data ML (1) (2) (3) (4) algorithm Online learning example: Ad targeting Website Ad1 Ad2 Ad3 = Suggested ads User clicked on Ad2 y t =Ad2 18 Input: x t User info, page text ML algorithm (t) (t+1) 9
10 Online learning problem Data arrives over each time step t: - Observe input x t Info of user, text of webpage - Make a prediction t Which ad to show - Observe true output y t Which ad user clicked on 19 Need ML algorithm to update coefficients each time step! Stochastic gradient ascent can be used for online learning!!! init w (1) =0, t=1 Each time step t: - Observe input x t - Make a prediction t - Observe true output y t - Update coefficients: for j=0,,d w j (t+1) w j (t) + η 20 10
11 Summary of online learning Data arrives over time Must make a prediction every time new data point arrives Observe true class after prediction made Want to update parameters immediately 21 Summary of stochastic gradient descent 11
12 What you can do now Significantly speedup learning algorithm using stochastic gradient Describe intuition behind why stochastic gradient works Apply stochastic gradient in practice Describe online learning problems Relate stochastic gradient to online learning 23 Decision Trees Emily Fox University of Washington January 27,
13 Predicting potential loan defaults What makes a loan risky? I want a to buy a new house! Credit History Income Loan Application Term Personal Info 26 13
14 Credit history explained Did I pay previous loans on time? Example: excellent, good, or fair Credit History Income Term Personal Info 27 Income What s my income? Example: $80K per year Credit History Income Term Personal Info 28 14
15 Loan terms How soon do I need to pay the loan? Example: 3 years, 5 years, Credit History Income Term Personal Info 29 Personal information Credit History Age, reason for the loan, marital status, Example: Home loan for a married couple Income Term Personal Info 30 15
16 Intelligent application Loan Applications Intelligent loan application review system 31 Classifier review i = +1 Loan Application Input: x i Classifier MODEL Output: Predicted class i =
17 This module... decision trees Start excellent Credit? poor fair 3 years Term? 5 years high Income? Low Term? 3 years 5 years 33 Scoring a loan application Start x i = (Credit = poor, Income = high, Term = 5 years) excellent Credit? fair poor 3 years Term? 5 years high Income? Low Term? 3 years 5 years i = 34 17
18 Decision tree learning task Decision tree learning problem Training data: N observations (x i,y i ) Credit Term Income y excellent 3 yrs high safe fair 5 yrs low risky fair 3 yrs high safe poor 5 yrs high risky excellent 3 yrs low risky fair 5 yrs low safe poor 3 yrs high risky poor 5 yrs low safe fair 3 yrs high safe Optimize quality metric on training data T(X) 36 18
19 Quality metric: Classification error Error measures fraction of mistakes Error = # incorrect predictions # examples - Best possible value : Worst possible value: How do we find the best tree? Exponentially large number of possible trees makes decision tree learning hard! T 1 (X) T 2 (X) T 3 (X) Learning the smallest decision tree is an NP-hard problem [Hyafil & Rivest 76] T 4 (X) T 5 (X) T 6 (X) 38 19
20 Greedy decision tree learning Our training data table Assume N = 40, 3 features Credit Term Income y excellent 3 yrs high safe fair 5 yrs low risky fair 3 yrs high safe poor 5 yrs high risky excellent 3 yrs low risky fair 5 yrs low safe poor 3 yrs high risky poor 5 yrs low safe fair 3 yrs high safe 40 20
21 Start with all the data Loan status: (all data) # of loans # of loans N = 40 examples 41 Compact visual notation: node Loan status: # of loans # of loans N = 40 examples 42 21
22 Decision stump: Single level tree Loan status: Split on Credit Credit? excellent 9 0 fair 9 4 poor 4 14 Subset of data with Credit = excellent Subset of data with Credit = fair Subset of data with Credit = poor 43 Visual notation: Intermediate nodes Loan status: Credit? excellent 9 0 fair 9 4 poor Intermediate nodes 22
23 Making predictions with a decision stump Loan status: root excellent 9 0 credit? fair 9 4 poor 4 14 For each intermediate node, set = majority value 45 Selecting best feature to split on 23
24 How do we learn a decision stump? Loan status: Find the best feature to split on! Credit? excellent 9 0 fair 9 4 poor How do we select the best feature? Choice 1: Split on Credit Loan status: Choice 2: Split on Term Loan status: Credit? Term? excellent 9 0 fair 9 4 poor years years
25 How do we measure effectiveness of a split? Loan status: Credit? Idea: Calculate classification error of this decision stump excellent 9 0 fair 9 4 poor 4 14 Error = # mistakes # data points 49 Calculating classification error Step 1: = class of majority of data in node Step 2: Calculate classification error of predicting for this data Loan status: Error =. 22 correct = majority class 18 mistakes = Tree Classification error (root)
26 Choice 1: Split on Credit history? Choice 1: Split on Credit Loan status: Does a split on Credit reduce classification error below 0.45? Credit? excellent 9 0 fair 9 4 poor Split on Credit: Classification error Choice 1: Split on Credit Loan status: Credit? Error =. 52 excellent 9 0 fair 9 4 poor mistakes 4 mistakes 4 mistakes Tree = Classification error (root) 0.45 Split on credit
27 Choice 2: Split on Term? Choice 2: Split on Term Loan status: Term? 3 years years Evaluating the split on Term Choice 2: Split on Term Loan status: 54 3 years 16 4 Term? 5 years mistakes 6 mistakes Error =. = Tree Classification error (root) 0.45 Split on credit 0.2 Split on term
28 Choice 1 vs Choice 2: Comparing split on Credit vs Term Tree Classification error (root) 0.45 split on credit 0.2 split on loan term 0.25 Choice 1: Split on Credit Loan status: Choice 2: Split on Term Loan status: Credit? Term? excellent 9 0 fair 8 4 poor years years Feature split selection algorithm Given a subset of data M (a node in a tree) For each feature h i (x): 1. Split data of M according to feature h i (x) 2. Compute classification error split Chose feature h * (x) with lowest classification error 56 28
29 Recursion & Stopping conditions We ve learned a decision stump, what next? Loan status: Credit? excellent 9 0 fair 9 4 poor 4 14 All data points are nothing else to do with this subset of data 58 Leaf node 29
30 Tree learning = Recursive stump learning Loan status: Credit? excellent 9 0 fair 9 4 poor 4 14 Build decision stump with subset of data where Credit = fair Build decision stump with subset of data where Credit = poor 59 Second level Loan status: excellent 9 0 Credit? fair 9 4 Term? poor 4 14 Income? 3 years years 9 0 high 4 5 Low Build another stump these data points 30
31 Final decision tree Loan status: poor 4 14 Credit? Income? excellent 9 0 Fair 9 4 high 4 5 low 0 9 Term? Term? 3 years years years years Simple greedy decision tree learning Pick best feature to split on Learn decision stump with this split For each leaf of decision stump, recurse 62 When do we stop??? 31
32 Stopping condition 1: All data agrees on y All data in these nodes have same Loan y value status: Nothing to do Credit? poor 4 14 Income? excellent 9 0 Fair 9 4 high 4 5 low 0 9 Term? Term? 3 years years years years Stopping condition 2: Already split on all features Already split on all possible features Loan status: Nothing to do Credit? poor 4 14 Income? excellent 9 0 Fair 9 4 high 4 5 low 0 9 Term? Term? 3 years years years years
33 Greedy decision tree learning Step 1: Start with an empty tree Step 2: Select a feature to split data For each split of the tree: Step 3: If nothing more to, make predictions Step 4: Otherwise, go to Step 2 & continue (recurse) on this split Pick feature split leading to lowest classification error Stopping conditions 1 & 2 Recursion 65 Is this a good idea? Proposed stopping condition 3: Stop if no split reduces the classification error 66 33
34 Stopping condition 3: Don t stop if error doesn t decrease??? y = x[1] xor x[2] x[1] x[2] y y values True False 2 2 False False False False True True True False True Error =. True True False = Tree Classification error (root) Consider split on x[1] y = x[1] xor x[2] x[1] x[2] y y values True False 2 2 False False False False True True True False True x[1] Error =. True True False True 1 1 False 1 1 = 68 Tree Classification error (root) 0.5 Split on x[1]
35 Consider split on x[2] 69 y = x[1] xor x[2] x[1] x[2] y False False False False True True True False True True True False y values True False Neither features improve training error Stop now??? True x[2] False 1 1 Tree Error = = 0.5 Classification error (root) 0.5 Split on x[1] 0.5 Split on x[2] 0.5 Final tree with stopping condition 3 y = x[1] xor x[2] x[1] x[2] y False False False False True True True False True True True False y values True False 2 2 Predict True Tree with stopping condition 3 Classification error
36 Without stopping condition 3 Condition 3 (stopping when training error doesn t improve) is not recommended! y = x[1] xor x[2] x[1] x[2] y False False False False True True True False True True True False y values True False True x[1] False 1 1 x[2] x[2] Tree with stopping condition 3 without stopping condition 3 71 Classification error 0.5 True 0 1 False False 1 0 True True 1 0 True False 0 1 False Decision tree learning: Real valued features 36
37 How do we use real values inputs? Income Credit Term y $105 K excellent 3 yrs $112 K good 5 yrs $73 K fair 3 yrs $69 K excellent 5 yrs $217 K excellent 3 yrs $120 K good 5 yrs $64 K fair 3 yrs $340 K excellent 5 yrs $60 K good 3 yrs 73 Threshold split Loan status: Split on the feature Income Income? < $60K 8 13 >= $60K 14 5 Subset of data with Income >= $60K 74 37
38 Finding the best threshold split Infinite possible values of t Income = t * Income Income < t * Income >= t * $10K $120K 75 Consider a threshold between points Same classification error for any threshold split between v A and v B Income v A v B $10K $120K 76 38
39 Only need to consider mid-points Finite number of splits to consider Income $10K $120K 77 Threshold split selection algorithm Step 1: Sort the values of a feature h j (x) : Let {v 1, v 2, v 3, v N } denote sorted values Step 2: - For i = 1 N-1 Consider split t i = (v i + v i+1 ) / 2 Compute classification error for treshold split h j (x) >= t i - Chose the t * with the lowest classification error 78 39
40 Visualizing the threshold split Income Threshold split is the line Age = 38 $80K $40K $0K Age 79 Split on Age >= 38 Income age >= 38 age < 38 Predict $80K $40K Predict $0K Age 80 40
41 Depth 2: Split on Income >= $60K Income Threshold split is the line Income = 60K $80K $40K $0K Age 81 Each split partitions the 2-D space Income Age < 38 Age >= 38 Income >= 60K $80K $40K $0K Age >= 38 Income < 60K Age 82 41
42 Summary of decision trees What you can do now Define a decision tree classifier Interpret the output of a decision trees Learn a decision tree classifier using greedy algorithm Traverse a decision tree to make predictions - Majority class predictions - Probability predictions - Multiclass classification 84 42
(Sub)Gradient Descent
(Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include
More informationLecture 1: Machine Learning Basics
1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3
More informationMachine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler
Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina
More informationCS Machine Learning
CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing
More informationPython Machine Learning
Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled
More informationArtificial Neural Networks written examination
1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14
More informationExperiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling
Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Notebook for PAN at CLEF 2013 Andrés Alfonso Caurcel Díaz 1 and José María Gómez Hidalgo 2 1 Universidad
More informationA Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and
A Decision Tree Analysis of the Transfer Student Emma Gunu, MS Research Analyst Robert M Roe, PhD Executive Director of Institutional Research and Planning Overview Motivation for Analyses Analyses and
More informationAxiom 2013 Team Description Paper
Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association
More informationAssignment 1: Predicting Amazon Review Ratings
Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for
More informationThe Good Judgment Project: A large scale test of different methods of combining expert predictions
The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania
More informationActive Learning. Yingyu Liang Computer Sciences 760 Fall
Active Learning Yingyu Liang Computer Sciences 760 Fall 2017 http://pages.cs.wisc.edu/~yliang/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed by Mark Craven,
More informationIntroduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition
Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Todd Holloway Two Lecture Series for B551 November 20 & 27, 2007 Indiana University Outline Introduction Bias and
More informationModule 12. Machine Learning. Version 2 CSE IIT, Kharagpur
Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should
More informationCSL465/603 - Machine Learning
CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am
More informationTesting A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA
Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA Testing a Moving Target How Do We Test Machine Learning Systems? Peter Varhol, Technology
More informationAnalysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems
Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems Ajith Abraham School of Business Systems, Monash University, Clayton, Victoria 3800, Australia. Email: ajith.abraham@ieee.org
More informationExploration. CS : Deep Reinforcement Learning Sergey Levine
Exploration CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 4 due on Wednesday 2. Project proposal feedback sent Today s Lecture 1. What is exploration? Why is it a problem?
More informationChapter 2 Rule Learning in a Nutshell
Chapter 2 Rule Learning in a Nutshell This chapter gives a brief overview of inductive rule learning and may therefore serve as a guide through the rest of the book. Later chapters will expand upon the
More informationSoftware Maintenance
1 What is Software Maintenance? Software Maintenance is a very broad activity that includes error corrections, enhancements of capabilities, deletion of obsolete capabilities, and optimization. 2 Categories
More informationMachine Learning from Garden Path Sentences: The Application of Computational Linguistics
Machine Learning from Garden Path Sentences: The Application of Computational Linguistics http://dx.doi.org/10.3991/ijet.v9i6.4109 J.L. Du 1, P.F. Yu 1 and M.L. Li 2 1 Guangdong University of Foreign Studies,
More informationProof Theory for Syntacticians
Department of Linguistics Ohio State University Syntax 2 (Linguistics 602.02) January 5, 2012 Logics for Linguistics Many different kinds of logic are directly applicable to formalizing theories in syntax
More informationRadius STEM Readiness TM
Curriculum Guide Radius STEM Readiness TM While today s teens are surrounded by technology, we face a stark and imminent shortage of graduates pursuing careers in Science, Technology, Engineering, and
More informationModel Ensemble for Click Prediction in Bing Search Ads
Model Ensemble for Click Prediction in Bing Search Ads Xiaoliang Ling Microsoft Bing xiaoling@microsoft.com Hucheng Zhou Microsoft Research huzho@microsoft.com Weiwei Deng Microsoft Bing dedeng@microsoft.com
More informationOn the Combined Behavior of Autonomous Resource Management Agents
On the Combined Behavior of Autonomous Resource Management Agents Siri Fagernes 1 and Alva L. Couch 2 1 Faculty of Engineering Oslo University College Oslo, Norway siri.fagernes@iu.hio.no 2 Computer Science
More informationDetailed course syllabus
Detailed course syllabus 1. Linear regression model. Ordinary least squares method. This introductory class covers basic definitions of econometrics, econometric model, and economic data. Classification
More informationWE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT
WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT PRACTICAL APPLICATIONS OF RANDOM SAMPLING IN ediscovery By Matthew Verga, J.D. INTRODUCTION Anyone who spends ample time working
More informationLearning goal-oriented strategies in problem solving
Learning goal-oriented strategies in problem solving Martin Možina, Timotej Lazar, Ivan Bratko Faculty of Computer and Information Science University of Ljubljana, Ljubljana, Slovenia Abstract The need
More informationarxiv: v1 [cs.lg] 15 Jun 2015
Dual Memory Architectures for Fast Deep Learning of Stream Data via an Online-Incremental-Transfer Strategy arxiv:1506.04477v1 [cs.lg] 15 Jun 2015 Sang-Woo Lee Min-Oh Heo School of Computer Science and
More informationTIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy
TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE Pierre Foy TIMSS Advanced 2015 orks User Guide for the International Database Pierre Foy Contributors: Victoria A.S. Centurino, Kerry E. Cotter,
More informationProbability and Statistics Curriculum Pacing Guide
Unit 1 Terms PS.SPMJ.3 PS.SPMJ.5 Plan and conduct a survey to answer a statistical question. Recognize how the plan addresses sampling technique, randomization, measurement of experimental error and methods
More informationDiscriminative Learning of Beam-Search Heuristics for Planning
Discriminative Learning of Beam-Search Heuristics for Planning Yuehua Xu School of EECS Oregon State University Corvallis,OR 97331 xuyu@eecs.oregonstate.edu Alan Fern School of EECS Oregon State University
More informationSARDNET: A Self-Organizing Feature Map for Sequences
SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu
More informationEvolutive Neural Net Fuzzy Filtering: Basic Description
Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:
More informationMYCIN. The MYCIN Task
MYCIN Developed at Stanford University in 1972 Regarded as the first true expert system Assists physicians in the treatment of blood infections Many revisions and extensions over the years The MYCIN Task
More informationNotes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1
Notes on The Sciences of the Artificial Adapted from a shorter document written for course 17-652 (Deciding What to Design) 1 Ali Almossawi December 29, 2005 1 Introduction The Sciences of the Artificial
More informationOn-Line Data Analytics
International Journal of Computer Applications in Engineering Sciences [VOL I, ISSUE III, SEPTEMBER 2011] [ISSN: 2231-4946] On-Line Data Analytics Yugandhar Vemulapalli #, Devarapalli Raghu *, Raja Jacob
More informationSoftprop: Softmax Neural Network Backpropagation Learning
Softprop: Softmax Neural Networ Bacpropagation Learning Michael Rimer Computer Science Department Brigham Young University Provo, UT 84602, USA E-mail: mrimer@axon.cs.byu.edu Tony Martinez Computer Science
More informationMajor Milestones, Team Activities, and Individual Deliverables
Major Milestones, Team Activities, and Individual Deliverables Milestone #1: Team Semester Proposal Your team should write a proposal that describes project objectives, existing relevant technology, engineering
More informationLearning Methods in Multilingual Speech Recognition
Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex
More informationISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM
Proceedings of 28 ISFA 28 International Symposium on Flexible Automation Atlanta, GA, USA June 23-26, 28 ISFA28U_12 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Amit Gil, Helman Stern, Yael Edan, and
More informationFoothill College Summer 2016
Foothill College Summer 2016 Intermediate Algebra Math 105.04W CRN# 10135 5.0 units Instructor: Yvette Butterworth Text: None; Beoga.net material used Hours: Online Except Final Thurs, 8/4 3:30pm Phone:
More informationGo fishing! Responsibility judgments when cooperation breaks down
Go fishing! Responsibility judgments when cooperation breaks down Kelsey Allen (krallen@mit.edu), Julian Jara-Ettinger (jjara@mit.edu), Tobias Gerstenberg (tger@mit.edu), Max Kleiman-Weiner (maxkw@mit.edu)
More informationGenerative models and adversarial training
Day 4 Lecture 1 Generative models and adversarial training Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University What is a generative model?
More informationWHEN THERE IS A mismatch between the acoustic
808 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 14, NO. 3, MAY 2006 Optimization of Temporal Filters for Constructing Robust Features in Speech Recognition Jeih-Weih Hung, Member,
More informationFoothill College Fall 2014 Math My Way Math 230/235 MTWThF 10:00-11:50 (click on Math My Way tab) Math My Way Instructors:
This is a team taught directed study course. Foothill College Fall 2014 Math My Way Math 230/235 MTWThF 10:00-11:50 www.psme.foothill.edu (click on Math My Way tab) Math My Way Instructors: Instructor:
More informationHow to set up gradebook categories in Moodle 2.
How to set up gradebook categories in Moodle 2. It is possible to set up the gradebook to show divisions in time such as semesters and quarters by using categories. For example, Semester 1 = main category
More informationAn empirical study of learning speed in backpropagation
Carnegie Mellon University Research Showcase @ CMU Computer Science Department School of Computer Science 1988 An empirical study of learning speed in backpropagation networks Scott E. Fahlman Carnegie
More informationTHE UNIVERSITY OF SYDNEY Semester 2, Information Sheet for MATH2068/2988 Number Theory and Cryptography
THE UNIVERSITY OF SYDNEY Semester 2, 2017 Information Sheet for MATH2068/2988 Number Theory and Cryptography Websites: It is important that you check the following webpages regularly. Intermediate Mathematics
More informationTransfer Learning Action Models by Measuring the Similarity of Different Domains
Transfer Learning Action Models by Measuring the Similarity of Different Domains Hankui Zhuo 1, Qiang Yang 2, and Lei Li 1 1 Software Research Institute, Sun Yat-sen University, Guangzhou, China. zhuohank@gmail.com,lnslilei@mail.sysu.edu.cn
More informationCorrective Feedback and Persistent Learning for Information Extraction
Corrective Feedback and Persistent Learning for Information Extraction Aron Culotta a, Trausti Kristjansson b, Andrew McCallum a, Paul Viola c a Dept. of Computer Science, University of Massachusetts,
More informationCS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus
CS 1103 Computer Science I Honors Fall 2016 Instructor Muller Syllabus Welcome to CS1103. This course is an introduction to the art and science of computer programming and to some of the fundamental concepts
More informationThe Strong Minimalist Thesis and Bounded Optimality
The Strong Minimalist Thesis and Bounded Optimality DRAFT-IN-PROGRESS; SEND COMMENTS TO RICKL@UMICH.EDU Richard L. Lewis Department of Psychology University of Michigan 27 March 2010 1 Purpose of this
More informationLearning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for
Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for Email Marilyn A. Walker Jeanne C. Fromer Shrikanth Narayanan walker@research.att.com jeannie@ai.mit.edu shri@research.att.com
More informationLikelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition Seltzer, M.L.; Raj, B.; Stern, R.M. TR2004-088 December 2004 Abstract
More informationBeyond the Pipeline: Discrete Optimization in NLP
Beyond the Pipeline: Discrete Optimization in NLP Tomasz Marciniak and Michael Strube EML Research ggmbh Schloss-Wolfsbrunnenweg 33 69118 Heidelberg, Germany http://www.eml-research.de/nlp Abstract We
More informationAUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION
JOURNAL OF MEDICAL INFORMATICS & TECHNOLOGIES Vol. 11/2007, ISSN 1642-6037 Marek WIŚNIEWSKI *, Wiesława KUNISZYK-JÓŹKOWIAK *, Elżbieta SMOŁKA *, Waldemar SUSZYŃSKI * HMM, recognition, speech, disorders
More informationGetting Started with Deliberate Practice
Getting Started with Deliberate Practice Most of the implementation guides so far in Learning on Steroids have focused on conceptual skills. Things like being able to form mental images, remembering facts
More informationShockwheat. Statistics 1, Activity 1
Statistics 1, Activity 1 Shockwheat Students require real experiences with situations involving data and with situations involving chance. They will best learn about these concepts on an intuitive or informal
More informationGiven a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations
4 Interior point algorithms for network ow problems Mauricio G.C. Resende AT&T Bell Laboratories, Murray Hill, NJ 07974-2070 USA Panos M. Pardalos The University of Florida, Gainesville, FL 32611-6595
More informationCalibration of Confidence Measures in Speech Recognition
Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE
More informationLearning From the Past with Experiment Databases
Learning From the Past with Experiment Databases Joaquin Vanschoren 1, Bernhard Pfahringer 2, and Geoff Holmes 2 1 Computer Science Dept., K.U.Leuven, Leuven, Belgium 2 Computer Science Dept., University
More informationQuantitative analysis with statistics (and ponies) (Some slides, pony-based examples from Blase Ur)
Quantitative analysis with statistics (and ponies) (Some slides, pony-based examples from Blase Ur) 1 Interviews, diary studies Start stats Thursday: Ethics/IRB Tuesday: More stats New homework is available
More information1.11 I Know What Do You Know?
50 SECONDARY MATH 1 // MODULE 1 1.11 I Know What Do You Know? A Practice Understanding Task CC BY Jim Larrison https://flic.kr/p/9mp2c9 In each of the problems below I share some of the information that
More informationLecture 1: Basic Concepts of Machine Learning
Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010
More informationProbability and Game Theory Course Syllabus
Probability and Game Theory Course Syllabus DATE ACTIVITY CONCEPT Sunday Learn names; introduction to course, introduce the Battle of the Bismarck Sea as a 2-person zero-sum game. Monday Day 1 Pre-test
More informationReducing Features to Improve Bug Prediction
Reducing Features to Improve Bug Prediction Shivkumar Shivaji, E. James Whitehead, Jr., Ram Akella University of California Santa Cruz {shiv,ejw,ram}@soe.ucsc.edu Sunghun Kim Hong Kong University of Science
More informationBuild on students informal understanding of sharing and proportionality to develop initial fraction concepts.
Recommendation 1 Build on students informal understanding of sharing and proportionality to develop initial fraction concepts. Students come to kindergarten with a rudimentary understanding of basic fraction
More informationChallenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley
Challenges in Deep Reinforcement Learning Sergey Levine UC Berkeley Discuss some recent work in deep reinforcement learning Present a few major challenges Show some of our recent work toward tackling
More informationLaboratorio di Intelligenza Artificiale e Robotica
Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning
More informationUnderstanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010)
Understanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010) Jaxk Reeves, SCC Director Kim Love-Myers, SCC Associate Director Presented at UGA
More informationCollege Pricing and Income Inequality
College Pricing and Income Inequality Zhifeng Cai U of Minnesota, Rutgers University, and FRB Minneapolis Jonathan Heathcote FRB Minneapolis NBER Income Distribution, July 20, 2017 The views expressed
More informationImproving Fairness in Memory Scheduling
Improving Fairness in Memory Scheduling Using a Team of Learning Automata Aditya Kajwe and Madhu Mutyam Department of Computer Science & Engineering, Indian Institute of Tehcnology - Madras June 14, 2014
More informationMath 098 Intermediate Algebra Spring 2018
Math 098 Intermediate Algebra Spring 2018 Dept. of Mathematics Instructor's Name: Office Location: Office Hours: Office Phone: E-mail: MyMathLab Course ID: Course Description This course expands on the
More informationA Version Space Approach to Learning Context-free Grammars
Machine Learning 2: 39~74, 1987 1987 Kluwer Academic Publishers, Boston - Manufactured in The Netherlands A Version Space Approach to Learning Context-free Grammars KURT VANLEHN (VANLEHN@A.PSY.CMU.EDU)
More informationBridging Lexical Gaps between Queries and Questions on Large Online Q&A Collections with Compact Translation Models
Bridging Lexical Gaps between Queries and Questions on Large Online Q&A Collections with Compact Translation Models Jung-Tae Lee and Sang-Bum Kim and Young-In Song and Hae-Chang Rim Dept. of Computer &
More informationSCT Banner Student Fee Assessment Training Workbook October 2005 Release 7.2
SCT HIGHER EDUCATION SCT Banner Student Fee Assessment Training Workbook October 2005 Release 7.2 Confidential Business Information --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
More informationIterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages
Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages Nuanwan Soonthornphisaj 1 and Boonserm Kijsirikul 2 Machine Intelligence and Knowledge Discovery Laboratory Department of Computer
More informationLaboratorio di Intelligenza Artificiale e Robotica
Laboratorio di Intelligenza Artificiale e Robotica A.A. 2008-2009 Outline 2 Machine Learning Unsupervised Learning Supervised Learning Reinforcement Learning Genetic Algorithms Genetics-Based Machine Learning
More informationENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering
ENME 605 Advanced Control Systems, Fall 2015 Department of Mechanical Engineering Lecture Details Instructor Course Objectives Tuesday and Thursday, 4:00 pm to 5:15 pm Information Technology and Engineering
More informationMulti-label classification via multi-target regression on data streams
Mach Learn (2017) 106:745 770 DOI 10.1007/s10994-016-5613-5 Multi-label classification via multi-target regression on data streams Aljaž Osojnik 1,2 Panče Panov 1 Sašo Džeroski 1,2,3 Received: 26 April
More informationAP Calculus AB. Nevada Academic Standards that are assessable at the local level only.
Calculus AB Priority Keys Aligned with Nevada Standards MA I MI L S MA represents a Major content area. Any concept labeled MA is something of central importance to the entire class/curriculum; it is a
More informationTruth Inference in Crowdsourcing: Is the Problem Solved?
Truth Inference in Crowdsourcing: Is the Problem Solved? Yudian Zheng, Guoliang Li #, Yuanbing Li #, Caihua Shan, Reynold Cheng # Department of Computer Science, Tsinghua University Department of Computer
More informationAn Introduction to Simulation Optimization
An Introduction to Simulation Optimization Nanjing Jian Shane G. Henderson Introductory Tutorials Winter Simulation Conference December 7, 2015 Thanks: NSF CMMI1200315 1 Contents 1. Introduction 2. Common
More informationContinual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots
Continual Curiosity-Driven Skill Acquisition from High-Dimensional Video Inputs for Humanoid Robots Varun Raj Kompella, Marijn Stollenga, Matthew Luciw, Juergen Schmidhuber The Swiss AI Lab IDSIA, USI
More informationManipulative Mathematics Using Manipulatives to Promote Understanding of Math Concepts
Using Manipulatives to Promote Understanding of Math Concepts Multiples and Primes Multiples Prime Numbers Manipulatives used: Hundreds Charts Manipulative Mathematics 1 www.foundationsofalgebra.com Multiples
More informationAn Effective Framework for Fast Expert Mining in Collaboration Networks: A Group-Oriented and Cost-Based Method
Farhadi F, Sorkhi M, Hashemi S et al. An effective framework for fast expert mining in collaboration networks: A grouporiented and cost-based method. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 27(3): 577
More informationPredicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks
Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks Devendra Singh Chaplot, Eunhee Rhim, and Jihie Kim Samsung Electronics Co., Ltd. Seoul, South Korea {dev.chaplot,eunhee.rhim,jihie.kim}@samsung.com
More informationProbabilistic Latent Semantic Analysis
Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview
More informationarxiv: v1 [math.at] 10 Jan 2016
THE ALGEBRAIC ATIYAH-HIRZEBRUCH SPECTRAL SEQUENCE OF REAL PROJECTIVE SPECTRA arxiv:1601.02185v1 [math.at] 10 Jan 2016 GUOZHEN WANG AND ZHOULI XU Abstract. In this note, we use Curtis s algorithm and the
More informationABSTRACT. A major goal of human genetics is the discovery and validation of genetic polymorphisms
ABSTRACT DEODHAR, SUSHAMNA DEODHAR. Using Grammatical Evolution Decision Trees for Detecting Gene-Gene Interactions in Genetic Epidemiology. (Under the direction of Dr. Alison Motsinger-Reif.) A major
More informationRule Learning With Negation: Issues Regarding Effectiveness
Rule Learning With Negation: Issues Regarding Effectiveness S. Chua, F. Coenen, G. Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX Liverpool, United
More informationComment-based Multi-View Clustering of Web 2.0 Items
Comment-based Multi-View Clustering of Web 2.0 Items Xiangnan He 1 Min-Yen Kan 1 Peichu Xie 2 Xiao Chen 3 1 School of Computing, National University of Singapore 2 Department of Mathematics, National University
More informationActivity 2 Multiplying Fractions Math 33. Is it important to have common denominators when we multiply fraction? Why or why not?
Activity Multiplying Fractions Math Your Name: Partners Names:.. (.) Essential Question: Think about the question, but don t answer it. You will have an opportunity to answer this question at the end of
More informationCal s Dinner Card Deals
Cal s Dinner Card Deals Overview: In this lesson students compare three linear functions in the context of Dinner Card Deals. Students are required to interpret a graph for each Dinner Card Deal to help
More informationTeam Formation for Generalized Tasks in Expertise Social Networks
IEEE International Conference on Social Computing / IEEE International Conference on Privacy, Security, Risk and Trust Team Formation for Generalized Tasks in Expertise Social Networks Cheng-Te Li Graduate
More informationNavigating the PhD Options in CMS
Navigating the PhD Options in CMS This document gives an overview of the typical student path through the four Ph.D. programs in the CMS department ACM, CDS, CS, and CMS. Note that it is not a replacement
More informationEmotional Variation in Speech-Based Natural Language Generation
Emotional Variation in Speech-Based Natural Language Generation Michael Fleischman and Eduard Hovy USC Information Science Institute 4676 Admiralty Way Marina del Rey, CA 90292-6695 U.S.A.{fleisch, hovy}
More informationAlgorithms and Data Structures (NWI-IBC027)
Algorithms and Data Structures (NWI-IBC027) Frits Vaandrager F.Vaandrager@cs.ru.nl Institute for Computing and Information Sciences 7th September 2017 Frits Vaandrager 7th September 2017 Lecture 1 1 /
More informationDegreeWorks Advisor Reference Guide
DegreeWorks Advisor Reference Guide Table of Contents 1. DegreeWorks Basics... 2 Overview... 2 Application Features... 3 Getting Started... 4 DegreeWorks Basics FAQs... 10 2. What-If Audits... 12 Overview...
More information