Machine Learning and Auto-Evaluation

Similar documents
Lecture 1: Basic Concepts of Machine Learning

CS 446: Machine Learning

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

CS Machine Learning

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Python Machine Learning

Assignment 1: Predicting Amazon Review Ratings

Learning From the Past with Experiment Databases

Decision Analysis. Decision-Making Problem. Decision Analysis. Part 1 Decision Analysis and Decision Tables. Decision Analysis, Part 1

Multivariate k-nearest Neighbor Regression for Time Series data -

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Truth Inference in Crowdsourcing: Is the Problem Solved?

Lecture 1: Machine Learning Basics

Generating Test Cases From Use Cases

Individual Component Checklist L I S T E N I N G. for use with ONE task ENGLISH VERSION

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Data Stream Processing and Analytics

Paper 2. Mathematics test. Calculator allowed. First name. Last name. School KEY STAGE TIER

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

(Sub)Gradient Descent

Model Ensemble for Click Prediction in Bing Search Ads

Unpacking a Standard: Making Dinner with Student Differences in Mind

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Carnegie Mellon University Department of Computer Science /615 - Database Applications C. Faloutsos & A. Pavlo, Spring 2014.

Science Fair Rules and Requirements

Linking Task: Identifying authors and book titles in verbose queries

Lecture 10: Reinforcement Learning

May To print or download your own copies of this document visit Name Date Eurovision Numeracy Assignment

Social Media Marketing BUS COURSE OUTLINE

Section 7, Unit 4: Sample Student Book Activities for Teaching Listening

EDEXCEL FUNCTIONAL SKILLS PILOT. Maths Level 2. Chapter 7. Working with probability

Visual CP Representation of Knowledge

Welcome to. ECML/PKDD 2004 Community meeting

An OO Framework for building Intelligence and Learning properties in Software Agents

Rule Learning With Negation: Issues Regarding Effectiveness

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Reducing Features to Improve Bug Prediction

CSL465/603 - Machine Learning

UCLA UCLA Electronic Theses and Dissertations

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Machine Learning and Development Policy

Rule Learning with Negation: Issues Regarding Effectiveness

Mathematics Success Grade 7

Modeling user preferences and norms in context-aware systems

Math Hunt th November, Sodalitas de Mathematica St. Xavier s College, Maitighar Kathmandu, Nepal

Name Class Date. Graphing Proportional Relationships

Social Media Journalism J336F Unique ID CMA Fall 2012

Australian Journal of Basic and Applied Sciences

1 3-5 = Subtraction - a binary operation

Test How To. Creating a New Test

preassessment was administered)

COMM370, Social Media Advertising Fall 2017

Axiom 2013 Team Description Paper

INSTRUCTIONAL FOCUS DOCUMENT Grade 5/Science

Cross-lingual Short-Text Document Classification for Facebook Comments

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

STAT 220 Midterm Exam, Friday, Feb. 24

A Vector Space Approach for Aspect-Based Sentiment Analysis

Learning Methods in Multilingual Speech Recognition

Natural Language Processing: Interpretation, Reasoning and Machine Learning

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Indian Institute of Technology, Kanpur

Redirected Inbound Call Sampling An Example of Fit for Purpose Non-probability Sample Design

Reinforcement Learning by Comparing Immediate Reward

Semi-Supervised Face Detection

Visit us at:

Distributed Weather Net: Wireless Sensor Network Supported Inquiry-Based Learning

Applications of data mining algorithms to analysis of medical data

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education

Leveraging Sentiment to Compute Word Similarity

AUTHOR COPY. Techniques for cold-starting context-aware mobile recommender systems for tourism

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Social Media Journalism J336F Unique Spring 2016

Optimizing to Arbitrary NLP Metrics using Ensemble Selection

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

M55205-Mastering Microsoft Project 2016

Cooperative evolutive concept learning: an empirical study

Ensemble Technique Utilization for Indonesian Dependency Parser

Detecting Online Harassment in Social Networks

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Welcome to ACT Brain Boot Camp

Competition in Information Technology: an Informal Learning

IBM Software Group. Mastering Requirements Management with Use Cases Module 6: Define the System

Towards Semantic Facility Data Management

Analyzing sentiments in tweets for Tesla Model 3 using SAS Enterprise Miner and SAS Sentiment Analysis Studio

arxiv: v2 [cs.cv] 30 Mar 2017

MKTG 611- Marketing Management The Wharton School, University of Pennsylvania Fall 2016

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

SETTING THE STAGE. News in Review January 2013 Teacher Resource Guide ROB FORD: Toronto s Controversial Mayor. Vocabulary Platform

Preliminary Report. 10. April. Bachelor thesis B16K06 Development of innovative solid multi zone heat storage system. for Hochschule Wismar

UNIT ONE Tools of Algebra

Mini Lesson Ideas for Expository Writing

Measurement & Analysis in the Real World

Radius STEM Readiness TM

A Coding System for Dynamic Topic Analysis: A Computer-Mediated Discourse Analysis Technique

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

Transcription:

Machine Learning and Auto-Evaluation In very simple terms, Machine Learning is about training or teaching computers to take decisions or actions without explicitly programming them. For example, whenever you read a tweet or movie review, you can figure out if the views expressed are positive or negative, but, can you teach a computer to determine the sentiment of that piece of text. This has many real life applications. Whenever Donald Trump makes a speech, Twitter goes crazy and there are both positive and negative sentiments all over the site. Trump s campaign management then determines what is the overall sentiment among the populace. Another example, Baidu predicted that Germany would take the World Cup 2014 before even the match was played. http://www.marketing-interactive.com/germany-will-win-world-cup-2014-baidu-predicts/ Weather Problem Lets take this small dataset of favourable weather conditions of playing some game. The problem here is to forecast whether one will be able to play the game in the given weather conditions. lssunny Hot High False No Definitions Feature/Attribute : Outlook, Temperature, Humidity and Windy w.r.t the above dataset are called features or attributes as they characterize or play a role in determine the favorable conditions. Outcome/Target: The problem is to predict what the outcome or target would be i.e whether you would be able to play the game or not. Vector: An ordered collection of features or attributes that correspond to a row in the dataset. Eg. (Sunny, Hot, High, False) corresponding to the first row is a vector with the target as No. ML Model: The algorithm or process or artifact that is generated as result of the learning process. There are lots of different types of ML models, e.g. Decision Trees, SVM, Naive Baiyes all capturing different aspects of the data. Error Metric/ Evaluation Metric: Whenever we train an ML model, we need to determine how correct or how well the model performs i.e how accurate the predictions are. There are again different types of evaluation metrics according to the type of ML problem.

Typical Learning Process

Supporting ML Problems on HackerEarth Our ML platform supports the machine learning flow as depicted in the above diagrams. We divide a given data set into training and test dataset. Training dataset is the data on which users have to train their models. Once the models are trained, users are expected to do the prediction on the test dataset and submit the prediction file. Test data that we give out to the users doesn t have the target variable. For e.g, for the following dataset of 10 rows, we can divide the dataset into train dataset with 60% of the rows and test dataset as the remaining 40% of the rows. Sunny Hot High False No Overcast Hot High False Yes Overcast Hot Normal False Yes Sunny Mild Normal True Yes Sunny Mild High False No Overcast Cool Normal True Yes Rainy Mild High True Yes Train Dataset (train.csv) Sunny Hot High False No Overcast Hot High False Yes Overcast Hot Normal False Yes Test Dataset (test.csv) Id Outlook Temperature Humidity Windy 1 Sunny Mild Normal True 2 Sunny Mild High False 3 Overcast Cool Normal True 4 Rainy Mild High True

Observe that in the test dataset, we hold out the target variable. Now, users train their models on train.csv, do the prediction on test.csv and submit a prediction file similar to the following file. User Prediction File (user_prediction.csv) Id Play 1 Yes 2 Yes 3 No 4 No On our server, we have the correct prediction file which we had separated out from the training set. Correct Prediction File(correct_prediction.csv) Id Play 1 Yes 2 No 3 Yes 4 Yes Evaluation Metric Once a user submits the prediction file, we give a score or evaluate the model performance only on 50% of the test dataset according to an evaluation metric which we define for the given problem. For the weather problem, we can define one of the evaluation metric as: Score = No of correct predictions / Total Rows For the submitted file, there is only 1 correct prediction for id=1 for the first 50% of the test dataset(id=1, 2), so, the score according to the above defined formula for the first two rows would be:- Score online = ½ = 0.5 Once the contest is over, we re-evaluate the file and give a score on the complete dataset. Now the score would be: Score offline = ¼ = 0.25 The point to observe here is that the score after the offline evaluation may go down or up. This is due to the overfitting i.e users may try to maximize their score during the online phase and

their models may not be generic enough to perform well on data outside of what was used in the training set. The practice of doing the evaluation on 50% of the test dataset during the online phase is to discourage overfitting by the users.