CS230: Lecture 5 Case Study

Similar documents
Lecture 1: Machine Learning Basics

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

Generative models and adversarial training

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Python Machine Learning

(Sub)Gradient Descent

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

CS Machine Learning

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Artificial Neural Networks written examination

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Forget catastrophic forgetting: AI that learns after deployment

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

San José State University Department of Psychology PSYC , Human Learning, Spring 2017

CS 446: Machine Learning

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

arxiv: v1 [cs.lg] 15 Jun 2015

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

INPE São José dos Campos

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Model Ensemble for Click Prediction in Bing Search Ads

Nutrition 10 Contemporary Nutrition WINTER 2016

Semi-Supervised Face Detection

STUDENTS' RATINGS ON TEACHER

Knowledge Transfer in Deep Convolutional Neural Nets

Cultivating DNN Diversity for Large Scale Video Labelling

Calibration of Confidence Measures in Speech Recognition

STAT 220 Midterm Exam, Friday, Feb. 24

WHEN THERE IS A mismatch between the acoustic

Data Structures and Algorithms

Rule Learning With Negation: Issues Regarding Effectiveness

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

arxiv: v1 [cs.cv] 10 May 2017

The University of Amsterdam s Concept Detection System at ImageCLEF 2011

Modeling function word errors in DNN-HMM based LVCSR systems

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Australian Journal of Basic and Applied Sciences

arxiv: v1 [cs.lg] 7 Apr 2015

The stages of event extraction

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

Rule Learning with Negation: Issues Regarding Effectiveness

CSC200: Lecture 4. Allan Borodin

Corrective Feedback and Persistent Learning for Information Extraction

Navigating the PhD Options in CMS

Linking Task: Identifying authors and book titles in verbose queries

B. How to write a research paper

CSL465/603 - Machine Learning

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Word Segmentation of Off-line Handwritten Documents

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Indian Institute of Technology, Kanpur

Speech Recognition at ICSI: Broadcast News and beyond

Applications of data mining algorithms to analysis of medical data

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

arxiv: v1 [cs.cl] 2 Apr 2017

MTH 215: Introduction to Linear Algebra

Deep Facial Action Unit Recognition from Partially Labeled Data

Course Specifications

A Web Based Annotation Interface Based of Wheel of Emotions. Author: Philip Marsh. Project Supervisor: Irena Spasic. Project Moderator: Matthew Morgan

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

SEMI-SUPERVISED ENSEMBLE DNN ACOUSTIC MODEL TRAINING

Visualizing Architecture

ACTL5103 Stochastic Modelling For Actuaries. Course Outline Semester 2, 2014

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

MYCIN. The MYCIN Task

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Dialog-based Language Learning

Evolution of Symbolisation in Chimpanzees and Neural Nets

A Case Study: News Classification Based on Term Frequency

An empirical study of learning speed in backpropagation

Accounting 312: Fundamentals of Managerial Accounting Syllabus Spring Brown

Test How To. Creating a New Test


Taste And Sight Anatomy Study Guide

Many instructors use a weighted total to calculate their grades. This lesson explains how to set up a weighted total using categories.

CS 101 Computer Science I Fall Instructor Muller. Syllabus

ReFresh: Retaining First Year Engineering Students and Retraining for Success

CS Course Missive

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

CHMB16H3 TECHNIQUES IN ANALYTICAL CHEMISTRY

Softprop: Softmax Neural Network Backpropagation Learning

Using EEG to Improve Massive Open Online Courses Feedback Interaction

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Number Line Moves Dash -- 1st Grade. Michelle Eckstein

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

GDP Falls as MBA Rises?

-Celebrating Your QI Success-

Evaluation of Respondus LockDown Browser Online Training Program. Angela Wilson EDTECH August 4 th, 2013

Why Did My Detector Do That?!

DOCENT VOLUNTEER EDUCATOR APPLICATION Winter Application Deadline: April 15, 2013

Transcription:

CS230: Lecture 5 Case Study Kian Katanforoosh

Problem statement: Live-Cell Detection Goal: determining which parts of a microscope image corresponds to which individual cells. Data: Doctors have collected 100,000 images from microscopes and gave them to you. Images have been taken from three types of microscopes: Type A Type B Type C 50,000 images 25,000 images 25,000 images Question: The doctors who hired you would like to use your algorithm on images from microscope C. How you would split this dataset into train, dev and test sets?

Data Question: The doctors who hired you would like to use your algorithm on images from microscope C. How you would split this dataset into train, dev and test sets? Answer: i) Split has to be roughly 90,5,5. Not 60,20,20. ii) Distribution of dev and test set have to be the same (contain images from C ). iii) There should be C images in the training as well, more than in the test/dev set. Question: Can you augment this dataset? If yes, give only 3 distinct methods you would use. If no, explain why (give only 2 reasons). Answer: Many augmentation methods would work in this case: cropping adding random noise changing contrast, blurring. flip rotate

Architecture and Loss Question: - What is the mathematical relation between nx and ny? - What s the last activation of your network? - What loss function should you use? Answer: i) nx = 3 ny ii) Sigmoid activation iii) Summation over all pixel value with cross entropy loss.

Transfer Learning First try: You have coded your neural network (model M1) and have trained it for 1000 epochs. It doesn t perform well. Transfer Learning: One of your friends suggested to use transfer learning using another labeled dataset made of 1,000,000 microscope images for skin disease classification (very similar images). A model (M2) has been trained on this dataset on a 10-class classification. Here is an example of input/output of the model M2. Question: You perform transfer learning from M2 to M1, what are the new hyperparameters that you ll have to tune?

Transfer Learning Question: You perform transfer learning from M2 to M1, what are the new hyperparameters that you ll have to tune?

Network modification Question: How can you correct your model and/or dataset to satisfy the doctors request? Answer: Modify the dataset in order to label the boundaries between cells. On top of that, change the loss function to give more weight to boundaries or penalize false positives.

Network modification New goal: They give you a dataset containing images similar to the previous ones. The difference is that each image is labeled as 0 (there are no cancer cells on the image) or 1 (there are cancer cells on the image). You easily build a state-of-the-art model to classify these images with 99% accuracy. The doctors are astonished and surprised, they ask you to explain your network s predictions. Question: Given an image classified as 1 (cancer present), how can you figure out based on which cell(s) the model predicted 1? Answer: Gradient of output w.r.t. input X Question: Your model detects cancer on cells (test set) images with 99% accuracy, while a doctor would on average perform 97% accuracy on the same task. Is this possible? Explain. Answer: If the dataset was entirely labeled by this one doctor with 97% accuracy, it is unlikely that the model can perform at 99% accuracy. However if annotated by multiple doctors, the network will learn from these several doctors and be able to outperform the one doctor with 97% accuracy. In this case, a panel composed of the doctors who labeled the data would likely perform at 99% accuracy or higher.

Duties for next week For next Tuesday 05/08, 9am: C4M1 Quiz: The basics of ConvNets Programming Assignment: Convolutional Neural Network - Step by Step Programming Assignment: Convolutional Neural Network - Application C4M2 Quiz: Convolutional models Programming Assignment: Keras Tutorial (optional, but highly recommended) Programming Assignment: Residual Networks Midterm, on 05/11: everything up to C4M2 (included) and next Tuesday s in-class lecture can be expected. This Friday (05/04): (optional) Hands-on TA session: GPU / Practical project advice