CS519: Deep Learning. Winter Fuxin Li

Similar documents
Python Machine Learning

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

(Sub)Gradient Descent

arxiv: v1 [cs.lg] 15 Jun 2015

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Lecture 1: Machine Learning Basics

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

CSL465/603 - Machine Learning

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

A Review: Speech Recognition with Deep Learning Methods

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Artificial Neural Networks written examination

arxiv: v1 [cs.cv] 10 May 2017

arxiv: v1 [cs.lg] 7 Apr 2015

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

THE world surrounding us involves multiple modalities

Residual Stacking of RNNs for Neural Machine Translation

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

arxiv: v2 [stat.ml] 30 Apr 2016 ABSTRACT

Generative models and adversarial training

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Calibration of Confidence Measures in Speech Recognition

SORT: Second-Order Response Transform for Visual Recognition

arxiv: v4 [cs.cl] 28 Mar 2016

CS224d Deep Learning for Natural Language Processing. Richard Socher, PhD

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Accounting 312: Fundamentals of Managerial Accounting Syllabus Spring Brown

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

Learning Methods for Fuzzy Systems

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Dropout improves Recurrent Neural Networks for Handwriting Recognition

arxiv: v1 [cs.cl] 27 Apr 2016

Second Exam: Natural Language Parsing with Neural Networks

Axiom 2013 Team Description Paper

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

A Deep Bag-of-Features Model for Music Auto-Tagging

Model Ensemble for Click Prediction in Bing Search Ads

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

San José State University Department of Psychology PSYC , Human Learning, Spring 2017

arxiv: v2 [cs.cv] 30 Mar 2017

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Softprop: Softmax Neural Network Backpropagation Learning

Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

arxiv: v1 [cs.dc] 19 May 2017

FF+FPG: Guiding a Policy-Gradient Planner

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

THE enormous growth of unstructured data, including

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

TD(λ) and Q-Learning Based Ludo Players

A study of speaker adaptation for DNN-based speech synthesis

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

Advanced Multiprocessor Programming

Learning to Schedule Straight-Line Code

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH AND LANGUAGE PROCESSING, VOL XXX, NO. XXX,

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

Human Emotion Recognition From Speech

Lip Reading in Profile

Improving Fairness in Memory Scheduling

Deep Neural Network Language Models

CS 3516: Computer Networks

What Can Neural Networks Teach us about Language? Graham Neubig a2-dlearn 11/18/2017

Evolution of Symbolisation in Chimpanzees and Neural Nets

Georgetown University at TREC 2017 Dynamic Domain Track

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

arxiv: v3 [cs.cl] 7 Feb 2017

Using Deep Convolutional Neural Networks in Monte Carlo Tree Search

MGMT3274 INTERNATONAL BUSINESS PROCESSES AND PROBLEMS

INTERMEDIATE ALGEBRA Course Syllabus

Knowledge Transfer in Deep Convolutional Neural Nets

MGT/MGP/MGB 261: Investment Analysis

Biscayne Bay Campus, Marine Science Building (room 250 D)

95723 Managing Disruptive Technologies

CS Course Missive

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

DNN ACOUSTIC MODELING WITH MODULAR MULTI-LINGUAL FEATURE EXTRACTION NETWORKS

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

CS Machine Learning

arxiv: v2 [cs.ir] 22 Aug 2016

Challenges in Deep Reinforcement Learning. Sergey Levine UC Berkeley

SEMI-SUPERVISED ENSEMBLE DNN ACOUSTIC MODEL TRAINING

INPE São José dos Campos

Lecture 1: Basic Concepts of Machine Learning

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Navigating the PhD Options in CMS

Math 181, Calculus I

The Good Judgment Project: A large scale test of different methods of combining expert predictions

SARDNET: A Self-Organizing Feature Map for Sequences

Dialog-based Language Learning

CRIJ 2328 Police Systems and Practices. Class Meeting Time:

Summarizing Answers in Non-Factoid Community Question-Answering

Streaming Video Control Review. Who am I?

Transcription:

CS519: Deep Learning Winter 2017 Fuxin Li

Course Information Instructor: Dr. Fuxin Li KEC 2077, lif@eecs.oregonstate.edu TA: Mingbo Ma: mam@oregonstate.edu Xu Xu: xux@oregonstate.edu My office hour: TBD (vote) Class Webpage: http://classes.engr.oregonstate.edu/eecs/winter2017/cs519-006/ Questions/Discussions on CANVAS

Prerequisites Significant knowledge on machine learning, especially the generics (not specific algorithms) CS 534 or equivalent knowledge Refresher will be provided in the next lecture Some knowledge of numerical optimization 1.5 weeks will be devoted to optimization and also deep network optimization

Grading Initial quiz (5%) based on participation only 3 Assignments (20%) No late assignments No downloading code from the Internet Quizzes (3 more quizzes totaling 30%) Based on whether you answer the questions correctly Final Project (45%) Final project is to be done with teams not more than 3 participants Grading will be done according to: Initial proposal (10%) Final oral presentation (10%) Final written presentation (25%)

Materials Book: I. Goodfellow, A. Courville, Y. Bengio. Deep Learning. MIT Press 2016. Electronic version: http://www.deeplearningbook.org/ More readings can be found at: http://deeplearning.net/reading-list/ http://colah.github.io/ http://karpathy.github.io/ https://www.coursera.org/course/neuralnets

Toolboxes A plethora of deep learning toolboxes around: Caffe Theano Torch TensorFlow CNTK, MXNet, Lasagne, Keras, Neon, etc. Toolbox policy: We stick to Keras for assignments (easiest learning curve) Final project: select the one you are most comfortable with

Outcome Understand the concepts of deep learning Gain some intuitions on deep networks Understand the training of deep learning Be able to use at least one deep learning toolbox to design and train a deep network Be able to design new algorithms and new architectures

What will be covered Basic neural network structure Training tricks (SGD, Momentum etc.) CNNs LSTMs Unsupervised neural networks Neural reinforcement learning (Dead week)

Final Project Groups of no more than 3 persons Jointly work on a significant project Must use deep learning CANNOT be just running an already-trained classifier on some images Try to solve a real problem One can elect projects from paper readings I will try to suggest some standard projects New neural architectures/changes to current architectures are welcome Grading based on the project merit, execution and presentation

Project Presentations 2 presentations for the final project Initial design (at least 1 month before finals week) Talk about what is your project about What you plan to do Re-grouping if several people are thinking about similar projects Final presentation (finals week) Need to identify who did what in the team 8 minutes per presentation Slides uploaded to a common computer Need to schedule 1 additional 2-hour session for it

Computing Resources Pelican cluster: 4 nodes with 2 GTX 980 Ti (6GB) each Accessible by SSH at pelican.eecs.oregonstate.edu Policy: 1 GPU per group otherwise risk your jobs be killed If you want to buy your own: Website will link you to a good article GTX Titan X PASCAL, GTX 1080 Ti (Mar 2017), GTX 1080, GTX 1070, GTX 1060 (sorted descendingly by price)

Approximate schedule (will be on website) Week 1 (Jan. 9-13) 1. Admin + General Introduction 2. Machine Learning Refresher (linear algorithms, empirical risk minimization, regularization, support vector machines) Week 2 (Jan. 16-20): Standard neural networks 3. Machine Learning Refresher (unfinished parts) + Basic Neural Networks with Hidden Layer (backpropagation) 4. Optimization Primer #1 (nonconvex optimization, stationary points and saddle points, optima, gradients) Week 3 (Jan. 23-27): Convolutional Networks 5. Convolutional Neural Networks (mostly in computer vision) 6. Continued CNN, Visualization of CNN Week 4 (Jan. 30 Feb. 3): Temporal Neural Models 7. Introduction of deep learning toolboxes (Caffe, Keras, automatic gradients) 8. Temporal Neural Models (RNNs and LSTMs) Week 5 (Feb. 6 Feb. 10): Deciding what project to work on 9. Continued Temporal Neural Models (LSTMs, GRUs, stacked together with CNNs) 10. An overview of other neural models Week 6 (Feb. 13-17): Project proposals 11. Project Proposals 12. Neural Network Optimization (stochastic mini-batch gradient descent, momentum, dropout, learning rate and weight decay)

Approximate schedule Week 7 (Feb. 20-24): Neural Network Optimization, Unsupervised Approaches 13. Neural Network Optimization (stochastic mini-batch gradient descent, momentum, dropout, learning rate and weight decay, automatic step-size methods) 14. Unsupervised Deep Learning (Autoencoders) Week 8 (Feb. 27 Mar. 3): Unsupervised Approaches, NLP applications 15. Unsupervised Deep Learning II (GANs) 16. Deep Learning in Natural Language Processing (Guest lecture from the Algorithms for Computational Linguistics group) Week 9 (Mar. 6 - Mar. 10): Deep Learning Frontiers 17. ResNet and New Architectures 18. Restricted Boltzmann Machines and Deep Belief Networks, convolutional DBN Week 10 (Mar. 13 - Mar. 17): Deep Reinforcement Learning 19. Deep reinforcement learning (guest lecture by Alan Fern) 20. Deep reinforcement learning (guest lecture by Alan Fern) Week 11 (Mar. 20 - Mar. 24): Finals Week 21. Project Presentations 22. Project Presentations