Development of Deep Learning & Attitude of Sharing. Jooyoul Lee LG CNS

Similar documents
Python Machine Learning

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

CSL465/603 - Machine Learning

(Sub)Gradient Descent

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Generative models and adversarial training

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Lecture 1: Machine Learning Basics

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Artificial Neural Networks written examination

A Neural Network GUI Tested on Text-To-Phoneme Mapping

arxiv: v1 [cs.cv] 10 May 2017

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

arxiv: v1 [cs.lg] 15 Jun 2015

Axiom 2013 Team Description Paper

CS Machine Learning

A Compact DNN: Approaching GoogLeNet-Level Accuracy of Classification and Domain Adaptation

Model Ensemble for Click Prediction in Bing Search Ads

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

SORT: Second-Order Response Transform for Visual Recognition

CS224d Deep Learning for Natural Language Processing. Richard Socher, PhD

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Attributed Social Network Embedding

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Indian Institute of Technology, Kanpur

Second Exam: Natural Language Parsing with Neural Networks

Top US Tech Talent for the Top China Tech Company

Modeling function word errors in DNN-HMM based LVCSR systems

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Residual Stacking of RNNs for Neural Machine Translation

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Human Emotion Recognition From Speech

arxiv: v4 [cs.cl] 28 Mar 2016

Forget catastrophic forgetting: AI that learns after deployment

Taxonomy-Regularized Semantic Deep Convolutional Neural Networks

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

Modeling function word errors in DNN-HMM based LVCSR systems

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

Learning Methods for Fuzzy Systems

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Artificial Neural Networks

Learning From the Past with Experiment Databases

Using Deep Convolutional Neural Networks in Monte Carlo Tree Search

arxiv: v1 [cs.cl] 27 Apr 2016

Calibration of Confidence Measures in Speech Recognition

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

Chamilo 2.0: A Second Generation Open Source E-learning and Collaboration Platform

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

Issues in the Mining of Heart Failure Datasets

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

A Review: Speech Recognition with Deep Learning Methods

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Softprop: Softmax Neural Network Backpropagation Learning

Lip Reading in Profile

Lecture 1: Basic Concepts of Machine Learning

Knowledge Transfer in Deep Convolutional Neural Nets

Word Segmentation of Off-line Handwritten Documents

Deep Neural Network Language Models

arxiv:submit/ [cs.cv] 2 Aug 2017

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

CS 446: Machine Learning

arxiv: v2 [cs.cv] 30 Mar 2017

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Laboratorio di Intelligenza Artificiale e Robotica

Cultivating DNN Diversity for Large Scale Video Labelling

THE enormous growth of unstructured data, including

Henry Tirri* Petri Myllymgki

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

GACE Computer Science Assessment Test at a Glance

MYCIN. The MYCIN Task

Discriminative Learning of Beam-Search Heuristics for Planning

Rule Learning With Negation: Issues Regarding Effectiveness

arxiv: v1 [cs.lg] 7 Apr 2015

Courses in English. Application Development Technology. Artificial Intelligence. 2017/18 Spring Semester. Database access

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Laboratorio di Intelligenza Artificiale e Robotica

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Dropout improves Recurrent Neural Networks for Handwriting Recognition

We are strong in research and particularly noted in software engineering, information security and privacy, and humane gaming.

Probabilistic Latent Semantic Analysis

AI Agent for Ice Hockey Atari 2600

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Beyond the Pipeline: Discrete Optimization in NLP

Rule Learning with Negation: Issues Regarding Effectiveness

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

EdX Learner s Guide. Release

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

arxiv: v2 [cs.cv] 4 Mar 2016

arxiv: v2 [cs.ir] 22 Aug 2016

INPE São José dos Campos

Transcription:

Development of Deep Learning & Attitude of Sharing Jooyoul Lee LG CNS

Agenda 1. Deep Learning overview 2. Why is Deep Learning growing so fast? 3. Deep Learning Tools & Open Source 4. Wrap-Up

1. Deep Learning overview

AI, Machine Learning, Deep Learning Strong AI Weak AI Machine Learning Artificial Neural Networks Deep Learning

Categorical Continuous 1.1 Machine Learning & Neural Nets Machine Learning Algorithms Unsupervised Supervised Clustering & Dimensionality reduction K-means LDA(Latent Dirichlet Allocation) SVD(Singular Value Decomposition) PCA(Principal Component Analysis) Regression Liner Polynomial Decision Tree Random Forests Artificial Neural Networks Gaussian Mixture HMM(Hidden Markov Model) Classification Logistic Regression Naïve-Bayes SVM(Support Vector Machine)

1.1 Machine Learning & Neural Nets Artificial Neural Networks (ANN) sigmoid

1.1 Machine Learning & Neural Nets More hidden layer, more powerful Input layer Hidden layer Output layer [Perceptron] [Multi-Layer Perceptron] Hidden Node

1.1 Machine Learning & Neural Nets But, More hidden layer, hard to train! Why? The vanishing gradient problem & so on Michael Nielsen (2015), http://neuralnetworksanddeeplearning.com/

1.1 Machine Learning & Neural Nets Breakthrough!

1.2 What is Deep Learning Deep Neural Networks + Machine Learning = Deep Learning Neural Networks - 1~2 hidden layers Deep Neural Networks - More hidden layers, complex

1.2 What is Deep Learning Deep Neural Networks + Machine Learning = Deep Learning Neural Networks - 1~2 hidden layers Deep Neural Networks - More hidden layers, complex

1.2 What is Deep Learning Deep Neural Networks + Machine Learning = Deep Learning Fully-Connected / Undirected DNN(Deep Neural Networks) RBM(Restricted Boltzmann Machine) DBN(Deep Belief Network) Convolutional LeNet AlexNet, VGGNet GoogleNet ResNet Recurrent LSTM(Long Short-Term Memory) GRU(Gated Recurrent Unit) Memory Networks

1.2 What is Deep Learning Deep Neural Networks + Machine Learning = Deep Learning Activation Function ReLU(Rectified Linear Units) Leaky ReLU Maxout Regularization Drop-Out Batch Normalization Optimization SGD(Stochastic Gradient Descent) AdaGrad RMSprop Adam

Error rate on ImageNet (Top 5) 1.3 Why Deep Learning? The state-of-the-art in speech recognition, visual object recognition Using Deep Learning Speech Recognition Visual Object Recognition (ILSVRC 1) ) 1) ImageNet Large Scale Visual Recognition Challenge (ILSVRC)

1.3 Why Deep Learning? Deep Learning Everywhere

1.3 Why Deep Learning? Why is Deep Learning working so well? Learning multiple levels of representation/abstraction Output Even more abstract features More abstract features Features Input

2. Why is Deep Learning growing so fast?

2.1 Why is Deep Learning growing so fast? Big Data & High Performance Computing(HPC) Big Data Deep Learning High Performance Computing (e.g. GPU, Distributed computing 1) ) 1) Hadoop, Spark

2.1 Why is Deep Learning growing so fast? The shoulders of giants If I have seen further than others, it is by standing Upon the shoulders of giants. Issac Newton

2.1 Why is Deep Learning growing so fast? Open Access & Open Source Deep Learning Open Access Open Source

2.2 Open Access Open access refers to online research outputs that are free of all restrictions on access (e.g. access tolls) and free of many restrictions on use (e.g. certain copyright and license restrictions). _from Wikipedia

2.2 Open Access arxiv.org : repository of e-prints of scientific papers Open access to e-prints in Physics, Mathematics, Computer Science, Quantitative Biology, Quantitative Finance and Statistics Many Deep Learning researchers submit papers So, you can find NIPS, ICML, ICLR 1) papers at arxiv.org 1) NIPS : Conference on Neural Information Processing Systems ICML : International Conference on Machine Learning ICLR : International Conference on Learning Representation

2.3 Open Source GitHub : a web-based Git repository hosting service Many Deep Learning researchers open the algorithm implementation codes The codes are mainly released under BSD, MIT, Apache license Open Access Open Source

2.4 GitHub + arxiv GitXiv.com : a space to share collaborative open computer science projects

3. Deep Learning Tools & Open Source

3.1 Why is Deep Learning growing so fast? Remind The shoulders of giants Deep Learning Open Source Deep Learning Tools

3.1 Why is Deep Learning growing so fast? Implement a deep neural networks Deep Neural Nets Computational Graphs Code (call API of Deep Learning Tool)

3.2 Deep Learning Tools So many Open Source Deep Learning Tools

3.2 Deep Learning Tools Caffe : Convolutional Architecture for Fast Feature Embedding From U.C. Berkeley, BSD 2-Clause license, Written in C++ Pros and Cons 1) : (+) Good for feedforward networks and image processing (+) Good for finetuning existing networks (+) Train models without writing any code (+) Python interface is pretty useful (-) Need to write C++ / CUDA for new GPU layers (-) Not good for recurrent neural networks (-) Not extensible, bit of a hairball 1) Stanford CS231n Andrej Karpathy, https://deeplearning4j.org/compare-dl4j-torch7-pylearn#caffe

3.2 Deep Learning Tools Torch From NYU, BSD license, Written in C/C++, Lua Used a lot in Facebook and DeepMind Pros and Cons 1) : (+) Lots of modular pieces that are easy to combine (+) Easy to write your own layer types and run on GPU (+) Lots of pre-trained models (-) Lua (-) You usually write your own training code (Less plug and play) 1) Stanford CS231n Andrej Karpathy, https://deeplearning4j.org/compare-dl4j-torch7-pylearn#torch

3.2 Deep Learning Tools Theano From University of Montreal, Written in Python High-level wrappers: Keras, Lasagne Pros and Cons 1) : (+) Python + Numpy (+) Computational graph is nice abstraction (+) RNNs fit nicely in computational graph (-) Hard to navigate, debug, refactor (-) Large models can have long compile times (-) Not enough pre-trained models 1) Stanford CS231n Andrej Karpathy, https://deeplearning4j.org/compare-dl4j-torch7-pylearn#theano

3.2 Deep Learning Tools TensorFlow From Google, Apache 2.0 license, Written in C++, Python Very similar to Theano Pros and Cons 1) : (+) Python + Numpy (+) Computational graph abstraction, like Theano (+) Easy Visualizations (+) Multi-node training (-) Hard to hack into the networks (-) Training time is slower than others right now (-) Not enough pre-trained models (but it will get better) 1) Stanford CS231n Andrej Karpathy, https://deeplearning4j.org/compare-dl4j-torch7-pylearn#tensorflow

3.2 Deep Learning Tools Overview 1) Caffe Torch Theano TensorFlow Language C++, Python Lua Python Python Pre-trained Model Yes ++ Yes ++ Yes (Lasagne) Not enough Multi-GPU: Data parallel Yes Yes Yes Yes Multi-GPU: Model parallel No Yes Experimental Yes (best) Readable source code Yes (C++) Yes (Lua) No No Good at RNN No Mediocre Yes Yes (best) 1) Stanford CS231n Andrej Karpathy

3.3 Top Deep Learning Projects Popular GitHub Projects related to Deep Learning 1) 1) https://github.com/aymericdamien/topdeeplearning

4. Wrap-Up

4.1 Open Access, Source & Dataset Open Dataset Deep Learning Open Access Open Source Open Dataset

4.1 Open Access, Source & Dataset Open Dataset ImageNet Microsoft COCO Dataset Google Open Images Dataset Youtube 8M Dataset Facebook Question Answering Dataset And so on

4.2 Become a Deep Learning Hacker You can become a Deep Learning Hacker! Open Access Open Source Open Dataset And Open Courseware (MOOC). Coursera. Udacity. edx

4.2 Become a Deep Learning Hacker George Hotz (comma.ai CEO) Hacked the ios device, Playstation3 Developed Deep Learning based self-driving car

4.2 Become a Deep Learning Hacker Jeff Dean (Google Senior Fellow) GFS(Google File System), MapReduce, BigTable TensorFlow, Google Brain Team Leader

Thank you.