An Analysis of Single-Layer Networks in Unsupervised Feature Learning. Adam Coates1, Honglak Lee2, Andrew Y. Ng1

Similar documents
Generative models and adversarial training

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Lecture 1: Machine Learning Basics

Python Machine Learning

A study of speaker adaptation for DNN-based speech synthesis

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Offline Writer Identification Using Convolutional Neural Network Activation Features

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

arxiv: v2 [cs.cv] 30 Mar 2017

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Word Segmentation of Off-line Handwritten Documents

Probabilistic Latent Semantic Analysis

arxiv: v1 [cs.lg] 15 Jun 2015

Modeling function word errors in DNN-HMM based LVCSR systems

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Calibration of Confidence Measures in Speech Recognition

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Modeling function word errors in DNN-HMM based LVCSR systems

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

CS Machine Learning

A Review: Speech Recognition with Deep Learning Methods

Semantic Segmentation with Histological Image Data: Cancer Cell vs. Stroma

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Deep Neural Network Language Models

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Human Emotion Recognition From Speech

A Deep Bag-of-Features Model for Music Auto-Tagging

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Speech Emotion Recognition Using Support Vector Machine

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Model Ensemble for Click Prediction in Bing Search Ads

Digital Signal Processing: Speaker Recognition Final Report (Complete Version)

Knowledge Transfer in Deep Convolutional Neural Nets

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Linking Task: Identifying authors and book titles in verbose queries

WHEN THERE IS A mismatch between the acoustic

Artificial Neural Networks written examination

Learning Methods for Fuzzy Systems

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Multi-tasks Deep Learning Model for classifying MRI images of AD/MCI Patients

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

DOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS. Elliot Singer and Douglas Reynolds

(Sub)Gradient Descent

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

Deep Facial Action Unit Recognition from Partially Labeled Data

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

INPE São José dos Campos

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Using focal point learning to improve human machine tacit coordination

Cultivating DNN Diversity for Large Scale Video Labelling

A Survey on Unsupervised Machine Learning Algorithms for Automation, Classification and Maintenance

DNN ACOUSTIC MODELING WITH MODULAR MULTI-LINGUAL FEATURE EXTRACTION NETWORKS

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

CSL465/603 - Machine Learning

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

Lecture 1: Basic Concepts of Machine Learning

Automatic Speaker Recognition: Modelling, Feature Extraction and Effects of Clinical Environment

Switchboard Language Model Improvement with Conversational Data from Gigaword

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Taxonomy-Regularized Semantic Deep Convolutional Neural Networks

Assignment 1: Predicting Amazon Review Ratings

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

arxiv: v1 [cs.cl] 2 Apr 2017

Action Recognition and Video

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Rule Learning With Negation: Issues Regarding Effectiveness

CS 446: Machine Learning

Semi-Supervised Face Detection

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Second Exam: Natural Language Parsing with Neural Networks

Computerized Adaptive Psychological Testing A Personalisation Perspective

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

K-Medoid Algorithm in Clustering Student Scholarship Applicants

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

A Compact DNN: Approaching GoogLeNet-Level Accuracy of Classification and Domain Adaptation

THE world surrounding us involves multiple modalities

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Rule Learning with Negation: Issues Regarding Effectiveness

Attributed Social Network Embedding

Speaker Identification by Comparison of Smart Methods. Abstract

Test Effort Estimation Using Neural Network

On the Formation of Phoneme Categories in DNN Acoustic Models

Comment-based Multi-View Clustering of Web 2.0 Items

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

FSL-BM: Fuzzy Supervised Learning with Binary Meta-Feature for Classification

TABLE OF CONTENTS TABLE OF CONTENTS COVER PAGE HALAMAN PENGESAHAN PERNYATAAN NASKAH SOAL TUGAS AKHIR ACKNOWLEDGEMENT FOREWORD

Transcription:

An Analysis of Single-Layer Networks in Unsupervised Feature Learning Adam Coates1, Honglak Lee2, Andrew Y. Ng1

Overview A Brief Introduction Unsupervised feature learning framework Experiments and Analysis Q&A

A Brief Intro Recent works focus on learning good feature representation greedily pre-train several layers of feature for each layer a set of parameters are chosen: Number of features to learn the location of the features encoding input and output A Major drawback is complexity and expense

A Brief Intro In this paper : First study the effect of these choices on single layer network It turns out that there are other ingredients: Whitening large number of features dense feature extraction

A Brief Intro What they did : Used a simple feature learning framework that incorporates an unsupervised learning algorithm as a black box Analyze performance impact of : whitening number of features trained step size between extracted features receptive field size

Unsupervised feature learning framework Steps to learn features : 1. Extract random patches form unlabeled training data 2. Apply pre-processing stage 3. Learn a feature mapping After learning features : 1. Extract features from equally spaced sub-patches 2. Pool features together to reduce number of feature values 3. Train a linear classifier to predict labels

Unsupervised feature learning framework Extract random sub-patches : each patch has dimension w-by-w and d channels each patch can be represented as a vector of size w.w.d the dataset consists of m randomly sampled patches

Unsupervised feature learning framework Pre-Processing : Normalize the data by subtracting the mean and dividing by standard deviation Perform whitening

Whitening

Unsupervised feature learning framework Unsupervised Learning : the black box takes dataset X and outputs a function f :! N! K that maps input vector to a feature vector of size K sparse auto-encoders sparse RBMs K-means clustering Gaussian mixtures

Unsupervised feature learning framework Feature Extraction and Classification using the learned feature extractor, given any image patch compute a representation for the patch do this for many sub-patches of images

Unsupervised feature learning framework Figure 1: Illustration showing feature extraction using a -by- receptive field and stride. We firs (a) K-means (with and without whitening) (b) GMM (with and without whitening) (c) Sparse Autoencoder (with and without whitening) (d) Sparse RBM (with and without whitening)

Experiments The above framework includes number of parameters : whitening number of features the step size receptive field

Experiments 80 Performance for Raw and Whitened Inputs 80 75 75 Cross Validation Accuracy (%) 70 65 60 55 50 kmeans (tri) raw kmeans (hard) raw gmm raw 60 autoencoder raw rbm raw kmeans (tri) white kmeans (hard) white 55 gmm white autoencoder white rbm white 100 200 400 800 1200 1600 50 # Features Figure 3: Effect of whitening and number of bases (or centroids). 70 65

Experiments 78 76 74 Performance vs. Receptive Field Size 78 76 74 80 75 Performance vs. Feature Stride kmeans (tri) kmeans (hard) autoencoder rbm 80 75 Cross Validation Accuracy (%) 72 70 68 66 72 70 68 66 Cross Validation Accuracy (%) 70 65 60 55 70 65 60 55 64 64 50 50 62 60 6 8 12 Receptive field size (pixels) kmeans (tri) kmeans (hard) autoencoder rbm Figure Figure 4: Effect 4: Effect of receptive of field size and stride. 62 60 45 40 1 2 4 8 Stride between extracted features (pixels) 45 40 e present our experimental results on the impact of these parameters on performance.

Experiments Table 1: Test recognition accuracy (and error) for NORB (normalized-uniform Algorithm Test accuracy (and error) Convolutional Neural Networks [14] 93.4% (6.6%) Deep Boltzmann Machines [25] 92.8% (7.2%) Deep Belief Networks [18] 95.0% (5.0%) (Best result of [10]) 94.4% (5.6%) K-means (Triangle) 97.0% (3.0%) K-means (Hard) 96.9% (3.1%) Sparse auto-encoder 96.9% (3.1%) Sparse RBM 96.2% (3.8%) Table 2: Test recognition accuracy on CIFAR-10 Algorithm Test accuracy Raw pixels (reported in [11]) 37.3% RBM with backpropagation [11] 64.8% 3-Way Factored RBM + ZCA (3 layers) [23] 65.3% Mean-covariance RBM (3 layers) [22] 71.0% Improved Local Coordinate Coding [31] 74.5% Convolutional RBM [12] 78.9% K-means (Triangle) 77.9% K-means (Hard) 68.6% Sparse auto-encoder 73.4% Sparse RBM 72.4% K-means (Triangle, 4k features) 79.6%

Questions?