Package ELMR. November 28, 2015

Similar documents
Python Machine Learning

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Artificial Neural Networks written examination

Lecture 1: Machine Learning Basics

Generative models and adversarial training

Word Segmentation of Off-line Handwritten Documents

Test Effort Estimation Using Neural Network

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

A Neural Network GUI Tested on Text-To-Phoneme Mapping

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Learning From the Past with Experiment Databases

Evolutive Neural Net Fuzzy Filtering: Basic Description

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

WHEN THERE IS A mismatch between the acoustic

Human Emotion Recognition From Speech

A study of speaker adaptation for DNN-based speech synthesis

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Analysis of Hybrid Soft and Hard Computing Techniques for Forex Monitoring Systems

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

arxiv: v1 [cs.lg] 15 Jun 2015

Kamaldeep Kaur University School of Information Technology GGS Indraprastha University Delhi

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Model Ensemble for Click Prediction in Bing Search Ads

Time series prediction

CSL465/603 - Machine Learning

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Softprop: Softmax Neural Network Backpropagation Learning

CS Machine Learning

Rule Learning With Negation: Issues Regarding Effectiveness

Knowledge Transfer in Deep Convolutional Neural Nets

I-COMPETERE: Using Applied Intelligence in search of competency gaps in software project managers.

INPE São José dos Campos

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Assignment 1: Predicting Amazon Review Ratings

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Learning Methods for Fuzzy Systems

On the Formation of Phoneme Categories in DNN Acoustic Models

Classification Using ANN: A Review

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Issues in the Mining of Heart Failure Datasets

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Australian Journal of Basic and Applied Sciences

Modeling function word errors in DNN-HMM based LVCSR systems

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Data Fusion Through Statistical Matching

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Rule Learning with Negation: Issues Regarding Effectiveness

Speech Emotion Recognition Using Support Vector Machine

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Mining Association Rules in Student s Assessment Data

Modeling function word errors in DNN-HMM based LVCSR systems

arxiv: v2 [cs.ir] 22 Aug 2016

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Speaker Identification by Comparison of Smart Methods. Abstract

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Evolution of Symbolisation in Chimpanzees and Neural Nets

Device Independence and Extensibility in Gesture Recognition

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Second Exam: Natural Language Parsing with Neural Networks

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

(Sub)Gradient Descent

Digital Signal Processing: Speaker Recognition Final Report (Complete Version)

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Attributed Social Network Embedding

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Using focal point learning to improve human machine tacit coordination

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Learning Methods in Multilingual Speech Recognition

Deep Neural Network Language Models

Reducing Features to Improve Bug Prediction

Calibration of Confidence Measures in Speech Recognition

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Early Model of Student's Graduation Prediction Based on Neural Network

The Method of Immersion the Problem of Comparing Technical Objects in an Expert Shell in the Class of Artificial Intelligence Algorithms

A Deep Bag-of-Features Model for Music Auto-Tagging

Predicting Early Students with High Risk to Drop Out of University using a Neural Network-Based Approach

Axiom 2013 Team Description Paper

arxiv: v1 [cs.cv] 10 May 2017

FSL-BM: Fuzzy Supervised Learning with Binary Meta-Feature for Classification

arxiv: v2 [cs.cv] 30 Mar 2017

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Indian Institute of Technology, Kanpur

Discriminative Learning of Beam-Search Heuristics for Planning

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

An OO Framework for building Intelligence and Learning properties in Software Agents

Forget catastrophic forgetting: AI that learns after deployment

Chamilo 2.0: A Second Generation Open Source E-learning and Collaboration Platform

The University of Amsterdam s Concept Detection System at ImageCLEF 2011

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Transcription:

Title Extreme Machine Learning (ELM) Version 1.0 Author Alessio Petrozziello [aut, cre] Package ELMR November 28, 2015 Maintainer Alessio Petrozziello <alessio.petrozziello@port.ac.uk> Training and prediction functions are provided for the Extreme Learning Machine algorithm (ELM). The ELM use a Single Hidden Layer Feedforward Neural Network (SLFN) with random generated weights and no gradient-based backpropagation. The training time is very short and the online version allows to update the model using small chunk of the training set at each iteration. The only parameter to tune is the hidden layer size and the learning function. Depends R (>= 3.2.2) License GPL-2 GPL-3 LazyData true RoxygenNote 5.0.1 NeedsCompilation no Repository CRAN Date/Publication 2015-11-28 14:53:50 R topics documented: OSelm_train.formula.................................... 2 OSelm_training....................................... 2 predict_elm......................................... 3 preprocess.......................................... 4 Index 5 1

2 OSelm_training OSelm_train.formula Trains an extreme learning machine with random weights Trains an extreme learning machine with random weights OSelm_train.formula(formula, data, Elm_type, nhiddenneurons, ActivationFunction, N0, Block) formula data Elm_type a symbolic description of the model to be fitted. training data frame containing the variables specified in formula. select if the ELM must perform a "regression" or "classification" nhiddenneurons number of neurons in the hidden layer ActivationFunction "rbf" for radial basis function with Gaussian kernels, "sig" for sigmoidal fucntion, "sin" for sine function, "hardlim" for hard limit function N0 Block size of the first block to be processed size of each chunk to be processed at each step returns all the parameters used in the function, the weight matrix, the labels for the classification, the number of classes found, the bias, the beta activation function and the accuracy on the trainingset OSelm_training Trains an online sequential extreme learning machine with random weights Trains an online sequential extreme learning machine with random weights OSelm_training(p, y, Elm_Type, nhiddenneurons, ActivationFunction, N0, Block)

predict_elm 3 p y Elm_Type dataset used to perform the training of the model classes vector for classiication or regressors for regression select if the ELM must perform a "regression" or "classification" nhiddenneurons number of neurons in the hidden layer ActivationFunction "rbf" for radial basis function with Gaussian kernels, "sig" for sigmoidal fucntion, "sin" for sine function, "hardlim" for hard limit function N0 Block size of the first block to be processed size of each chunk to be processed at each step returns all the parameters used in the function, the weight matrix, the labels for the classification, the number of classes found, the bias, the beta activation function and the accuracy on the trainingset References [1] N.-Y. Liang, G.-B. Huang, P. Saratchandran, and N. Sundararajan, A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006 Examples x = runif(100, 0, 50) train = data.frame(y,x) train = data.frame(preprocess(train)) OSelm_train.formula(y~x, train, "regression", 100, "hardlim", 10, 10) predict_elm Prediction function for the ELM model generated with the elm_training() function Prediction function for the ELM model generated with the elm_training() function predict_elm(model, test) model test the output of the elm_training() function dataset used to perform the testing of the model, the first column must be the column to be fitted for the regression or the labels for the classification

4 preprocess returns the accuracy on the testset References [1] N.-Y. Liang, G.-B. Huang, P. Saratchandran, and N. Sundararajan, "A Fast and Accurate Online Sequential Learning Algorithm for Feedforward Networks" IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006 Examples x = runif(100, 0, 50) train = data.frame(y,x) train = data.frame(preprocess(train)) model = OSelm_train.formula(y~x, train, "regression", 100, "hardlim", 10, 10) # x = runif(100, 0, 50) test = data.frame(y,x) test = data.frame(preprocess(train)) accuracy = predict_elm(model, test) preprocess Pre processing function for the training and test data set. Each numeric variable is standardized between -1 and 1 and each categorical variable is coded with a dummy coding. Pre processing function for the training and test data set. Each numeric variable is standardized between -1 and 1 and each categorical variable is coded with a dummy coding. preprocess(data) data to be preprocesses return the pre processed dataset

Index OSelm_train.formula, 2 OSelm_training, 2 predict_elm, 3 preprocess, 4 5