Big Data Meets Advanced Analytics Concepts and Practical Examples Dr. Gerhard Svolba SAS DACH Competence Center Analytics

Similar documents
Python Machine Learning

arxiv: v2 [cs.cv] 30 Mar 2017

Generative models and adversarial training

Lecture 1: Machine Learning Basics

Lecture 1: Basic Concepts of Machine Learning

Artificial Neural Networks written examination

Assignment 1: Predicting Amazon Review Ratings

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

(Sub)Gradient Descent

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

arxiv: v1 [cs.lg] 15 Jun 2015

CS 446: Machine Learning

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

CSL465/603 - Machine Learning

Australian Journal of Basic and Applied Sciences

Word Segmentation of Off-line Handwritten Documents

Learning Methods for Fuzzy Systems

Probabilistic Latent Semantic Analysis

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Calibration of Confidence Measures in Speech Recognition

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Time series prediction

Model Ensemble for Click Prediction in Bing Search Ads

A study of speaker adaptation for DNN-based speech synthesis

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Speech Emotion Recognition Using Support Vector Machine

Learning From the Past with Experiment Databases

INPE São José dos Campos

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

CS Machine Learning

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Issues in the Mining of Heart Failure Datasets

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Speaker Identification by Comparison of Smart Methods. Abstract

Reducing Features to Improve Bug Prediction

Indian Institute of Technology, Kanpur

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Active Learning. Yingyu Liang Computer Sciences 760 Fall

A deep architecture for non-projective dependency parsing

Multi-tasks Deep Learning Model for classifying MRI images of AD/MCI Patients

Human Emotion Recognition From Speech

Comparison of network inference packages and methods for multiple networks inference

DNN ACOUSTIC MODELING WITH MODULAR MULTI-LINGUAL FEATURE EXTRACTION NETWORKS

Modeling function word errors in DNN-HMM based LVCSR systems

Softprop: Softmax Neural Network Backpropagation Learning

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

An OO Framework for building Intelligence and Learning properties in Software Agents

Second Exam: Natural Language Parsing with Neural Networks

Modeling function word errors in DNN-HMM based LVCSR systems

Learning to Schedule Straight-Line Code

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

THE enormous growth of unstructured data, including

Data Fusion Through Statistical Matching

School of Innovative Technologies and Engineering

WHEN THERE IS A mismatch between the acoustic

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Knowledge Transfer in Deep Convolutional Neural Nets

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

Rule Learning With Negation: Issues Regarding Effectiveness

A Review: Speech Recognition with Deep Learning Methods

Deep Neural Network Language Models

A survey of multi-view machine learning

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Linking Task: Identifying authors and book titles in verbose queries

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Deep Bag-of-Features Model for Music Auto-Tagging

Axiom 2013 Team Description Paper

Using dialogue context to improve parsing performance in dialogue systems

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

Learning Methods in Multilingual Speech Recognition

Shockwheat. Statistics 1, Activity 1

A Case Study: News Classification Based on Term Frequency

Applications of data mining algorithms to analysis of medical data

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Mining Association Rules in Student s Assessment Data

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Welcome to. ECML/PKDD 2004 Community meeting

Switchboard Language Model Improvement with Conversational Data from Gigaword

Artificial Neural Networks

Using EEG to Improve Massive Open Online Courses Feedback Interaction

Universidade do Minho Escola de Engenharia

arxiv: v1 [cs.cv] 10 May 2017

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Transcription:

Big Data Meets Advanced Analytics Concepts and Practical Examples Dr. Gerhard Svolba SAS DACH Competence Center Analytics #analyticsx

Concepts when Handling Big Data Use Advanced Machine Learning methods to describe the relationships in your data Understand specifics of complex systems by using Monte Carlo simulations Run SAS High Performance Analytics procedures in distributed mode

Two Practical Big Data Examples Deep Learning with Stacked Denoising Autoencoders Recognize Handwritten Digits Compress information into a few variables Monte Carlo Simulation of the Monopoly Board Game Distribution of the visit frequency on the fields Studying the profitability of different investment decisions

Machine learning ML wird häufig eingesetzt, wenn Die Vorhersagegüte eines Modells wichtiger ist als die Erklärbarkeit Traditionelle Ansätze ungeeignet erscheinen» Mehr Variablen als Beobachtungen» Viele hochkorrelierte Variablen» Unstrukturierte Daten» Fundamental nicht-lineare Zusammenhänge Anwendung : Mustererkennung

Data Mining Machine Learning #analyticsx SUPERVISED LEARNING UNSUPERVISED LEARNING SEMI-SUPERVISED LEARNING TRANSDUCTION Regression LASSO regression Logistic regression Ridge regression Decision tree Gradient boosting Random forests Neural networks SVM Naïve Bayes Neighbors Gaussian processes A priori rules Clustering k-means clustering Mean shift clustering Spectral clustering Kernel density estimation Nonnegative matrix factorization PCA Kernel PCA Sparse PCA Singular value decomposition SOM Prediction and classification* Clustering* EM TSVM Manifold regularization Autoencoders Multilayer perceptron Restricted Boltzmann machines REINFORCEMENT LEARNING DEVELOPMENTAL LEARNING *In semi-supervised learning, supervised prediction and classification algorithms are often combined with clustering.

Handwritten Digits as Training Data Classic MNIST training data 784 features form a 28x28 digital grid Greyscale features range from 0 to 255 60,000 labeled training images (785 variables, including 1 nominal target) 10,000 unlabeled test images (784 input variables)

Semi-Supervised Learning Extract a few representative features to discriminate the digits 0-9 Compress information of 784 variables into 2 features Use a convolutional neural network (deep learning)

Deep-Learning using a Stacked De-noising Autoencoder h5 h4 Uncorrupted Output Features Hidden Neurons Hidden Neurons Target Layer h3 Hidden Neurons Extractable Features h2 Hidden Neurons h1 Hidden Neurons Partially Corrupted Input Features Input Layer

Using SAS Code to Solve the Problem Noised INPUT (i) H1 H2 H3 H4 H5 INPUT(t) i=>h1 H1=>H2 H2=>H3 H3=>H4 H4=>H5 proc neural data= autoencodertraining dmdbcat= work.autoencodertrainingcat performance compile details cpucount= 12 threads= yes; archi MLP hidden= 5; hidden 300 / id= h1; hidden 100 / id= h2; hidden 2 / id= h3 act= linear; hidden 100 / id= h4; hidden 300 / id= h5; input corruptedpixel1-corruptedpixel400 / id= i level= int std= std; target pixel1-pixel400 / act= identity id= t level= int std= std; initial random= 123; prelim 10 preiter= 10; freeze h1->h2; freeze h2->h3; freeze h3->h4; freeze h4->h5; train technique= congra maxtime= 129600 maxiter= 1000; freeze i->h1; thaw h1->h2; train technique= congra maxtime= 129600 maxiter= 1000; freeze h1->h2; thaw h2->h3; train technique= congra maxtime= 129600 maxiter= 1000; freeze h2->h3; thaw h3->h4; train technique= congra maxtime= 129600 maxiter= 1000; freeze h3->h4; thaw h4->h5; train technique= congra maxtime= 129600 maxiter= 1000; thaw i->h1; thaw h1->h2; thaw h2->h3; thaw h3->h4; train technique= congra maxtime= 129600 maxiter= 1000; code file= 'C:\Path\to\code.sas'; run;

Studying a certain section in detail Target 1 Target 2 Target 3 Target 4 W51 W52 W53 W54 h51 h52

Edge Weights of the 5 th layer are loaded with discriminative information

Visualization of the separation of the two middle hidden layers

Our method results in much better separation that simple principal components analysis

Summary: Semi-Supervised Learning Extremely accurate predictions using deep neural networks. Target Variable Digit 0-9 has not been used in the model! Feature Extraction as pre-step in predictive modeling Requires Model-Tuning The most common applications of deep learning involve pattern recognition in unstructured data, such as text, photos, videos and sound.

The Monopoly board game is a complex system Set of Complex Rules Monetary Dimension Dynamic Component Additional Instructions Framework of Opportunities and Events Random Components

Questions of Interest What is the distribution of visits on the fields of the board game? Which fields are most profitable? Which fields to have a high variability in profitability? These questions can be transferred to many other simulations studies of complex systems.

Locating the Token Influential Factors Sum of 2 Dice Go to Jail! Event Fields Accelerator Dice

Almost Even Distribution Sum of 2 Dice

All Field-40 visits are relocated to 14 Sum of 2 Dice Go to Jail!

Event Fields relocate to other fields Sum of 2 Dice Go to Jail! Event Fields

Red Dice introduces high variability Sum of 2 Dice Go to Jail! Event Fields Accelerator Dice

Effect of the accelerator dice after 20 rounds If the 3rd dice shows the Monopoly man: Move forward to the next free propertyfield The the next property field otherwiese

Effect of the accelerator dice after 70 rounds

Example for a Relocation

Profitability Distribution after 40 rounds

Profitability Distribution after 70 rounds

Implementation in SAS

Summary Applying advanced analytical methods to big data allows you to better understand relationshiops in the underlying processes. You receive results that would otherweise remain undiscovered. SAS offers a full set of methods to handle big data in advanced analytics applications

Links Patric Hall: Overview of Machine Learning with SAS Enterprise Miner http://support.sas.com/resources/papers/proceedings 14/SAS313-2014.pdf Rick Wicklin: Simulating Data with SAS http://support.sas.com/publishing/authors/wicklin.ht ml Gerhard Svolba: Applying Data Science: Business Case Studies Using SAS (SAS Press, expected 2017)

Contact Information Gerhard Svolba Analytic Solution Architect SAS-Austria Sastools.by.gerhard@gmx.net http://www.sascommunity.org/wiki/gerhard_svolba LinkedIn XING PictureBlog Data Quality for Analytics Using SAS SAS Press 2012 http://www.sascommunity.org/wiki/data_quality_for_analytics Data Preparation for Analytics Using SAS SAS Press 2006 http://www.sascommunity.org/wiki/data_preparation_for_analytics