Physical Activity Recognition from Accelerometer Data Using a Multi Scale Ensemble Method

Similar documents
OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Activity Recognition from Accelerometer Data

Human Emotion Recognition From Speech

Lecture 1: Machine Learning Basics

Speech Emotion Recognition Using Support Vector Machine

Assignment 1: Predicting Amazon Review Ratings

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Python Machine Learning

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Learning From the Past with Experiment Databases

Understanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010)

Speech Recognition by Indexing and Sequencing

Rule Learning With Negation: Issues Regarding Effectiveness

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Word Segmentation of Off-line Handwritten Documents

Automatic Pronunciation Checker

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

A study of speaker adaptation for DNN-based speech synthesis

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Multivariate k-nearest Neighbor Regression for Time Series data -

Learning Methods in Multilingual Speech Recognition

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Rule Learning with Negation: Issues Regarding Effectiveness

Why Did My Detector Do That?!

Probability and Statistics Curriculum Pacing Guide

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Australian Journal of Basic and Applied Sciences

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Modeling function word errors in DNN-HMM based LVCSR systems

Large-Scale Web Page Classification. Sathi T Marath. Submitted in partial fulfilment of the requirements. for the degree of Doctor of Philosophy

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Chapters 1-5 Cumulative Assessment AP Statistics November 2008 Gillespie, Block 4

Modeling function word errors in DNN-HMM based LVCSR systems

This scope and sequence assumes 160 days for instruction, divided among 15 units.

A Vector Space Approach for Aspect-Based Sentiment Analysis

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Statewide Framework Document for:

The stages of event extraction

arxiv: v1 [cs.lg] 3 May 2013

Data Fusion Through Statistical Matching

Broward County Public Schools G rade 6 FSA Warm-Ups

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Reducing Features to Improve Bug Prediction

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method

CS 1103 Computer Science I Honors. Fall Instructor Muller. Syllabus

Speech Recognition at ICSI: Broadcast News and beyond

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

(Sub)Gradient Descent

Montana Content Standards for Mathematics Grade 3. Montana Content Standards for Mathematical Practices and Mathematics Content Adopted November 2011

CS Machine Learning

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Generative models and adversarial training

learning collegiate assessment]

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Artificial Neural Networks written examination

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Ensemble Technique Utilization for Indonesian Dependency Parser

Corpus Linguistics (L615)

Universidade do Minho Escola de Engenharia

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Visit us at:

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

QUESTIONS and Answers from Chad Rice?

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Indian Institute of Technology, Kanpur

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Ph.D in Advance Machine Learning (computer science) PhD submitted, degree to be awarded on convocation, sept B.Tech in Computer science and

Cooper Upper Elementary School

STT 231 Test 1. Fill in the Letter of Your Choice to Each Question in the Scantron. Each question is worth 2 point.

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Calibration of Confidence Measures in Speech Recognition

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

Linking Task: Identifying authors and book titles in verbose queries

Issues in the Mining of Heart Failure Datasets

Switchboard Language Model Improvement with Conversational Data from Gigaword

Detection and Classification of Mu Rhythm using Phase Synchronization for a Brain Computer Interface

Model Ensemble for Click Prediction in Bing Search Ads

Combining Proactive and Reactive Predictions for Data Streams

Experiment Databases: Towards an Improved Experimental Methodology in Machine Learning

Historical maintenance relevant information roadmap for a self-learning maintenance prediction procedural approach

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

WHEN THERE IS A mismatch between the acoustic

Speaker recognition using universal background model on YOHO database

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

*Lesson will begin on Friday; Stations will begin on the following Wednesday*

Evolutive Neural Net Fuzzy Filtering: Basic Description

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Transcription:

Physical Activity Recognition from Accelerometer Data Using a Multi Scale Ensemble Method Yonglei Zheng, Weng Keen Wong, Xinze Guan (Oregon State University) Stewart Trost (University of Queensland)

Introduction Goal: accurate, objective and detailed measurement of physical activity Why? Many health related reasons Understand relationship between physical activity and health outcomes Detecting at risk populations Measure effectiveness of intervention strategies

Introduction Accelerometers are a cheap, reliable and unobtrusive way to measure physical activity Capture acceleration in different planes (typically triaxial) Typically attached at the wrist or hip Actigraph s GT3X+ accelerometer Dimensions: 4.6cm x 3.3cm x 1.9cm Weight: 19 g

Introduction The challenge: interpreting this data Lying Down / Sitting Standing Walking

Introduction Segment and classify freeliving data Amplitude LiME Data Sample 2 1.5 1 0.5 0 0 100 200 300 400 500 0.5 Followup paper (not this talk) 1 1.5 Classify already segmented data 2 Time (Seconds) This talk Walking Running

Related Work 1. Time series Classification (see Xing, Pei and Keogh 2010) Nearest neighbor approaches with different distances metrics eg. Euclidean (Keogh and Kasetty 2003), Dynamic time warping (Wang et al. 2010) Supervised Learning eg. decision trees (Bonomi et al. 2009), neural networks (Staudenmayer et al. 2009), support vector regression (Su et al. 2005), ensembles (Ravi et al. 2005) Many different representations used eg. symbolic (Lin et al. 2003), shapelets (Ye and Keogh 2009), etc. 2. Segmentation Hidden Markov Models (Lester et al. 2005, Pober et al. 2006) Conditional Random Fields (van Kasteren et al. 2008, Gu et al. 2009, Wu et al. 2009)

Introduction Things to note: Each window of data consists of a single activity Repetitive pattern Discriminative features at different scales Supervised learning approach works very well on our data

Methodology Supervised Learning Approach Cut time series into non overlapping windows Time Axis 1 Axis 2 Axis 3 16:34:00 191 14 72 16:34:01 36 18 63 16:34:02 6 19 22 16:34:03 21 60 79 Feature Value X1 0.1 X2 15 X3 2 Supervised learning approaches

Methodology Two issues when applying supervised learning to time series data 1. What features to use? Feature extraction ultimately needs to be efficient Bag of features + regularization works very well

10 Features Axis 1 1. Percentiles: 10 th,25 th,50 th,75 th,9 0 th 2. Lag oneautocorrelation 3. Sum 4. Mean 5. Standard deviation 6. Coefficients of variation 7. Peak to peak amplitude 8. Interquartile range 9. Skewness 10. Kurtosis 11. Signal power 12. Log energy 13. Peak intensity 14. Zero crossings Axis 2 1. Percentiles: 10 th,25 th,50 th,75 th,9 0 th 2. Lag oneautocorrelation 3. Sum 4. Mean 5. Standard deviation 6. Coefficients of variation 7. Peak to peak amplitude 8. Interquartile range 9. Skewness 10. Kurtosis 11. Signal power 12. Log energy 13. Peak intensity 14. Zero crossings Axis 3 1. Percentiles: 10 th,25 th,50 th,75 th,9 0 th 2. Lag oneautocorrelation 3. Sum 4. Mean 5. Standard deviation 6. Coefficients of variation 7. Peak to peak amplitude 8. Interquartile range 9. Skewness 10. Kurtosis 11. Signal power 12. Log energy 13. Peak intensity 14. Zero crossings Between two axes 1. Correlation between axis 1 and axis2 2. Correlation between axis 2 and axis3 3. Correlation between axis 1 and axis3

Methodology Two issues when applying supervised learning to time series data 1. What features to use? 2. How big of a window? Too big: features too coarse, high latency of activity recognition Too small: features meaningless Need multi scale approach

Subwindow Ensemble Model Training data from other time series {t 1, t 2,, t 10 } 10 subwindows Training data from other time series {t 1, t 2,, t 6 } 6 subwindows Training data from other time series {t 1 } 1 subwindow Single scale model (1 sec) Single scale model (5 sec) Single scale model (10 sec) Majority Vote Final Prediction 12

Experiments Datasets Human Activity Sensing Challenge (triaxial, 100 Hz, 7 subjects, 6 classes) OSU Hip (triaxial, 30Hz, 53 subjects, 7 classes) OSU Wrist (triaxial, 30 Hz, 18 subjects, 7 classes) Experimental Setup Split by subject into train/validate/test splits Averaged over 30 splits

Experiments Algorithms 1. 1 NN (Euclidean distance, DTW) 2. (Single scale) Supervised Learning Algorithms (ANN, SVM) with 10 second windows 3. (Multi scale) SWEM (SVM) with 10 ensemble members

Results Algorithm HASC (Macro F1) OSU Hip (Macro F1) OSU Wrist (Macro F1) SWEM (SVM) 0.820* 0.942* 0.896* SVM (W=10) 0.794 0.937 0.886 ANN (W=10) 0.738 0.919 0.787 1NN (EUC) 0.648 0.572 0.456 1NN (DTW) 0.648 0.561 0.494

Results We can also analyze the performance of each ensemble member by itself:

Conclusion Subwindow Ensemble Model able to capture discriminative features at different scales without committing to a single window size Outperforms baseline algorithms High F1 indicates it is viable for deployment Future work: free living data segmentation, online algorithms

Acknowledgements This work was supported in part by funding from the Eunice Kennedy Shriver National Institute of Child Health and Human Development (NICHD R01 55400A)

Questions?

OSU Hip

HASC 21