BCI Competition 2008 Graz data set B

Similar documents
Detection and Classification of Mu Rhythm using Phase Synchronization for a Brain Computer Interface

East Tennessee State University, Johnson City b Algoma University, Sault Ste. Marie, Ontario, Canada. Online publication date: 28 December 2010

LEGO MINDSTORMS Education EV3 Coding Activities

TeacherPlus Gradebook HTML5 Guide LEARN OUR SOFTWARE STEP BY STEP

Python Machine Learning

Robot manipulations and development of spatial imagery

WHEN THERE IS A mismatch between the acoustic

Assignment 1: Predicting Amazon Review Ratings

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

COMPUTER INTERFACES FOR TEACHING THE NINTENDO GENERATION

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Speech Emotion Recognition Using Support Vector Machine

Using EEG to Improve Massive Open Online Courses Feedback Interaction

Modeling function word errors in DNN-HMM based LVCSR systems

Curriculum Design Project with Virtual Manipulatives. Gwenanne Salkind. George Mason University EDCI 856. Dr. Patricia Moyer-Packenham

Human Emotion Recognition From Speech

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

SOFTWARE EVALUATION TOOL

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Millersville University Degree Works Training User Guide

Introduction to Causal Inference. Problem Set 1. Required Problems

Does the Difficulty of an Interruption Affect our Ability to Resume?

Learning Methods for Fuzzy Systems

DegreeWorks Advisor Reference Guide

ecampus Basics Overview

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

A study of speaker adaptation for DNN-based speech synthesis

Using SAM Central With iread

PowerTeacher Gradebook User Guide PowerSchool Student Information System

On the Formation of Phoneme Categories in DNN Acoustic Models

Speaker Identification by Comparison of Smart Methods. Abstract

Modeling function word errors in DNN-HMM based LVCSR systems

Diploma in Library and Information Science (Part-Time) - SH220

Creating Your Term Schedule

Appendix L: Online Testing Highlights and Script

CHANCERY SMS 5.0 STUDENT SCHEDULING

Many instructors use a weighted total to calculate their grades. This lesson explains how to set up a weighted total using categories.

Non-Secure Information Only

Lecture 1: Machine Learning Basics

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Test Administrator User Guide

Driving Author Engagement through IEEE Collabratec

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

STUDENT MOODLE ORIENTATION

Proceedings of Meetings on Acoustics

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

LESSON PLANS: AUSTRALIA Year 6: Patterns and Algebra Patterns 50 MINS 10 MINS. Introduction to Lesson. powered by

Field Experience Management 2011 Training Guides

1. REFLEXES: Ask questions about coughing, swallowing, of water as fast as possible (note! Not suitable for all

INTERMEDIATE ALGEBRA PRODUCT GUIDE

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Eye Movements in Speech Technologies: an overview of current research

Grade 2: Using a Number Line to Order and Compare Numbers Place Value Horizontal Content Strand

CS Machine Learning

The Perception of Nasalized Vowels in American English: An Investigation of On-line Use of Vowel Nasalization in Lexical Access

Getting Started with TI-Nspire High School Science

Word Segmentation of Off-line Handwritten Documents

Running head: DELAY AND PROSPECTIVE MEMORY 1

Tracking decision makers under uncertainty

SARDNET: A Self-Organizing Feature Map for Sequences

GRADUATE STUDENT HANDBOOK Master of Science Programs in Biostatistics

(De-)Accentuation and the Processing of Information Status: Evidence from Event- Related Brain Potentials

A Case-Based Approach To Imitation Learning in Robotic Agents

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

Mandarin Lexical Tone Recognition: The Gating Paradigm

4-3 Basic Skills and Concepts

School Year 2017/18. DDS MySped Application SPECIAL EDUCATION. Training Guide

School of Innovative Technologies and Engineering

Enduring Understandings: Students will understand that

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Outreach Connect User Manual

CS 101 Computer Science I Fall Instructor Muller. Syllabus

Eliciting Language in the Classroom. Presented by: Dionne Ramey, SBCUSD SLP Amanda Drake, SBCUSD Special Ed. Program Specialist

A student diagnosing and evaluation system for laboratory-based academic exercises

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Calibration of Confidence Measures in Speech Recognition

A Case Study: News Classification Based on Term Frequency

Managing the Student View of the Grade Center

Speech/Language Pathology Plan of Treatment

Multisensor Data Fusion: From Algorithms And Architectural Design To Applications (Devices, Circuits, And Systems)

Linking object names and object categories: Words (but not tones) facilitate object categorization in 6- and 12-month-olds

A Neural Network GUI Tested on Text-To-Phoneme Mapping

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

IVY TECH COMMUNITY COLLEGE

AC : FACILITATING VERTICALLY INTEGRATED DESIGN TEAMS

Office of Planning and Budgets. Provost Market for Fiscal Year Resource Guide

Reducing Features to Improve Bug Prediction

Rule Learning With Negation: Issues Regarding Effectiveness

Lisa Forster Student Functional Group - ITS. SI-net: Student Placements

Evaluation of Various Methods to Calculate the EGG Contact Quotient

Using Blackboard.com Software to Reach Beyond the Classroom: Intermediate

PDF hosted at the Radboud Repository of the Radboud University Nijmegen

Preferences...3 Basic Calculator...5 Math/Graphing Tools...5 Help...6 Run System Check...6 Sign Out...8

Early Model of Student's Graduation Prediction Based on Neural Network

GRAPHIC DESIGN TECHNOLOGY Associate in Applied Science: 91 Credit Hours

Your School and You. Guide for Administrators

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2

Transcription:

BCI Competition 2008 Graz data set B R. Leeb 1, C. Brunner 1, G. R. Müller-Putz 1, A. Schlögl 2, and G. Pfurtscheller 1 1 Institute for Knowledge Discovery, Graz University of Technology, Austria 2 Institute for Human-Computer Interfaces, Graz University of Technology, Austria Experimental paradigm This data set consists of EEG data from 9 subjects of a study published in [1]. The subjects were right-handed, had normal or corrected-to-normal vision and were paid for participating in the experiments. All volunteers were sitting in an armchair, watching a flat screen monitor placed approximately 1 m away at eye level. For each subject 5 sessions are provided, whereby the first two sessions contain training data without feedback (screening), and the last three sessions were recorded with feedback. Each session consists of several runs, illustrated in Figure 1. At the beginning of each session, a recording of approximately 5 minutes was performed to estimate the EOG influence. The recording was divided into 3 blocks: (1) two minutes with eyes open (looking at a fixation cross on the screen), (2) one minute with eyes closed, and (3) one minute with eye movements. The artifact block was divided into four sections (15 seconds artifacts with 5 seconds resting in between) and the subjects were instructed with a text on the monitor to perform either eye blinking, rolling, up-down or left-right movements. At the beginning and at the end of each task a low and high warning tone were presented, respectively. Note that due to technical problems no EOG block is available in session B0102T and B0504E, (see Table 1 for a list of all subjects) Figure 1: Timing scheme of one session (for screening and feedback sessions). 1

Figure 2: Electrode montage of the three monopolar EOG channels. Data recording Three bipolar recordings (C3, Cz, and C4) were recorded with a sampling frequency of 250 Hz. The recordings had a dynamic range of ±100 µv for the screening and ±50 µv for the feedback sessions. They were bandpassfiltered between 0.5 Hz and 100 Hz, and a notch filter at 50 Hz was enabled. The placement of the three bipolar recordings (large or small distances, more anterior or posterior) were slightly different for each subject (for more details see [1]). The electrode position Fz served as EEG ground. In addition to the EEG channels, the electrooculogram (EOG) was recorded with three monopolar electrodes (see Figure 2, left mastoid serving as reference) using the same amplifier settings, but with a dynamic range of ±1 mv. The EOG channels are provided for the subsequent application of artifact processing methods [2] and must not be used for classification. The cue-based screening paradigm (see Figure 3a) consisted of two classes, namely the motor imagery (MI) of left hand (class 1) and right hand (class 2). Each subject participated in two screening sessions without feedback recorded on two different days within two weeks. Each session consisted of six runs with ten trials each and two classes of imagery. This resulted in 20 trials per run and 120 trials per session. Data of 120 repetitions of each MI class were available for each person in total. Prior to the first motor imagery training the subject executed and imagined different movements for each body part and selected the one which they could imagine best (e. g., squeezing a ball or pulling a brake). Each trial started with a fixation cross and an additional short acoustic warning tone (1 khz, 70 ms). Some seconds later a visual cue (an arrow pointing either to the left or right, according to the requested class) was presented for 1.25 seconds. Afterwards the subjects had to imagine the corresponding hand movement over a period of 4 seconds. Each trial was followed by a short break of at least 1.5 seconds. A randomized time of up to 1 second was added to the break to avoid adaptation. For the three online feedback sessions four runs with smiley feedback were recorded (see Figure 3b), whereby each run consisted of twenty trials for each type of motor imagery. At the beginning of each trial (second 0) the 2

(a) Screening Imagination of left hand movement Imagination of right hand movement beep Fixation cross Cue Imagery Period Pause 0 1 2 3 4 5 6 7 8 9 time in s (b) Smiley Feedback beep Smiley (grey) Cue Feedback Period (Smiley) Pause 0 1 2 3 4 5 6 7 8 9 time in s Figure 3: Timing scheme of the paradigm. (a) The first two sessions (01T, 02T) contain training data without feedback, and (b) the last three sessions (03T, 04E, 05E) with smiley feedback. feedback (a gray smiley) was centered on the screen. At second 2, a short warning beep (1 khz, 70 ms) was given. The cue was presented from second 3 to 7.5. Depending on the cue, the subjects were required to move the smiley towards the left or right side by imagining left or right hand movements, respectively. During the feedback period the smiley changed to green when moved in the correct direction, otherwise it became red. The distance of the smiley from the origin was set according to the integrated classification output over the past two seconds (more details see [1]). Furthermore, the classifier output was also mapped to the curvature of the mouth causing the smiley to be happy (corners of the mouth upwards) or sad (corners of the mouth downwards). At second 7.5 the screen went blank and a random interval between 1.0 and 2.0 seconds was added to the trial. The subject was instructed to keep the smiley on the correct side for as long as possible and therefore to perform the MI as long as possible. Data file description All data sets are stored in the General Data Format for biomedical signals (GDF), one file per subject and session. However, only the first three sessions contain the class labels for all trials, whereas the remaining two sessions are used to test the classifier and hence to evaluate the performance. All files are listed in Table 1. The GDF files can be loaded using the open-source toolbox BioSig, available for free at http://biosig.sourceforge.net/. There are 3

ID Training Evaluation 1 B0101T, B0102T, B0103T B0104E, B0105E 2 B0201T, B0202T, B0203T B0204E, B0205E 3 B0301T, B0302T, B0303T B0304E, B0305E 4 B0401T, B0402T, B0403T B0404E, B0405E 5 B0501T, B0502T, B0503T B0504E, B0505E 6 B0601T, B0602T, B0603T B0604E, B0605E 7 B0701T, B0702T, B0703T B0704E, B0705E 8 B0801T, B0802T, B0803T B0804E, B0805E 9 B0901T, B0902T, B0903T B0904E, B0905E Table 1: List of all files contained in the data set, the striked out evaluation data sets will be provided after the deadline of the competition. The first two sessions (...01T,...02T) contain training data without feedback, and the last three sessions (...03T,...04E,...05E) with smiley feedback. Note: Due to technical problems no recording for EOG estimation (eyes open, closed, eye movements) exists in session B0102T and B0504E. versions for Octave 1 /MATLAB 2 as well as a library for C/C++. A GDF file can be loaded with the BioSig toolbox with the following command in Octave/MATLAB (for C/C++, the corresponding function HDRTYPE* sopen and size t sread must be called): [s, h] = sload( B0101T.gdf ); Note that the runs are separated by 100 missing values, which are encoded as not-a-numbers (NaN) by default. Alternatively, this behavior can be turned off and the missing values will be encoded as the negative maximum values as stored in the file with: [s, h] = sload( A01T.gdf, 0, OVERFLOWDETECTION:OFF ); The workspace will then contain two variables, namely the signals s and a header structure h. The signal variable contains 6 channels (the first 3 are EEG and the last 3 are EOG signals). The header structure contains event information that describes the structure of the data over time. The following fields provide important information for the evaluation of this data set: h.event.typ h.event.pos h.event.dur 1 http://www.gnu.org/software/octave/ 2 The MathWorks, Inc., Natick, MA, USA 4

Event type Description 276 0x0114 Idling EEG (eyes open) 277 0x0115 Idling EEG (eyes closed) 768 0x0300 Start of a trial 769 0x0301 Cue onset left (class 1) 770 0x0302 Cue onset right (class 2) 781 0x030D BCI feedback (continuous) 783 0x030F Cue unknown 1023 0x03FF Rejected trial 1077 0x0435 Horizontal eye movement 1078 0x0436 Vertical eye movement 1079 0x0437 Eye rotation 1081 0x0439 Eye blinks 32766 0x7FFE Start of a new run Table 2: List of event types (the first column contains decimal values and the second hexadecimal values). The position of an event in samples is contained in h.event.pos. The corresponding type can be found in h.event.typ, and the duration of that particular event is stored in h.event.dur. The types used in this data set are described in Table 2 (hexadecimal values, decimal notation in parentheses). Note that the class labels (i. e., 1 and 2, corresponding to event types 769 and 770) are only provided for the training data and not for the testing data. The trials containing artifacts as scored by experts are marked as events with the type 1023. In addition, h.artifactselection contains a list of all trials, with 0 corresponding to a clean trial and 1 corresponding to a trial containing an artifact. In order to view the GDF files, the viewing and scoring application SigViewer v0.2 or higher (part of BioSig) can be used. Evaluation Participants should provide a continuous classification output for each sample in the form of classlabels (1, 2), including labeled trials and trials marked as artifact. A confusion matrix will then be built from all artifact-free trials for each time point. From these confusion matrices, the time course of the accuracy as well as the kappa coefficient will be obtained [3]. The algorithm used for this evaluation will be provided in BioSig. The winner is the algorithm with the largest kappa value X.KAP00. Due to the fact that the evaluation data sets will not be distributed until the end of the competition, the submissions must be programs that accept 5

EEG data (the structure of this data must be the same as used in all training sets 3 ) as input and produce the aforementioned class label vector. Since three EOG channels are provided, it is required to remove EOG artifacts before the subsequent data processing using artifact removal techniques such as highpass filtering or linear regression [4]. In order to enable the application of other correction methods, we have opted for a maximum transparency approach and provided the EOG channels; at the same time we request that artifacts do not influence the classification result. All algorithms must be causal, meaning that the classification output at time k may only depend on the current and past samples x k, x k 1,..., x 0. In order to check whether the causality criterion and the artifact processing requirements are fulfilled, all submissions must be open source, including all additional libraries, compilers, programming languages, and so on (for example, Octave/FreeMat, C++, Python,... ). Note that submissions can also be written in the closed-source development environment MATLAB as long as the code is executable in Octave. Similarily, C++ programs can be written and compiled with a Microsoft or Intel compiler, but the code must also compile with g++. References [1] R. Leeb, F. Lee, C. Keinrath, R. Scherer, H. Bischof, G. Pfurtscheller. Brain-computer communication: motivation, aim, and impact of exploring a virtual apartment. IEEE Transactions on Neural Systems and Rehabilitation Engineering 15, 473 482, 2007. [2] M. Fatourechi, A. Bashashati, R. K. Ward, G. E. Birch. EMG and EOG artifacts in brain computer interface systems: a survey. Clinical Neurophysiology 118, 480 494, 2007. [3] A. Schlögl, J. Kronegg, J. E. Huggins, S. G. Mason. Evaluation criteria in BCI research. In: G. Dornhege, J. d. R. Millán, T. Hinterberger, D. J. McFarland, K.-R. Müller (Eds.). Toward brain-computer interfacing, MIT Press, 327 342, 2007. [4] A. Schlögl, C. Keinrath, D. Zimmermann, R. Scherer, R. Leeb, G. Pfurtscheller. A fully automated correction method of EOG artifacts in EEG recordings. Clin.Neurophys. 2007 Jan;118(1):98-104. 3 One evaluation data set is distributed from the beginning of the competition to enable participants to test their program and to ensure that it produces the desired output. 6