SPEECH EMOTION RECOGNITION WITH SKEW-ROBUST NEURAL NETWORKS

Similar documents
Speech Emotion Recognition Using Support Vector Machine

A study of speaker adaptation for DNN-based speech synthesis

Human Emotion Recognition From Speech

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Python Machine Learning

Lecture 1: Machine Learning Basics

WHEN THERE IS A mismatch between the acoustic

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Assignment 1: Predicting Amazon Review Ratings

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Affective Classification of Generic Audio Clips using Regression Models

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Learning Methods in Multilingual Speech Recognition

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Evolutive Neural Net Fuzzy Filtering: Basic Description

Rule Learning With Negation: Issues Regarding Effectiveness

Artificial Neural Networks written examination

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Reducing Features to Improve Bug Prediction

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

(Sub)Gradient Descent

Rule Learning with Negation: Issues Regarding Effectiveness

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

Australian Journal of Basic and Applied Sciences

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

INPE São José dos Campos

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

CS Machine Learning

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Calibration of Confidence Measures in Speech Recognition

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

The Good Judgment Project: A large scale test of different methods of combining expert predictions

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Speaker Identification by Comparison of Smart Methods. Abstract

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Segmental Conditional Random Fields with Deep Neural Networks as Acoustic Models for First-Pass Word Recognition

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Word Segmentation of Off-line Handwritten Documents

Generative models and adversarial training

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language

Softprop: Softmax Neural Network Backpropagation Learning

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Probabilistic Latent Semantic Analysis

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition

Multi-Lingual Text Leveling

Learning From the Past with Experiment Databases

The NICT/ATR speech synthesis system for the Blizzard Challenge 2008

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

Knowledge Transfer in Deep Convolutional Neural Nets

Multivariate k-nearest Neighbor Regression for Time Series data -

International Journal of Advanced Networking Applications (IJANA) ISSN No. :

Probability and Statistics Curriculum Pacing Guide

Large-Scale Web Page Classification. Sathi T Marath. Submitted in partial fulfilment of the requirements. for the degree of Doctor of Philosophy

Using dialogue context to improve parsing performance in dialogue systems

Speaker recognition using universal background model on YOHO database

Whodunnit Searching for the Most Important Feature Types Signalling Emotion-Related User States in Speech

Issues in the Mining of Heart Failure Datasets

Indian Institute of Technology, Kanpur

Automatic Pronunciation Checker

Time series prediction

Speech Recognition at ICSI: Broadcast News and beyond

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

arxiv: v1 [cs.lg] 3 May 2013

The stages of event extraction

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

Model Ensemble for Click Prediction in Bing Search Ads

Semi-Supervised Face Detection

Automatic Speaker Recognition: Modelling, Feature Extraction and Effects of Clinical Environment

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

Speech Recognition by Indexing and Sequencing

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

A Case Study: News Classification Based on Term Frequency

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

TD(λ) and Q-Learning Based Ludo Players

CSL465/603 - Machine Learning

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

DOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS. Elliot Singer and Douglas Reynolds

Switchboard Language Model Improvement with Conversational Data from Gigaword

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

SUPRA-SEGMENTAL FEATURE BASED SPEAKER TRAIT DETECTION

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Transcription:

SPEECH EMOTION RECOGNITION WITH SKEW-ROBUST NEURAL NETWORKS Po-Yuan Shih and Chia-Ping Chen National Sun Yat-sen University Computer Science and Engineering Kaohsiung, Taiwan ROC Hsin-Min Wang Academia Sinica Institute of Information Science Taipei, Taiwan, ROC ABSTRACT We propose a neural-network training algorithm that is robust to data imbalance in classification. In our proposed algorithm, weights are introduced to training examples, effectively modifying the trajectory traversed in the parameter space during the learning process. Furthermore, the proposed algorithm would reduce to the normal stochastic gradient decent learning if the data is balanced. On the FAU-Aibo database, which is known to be used in Interspeech Emotion Challenge, the proposed method achieves an unweighted average (UA) recall rate of 45.3% on the 5-class speech emotion recognition task. Within the static modeling framework, where each example is represented as a fixed-length vector, this performance is one of the best performance ever achieved on the 5-class task. Index Terms speech emotion recognition, data imbalance, neural networks 1. INTRODUCTION A speech emotion recognition (SER) system takes a speech waveform as input, and outputs one of the emotional categories known to the classification system and hypothetically conveyed in the input speech. For examples of applications, SER can be incorporated in automatic speech recognition systems or spoken dialogue systems to improve recognition accuracy or user experience. SER evaluation plans have been implemented to promote global researches. The Interspeech 2009 Emotion Challenge (henceforth referred to as the Challenge) is a large-scale evaluation project to advance the technology of speech emotion recognition [1]. Since the release of the Challenge, many methods have been proposed to raise the bar of SER [2, 3, 4, 5, 6, 7, 8]. It may be hard to believe, but sound performance levels for the SER tasks as defined in the Challenge have not been achieved as of today. The highest unweighted average (UA) recall rate among submissions to the Challenge was 41.7% Thanks to the Ministry of Science and Technology for funding. Thanks to the Ministry of Science and Technology for funding. according to a summarization paper [9]. Following the Challenge, modern approaches for classification have been proposed to improve the performance on the 5-class task. For the static modeling framework in which each speech chunk is represented by a fixed-size vector, the highest UA achieved is 44.0% with anchor models [4]. For the dynamic modeling framework in which each speech chunk is represented by a variable-length sequence of feature vectors, the highest UA achieved is 45.6% with hybrid HMM-DNN systems [6]. Admittedly, there is much room for improvement. Two of the reasons for such difficulty, we believe, are the issue of skewed database and the intrinsic ambiguity of emotion expression. Both issues pose general challenges in machine learning. Using skewed data for system development is not an unusual scenario, as the data collection for a machinelearning system is often automated nowadays to reduce human factors. Further, as classification tasks move from areas of well-defined classes to uncharted territories, uncertainty in the labels is bound to happen either due to crowd-sourcing or the intrinsic ambiguity between target classes. In this paper, we propose skew-robust neural networks which use data weighting to deal with skewed data. New training objective functions are incorporated into learning process, to emphasize the examples of the small emotional classes and de-emphasize the examples of the large emotional classes. Furthermore, as part of the ambiguity of emotion expression comes from the difference between speakers, we also apply cross-speaker histogram equalization method to ameliorate such differences. The following sections of this paper are organized as follows. We introduce the basic ideas and describe the proposed methods in Section 2. Experiments and evaluation results are presented in Section 3. Finally, the concluding remarks are given in Section 4. 2. PROPOSED METHOD FAU-Aibo database consists of highly skewed data set, which posts a serious challenge on emotion classification. The number of speech chunks in FAU-Aibo is shown in Figure 1. A naïve classifier that simply assigns each test chunk to the Neu- 978-1-5090-4117-6/17/$31.00 2017 IEEE 2751 ICASSP 2017

6000 5000 4000 3000 2000 1000 0 A E N P R Training 881 2093 5590 674 721 Test 611 1508 5377 215 546 Training Fig. 1. Decomposition of class-wise speech chunks in FAU- Aibo corpus: anger (A), emphatic (E), Neutral (N), positive (P), and rest (R). tral class would achieve a weighted average (WA) recall rate of 65%. Thus, WA is not a good measure of performance. The measure of performance adopted in the Challenge is the unweighted average (UA) recall rate, which is the average of the recall rates of different classes. The above classifier would achieve an UA of 20%, which would be more reasonable. Test the training set equally. Without further ado, the parameters of the network will be trained to favor the class with the most training examples. In this paper, in order to make the most use of the precious data of small class and to reduce the swamping effect of large class, we propose to modify the objective function to where n=1 k=1 K t nk log y nk (2) t nk = r nk t nk (3) and r nk is the balance factor of class k for training example x n. More specifically, let C t(n) be the class of x n, then we have { 0, if k t(n) r nk = αn 1 (4) k, if k = t(n) Being inversely proportional to N k in Eq. (4), r nk has the effects of emphasizing the impacts of the errors of smallclass examples and of subduing the impacts of the errors of large-class examples during error back-propagation. Through r nk s, the relative contributions to the gradient of the objective function by examples of different classes are continuously adjusted, which increases the robustness of the training algorithm to skewed data. 2.1. Skew-Robust Neural Networks A classification system trained with unbalanced training data tends to predict a class that is populous in the training data. In this paper, we propose the following method, called skewrobust neural networks, to reduce such undesirable effects. We first introduce the basic ideas. Let the classes be C 1, C 2,..., C K where K is the number of classes. Let the number of training examples in these classes be N 1, N 2,..., N K respectively. In discriminative models such as neural networks, it is common practice to use one-hot (a.k.a. bit vector) representation for class label, and minimize the sum of cross-entropies 1 between targets and outputs, i.e. n=1 k=1 K t nk log y nk (1) where t nk is the target value of class k for example x n and y nk is the network output of class k, for training example x n. Parameters are often updated using error back propagation. The objective function Eq. (1) treats each example in 1 This function is the negative logarithm of the data-likelihood function, i.e. the error function. 2.2. Motivation and Benefits The proposed objective function Eq. (2) and the original objective function Eq. (1) are different in the effects they create during error back propagation for parameter learning. Specifically, using r nk as defined in Eq. (4), we get a better gradient that is emphasized (i.e. more aggressive error back propagation) in the direction to get the output prediction right for an example of a small class. The proposed method renders off-line data-balancing methods, such as SMOTE [10] for up-sampling or spread sub-sample for down-sampling, unnecessary. In principle, the proposed method is important as learning, whether by machine or by nature, is often based on highly skewed data sets. In practice, it is also important as up-sampling and down-sampling methods are time-consuming and introduces random artifacts that are unexpected. 2.3. Speaker Normalization In order to eliminate the difference due to speakers and other non-emotion factors, we apply cross-speaker histogram equalization (CSHE) to normalize data [8]. The principle of CSHE is outlined as follows. Suppose we have the cumulative distribution function (CDF) c Y (y) of a random variable Y. Furthermore, suppose we have examples for another random variable X D X = {x 1,.., x n }. (5) 2752

xˆ D X histogram estimation histogram estimation c X (x) (y) histogram mapping c ( yˆ) c ( xˆ ) Y D Y X c Y ŷ Table 1. Low-level descriptors (LLD) and functionals. LLDs (16) Functionals (12) ZCR mean RMS Energy standard deviation F0 kurtosis, skewness HNR extremes, position, range MFCC 1 12 regression coefficients, MSE Fig. 2. Cross-speaker histogram equalization. From D X, the CDF c X (x) can be estimated. HE transforms an instance x of X to the value y of Y such that c Y (y) = c X (x), (6) that is y = c 1 Y (c X(x)). (7) Thus, x and y are equalized in their CDF values and they correspond to the same bin in the respective histograms. The implementation of CSHE is illustrated in Figure 2. FAU-Aibo consists of speech data from 26 speakers for the training set and 25 speakers for the test set. We take all the training data distribution as the target distribution of a virtual speaker, and convert the feature distribution of all speakers to this target distribution. Different versions of CSHE are further specified in Section 3.3. 3. EXPERIMENT 3.1. Data and Experimental Setting 3.1.1. Data FAU-Aibo corpus contains recordings of spontaneous speech from 51 children as they are interacting with a SONY s pet robot Aibo. The data used in the Challenge consists of 9,959 chunks as training set and 8,257 chunks as test set. For the 5-class classification task, the speech chunks are labelled as anger (A), emphatic (E), Neutral (N), positive (P), and rest (R). Each speech chunk is emotionally-labeled by five persons, and the final label is decided by majority voting. As it turns out, the numbers of chunks of the classes are highly unbalanced: A (8.8%), E (21%), N (56.1%), P (6.8%), and R (7.2%). 3.1.2. Feature Extraction For speech features, we use the feature set shown in Table 1, including features related to prosody, spectral shape, voice quality, and their derivatives. Specifically, the 16 lowlevel descriptors (LLDs) are zero-crossing rate (ZCR), root Table 2. Weights for skew-robust neural networks in the FAU-Aibo Emotional Challenge 5-class task. N k is the number of chunks for class C k, and r nk is the weight. N k r nk for k = t(n) Anger 831 1.1 Emphatic 2093 0.5 Neutral 5590 0.2 Positive 674 1.5 Rest 721 1.4 mean square (RMS), pitch frequency (normalized to 500 Hz), harmonics-to-noise ratio (HNR), and mel-frequency cepstral coefficients (MFCC). The delta of LLDs are extracted as well. The 12 functionals are mean, standard deviation, kurtosis, skewness, minimum and maximum values, relative position and range, and the coefficients and the mean squared error (MSE) of linear regression. Thus, the final feature vector contains 16 2 12 = 384 attributes per speech chunk. 3.2. Skew-Robust Neural Networks The weights we use for r nk in Eq. (4) are listed in Table 2. These weights are decided by making them inversely proportional to N k and having magnitudes near 1. Furthermore, it is guaranteed that errors are magnified in the classes of Positive, Rest, and Anger, which have smaller numbers of examples than Neutral and Emphatic. The following settings are experimented to investigate the effectiveness of the proposed skew-robust methods. SVM: A support vector machine (SVM) system based on importance weights [11]. NN: A multi-layer perceptron with a hidden layer. DNN: A feed-forward deep neural network with 2 hidden layers. Here SVM is trained by the modified objective function of w 2 + C r nt(n) ξ n (8) n=1 where ξ n 0 is the penalty incurred by the nth data example, C > 0 controls the trade-off between the hyperplane margin 2753

Table 3. Unweighted average recall rates (UA) of the FAU- Aibo Emotional Challenge 5-class task with skew-robust neural networks. Please see text for details. SVM NN DNN baseline 29.8 30.1 28.6 +SR 41.1 39.6 39.9 +CSHE+SR 42.6 45.3 44.9 1.2 1 0.8 0.6 0.4 all basic/train-only one-speaker speaker-knn and the penalty, and the same weights r nk as listed in Table 2 are used. The topology of MLP and DNN are respectively 0.2 384 30 5, 384 20 15 5. Experimental results are shown in Table 3. We have the following comments. With NN, the proposed algorithm leads to significant improvements from 30.1% UA to 39.6% UA. When combined with CSHE, the performance is 45.3%, which is the best performance we achieve in this work. The proposed method outperforms the off-line databalancing method SMOTE (approximately 39% [1]). With SVM, the proposed method achieves an UA of 41.1%, and the combination with CSHE achieves 42.9%. The idea of giving more weights to examples of small classes also works with SVM. With DNN, the proposed method achieves an UA of 39.9%, and the combination with CSHE achieves 44.9%. In this case, i.e. adding an additional layer of hidden units to NN, DNN is not better than NN due to the limited amount of training data. 3.3. Speaker Normalization The following settings are experimented to investigate CSHE in different scenarios. In basic, CSHE is not applied to train or test data. In tr-only, CSHE is applied to train data, but not to test data. In one-spkr, the test data is assumed to be from one virtual speaker when CSHE is applied. In spkr-knn, a test example is assumed to be from the speaker recognized by knn. In all, CSHE is applied to every speaker in training and test set, using ground-truth speaker information. Note that all the above versions do not require speaker information during testing phase, except for the case of all, which is applicable only when the speaker information is available, i.e., speaker-dependent speech emotion recognition. Versions of CSHE can be applied to speaker-independent speech emotion recognition. In the case of spkr-knn, we implement a speaker recognition system based on knn, where the number of voting nearest neighbors is set to k = 90. The cumulative distributions of 0 0 0.01 0.02 0.03 Fig. 3. Cumulative distributions of the RMS energy mean feature of the test set data for different scenarios of CSHE. Table 4. Unweighted average recall rates (UA) of the FAU- Aibo Emotional Challenge 5-class task with different versions of CSHE. SVM SVM+SR NN NN+SR basic 29.8 41.1 30.1 39.6 all 32.3 42.6 33.7 45.3 tr-only 29.7 40.3 35.4 42.9 one-spkr 31.1 41.7 29.5 41.1 spkr-knn 28.0 37.5 27.9 40.1 test data in different scenarios of CSHE is shown in Figure 3. These scenarios lead to different empirical data distributions. The results of the FAU-Aibo 5-class task are shown in Table 4. For the proposed skew-robust neural networks, the best performance of UA is achieved in the case all at 45.3%. The performance of NN+SR in the case basic at 39.6% is improved by every version of CSHE, i.e. one-spkr at 41.1%, tr-only at 42.9%, and spkr-knn at 40.1%. Thus, the speaker-dependent system is more accurate in emotion recognition than speaker-independent system, which can be improved by unsupervised speaker normalization. 4. CONCLUSION We investigate a skew-robust parameter-learning method for neural networks. The main idea is to introduce weights to training examples, which effectively modifies the traversed trajectory in the parameter space during the learning process. In addition, we apply cross-speaker histogram equalization method to reduce the emotion expression of different speakers. Evaluated on the FAU-Aibo 5-class task, the proposed methods achieve an unweighted average recall rate of 45.3%. In light of the status quo as reviewed in Section 1, this is one of the best performance in the static modeling framework. 2754

5. REFERENCES [1] Björn Schuller, Stefan Steidl, and Anton Batliner, The Interspeech 2009 emotion challenge, in Proceedings of Interspeech 2009, pp. 312 315. [11] Andrew Rosenberg, Classifying skewed data: Importance weighting to optimize average recall, in Proceedings of Interspeech 2012, pp. 2242 2245. [2] Norhaslinda Kamaruddin and Abdul Wahab, Emulating human cognitive approach for speech emotion using MLP and GenSofNN, in Proceedings of Information and Communication Technology for the Muslim World 2013. [3] Chi-Chun Lee, Emily Mower, Carlos Busso, Sungbok Lee, and Shrikanth Narayanan, Emotion recognition using a hierarchical binary decision tree approach, Speech Communication, vol. 53, no. 9, pp. 1162 1171, 2011. [4] Yazid Attabi and Pierre Dumouchel, Anchor models for emotion recognition from speech, Affective Computing, IEEE Transactions on, vol. 4, no. 3, pp. 280 290, 2013. [5] Cheng Zha, Ping Yang, Xinran Zhang, and Li Zhao, Spontaneous speech emotion recognition via multiple kernel learning, in Proceedings of international Conference on Measuring Technology and Mechatronics Automation (ICMTMA), 2016. [6] Longfei Li, Yong Zhao, Dongmei Jiang, Yanning Zhang, Fengna Wang, Ivan Gonzalez, Emmanuel Valentin, and Hichem Sahli, Hybrid deep neural network hidden Markov model (DNN-HMM) based speech emotion recognition, in Proceedings of Affective Computing and Intelligent Interaction 2013, pp. 312 317. [7] Vidhyasaharan Sethu, Eliathamby Ambikairajah, and Julien Epps, Speaker normalisation for speech-based emotion detection, in Proceedings of International Conference on Digital Signal Processing 2007, pp. 611 614. [8] Bo-Chang Chiou, Cross-lingual automatic speech emotion recognition, M.S. thesis, National Sun Yat-sen University, 2014. [9] Björn Schuller, Anton Batliner, Stefan Steidl, and Dino Seppi, Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge, Speech Communication, vol. 53, no. 9, pp. 1062 1087, 2011. [10] Nitesh V. Chawla, Kevin W. Bowyer, Lawrence O. Hall, and W. Philip Kegelmeyer, SMOTE: synthetic minority over-sampling technique, Artificial Intelligence Research, pp. 321 357, 2002. 2755