SPEAKER INDEXING IN LARGE AUDIO DATABASES USING ANCHOR MODELS. D. E. Sturim 1 D. A. Reynolds 2, E. Singer 1 and J. P. Campbell 3

Similar documents
Learning Methods in Multilingual Speech Recognition

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Intelligent Agent Technology in Command and Control Environment

DOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS. Elliot Singer and Douglas Reynolds

Investigation on Mandarin Broadcast News Speech Recognition

Support Vector Machines for Speaker and Language Recognition

Modeling function word errors in DNN-HMM based LVCSR systems

AD (Leave blank) PREPARED FOR: U.S. Army Medical Research and Materiel Command Fort Detrick, Maryland

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Speech Recognition at ICSI: Broadcast News and beyond

A study of speaker adaptation for DNN-based speech synthesis

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

Modeling function word errors in DNN-HMM based LVCSR systems

Human Emotion Recognition From Speech

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Probabilistic Latent Semantic Analysis

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

SEDETEP Transformation of the Spanish Operation Research Simulation Working Environment

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

Speech Emotion Recognition Using Support Vector Machine

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

On-Line Data Analytics

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Automatic Speaker Recognition: Modelling, Feature Extraction and Effects of Clinical Environment

A Case Study: News Classification Based on Term Frequency

WHEN THERE IS A mismatch between the acoustic

Improvements to the Pruning Behavior of DNN Acoustic Models

CyberCIEGE: An Extensible Tool for Information Assurance Education

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

NCEO Technical Report 27

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Speech Recognition by Indexing and Sequencing

First Grade Standards

Word Segmentation of Off-line Handwritten Documents

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Spoofing and countermeasures for automatic speaker verification

Montana Content Standards for Mathematics Grade 3. Montana Content Standards for Mathematical Practices and Mathematics Content Adopted November 2011

Speaker recognition using universal background model on YOHO database

Digital Signal Processing: Speaker Recognition Final Report (Complete Version)

Reducing Features to Improve Bug Prediction

Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology

Non intrusive multi-biometrics on a mobile device: a comparison of fusion techniques

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Calibration of Confidence Measures in Speech Recognition

On the Combined Behavior of Autonomous Resource Management Agents

Python Machine Learning

Evaluation of a College Freshman Diversity Research Program

Grade 6: Correlated to AGS Basic Math Skills

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Detecting Wikipedia Vandalism using Machine Learning Notebook for PAN at CLEF 2011

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method

Cross Language Information Retrieval

ACOUSTIC EVENT DETECTION IN REAL LIFE RECORDINGS

Lecture Notes in Artificial Intelligence 4343

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

Wisconsin 4 th Grade Reading Results on the 2015 National Assessment of Educational Progress (NAEP)

Knowledge Transfer in Deep Convolutional Neural Nets

On the Formation of Phoneme Categories in DNN Acoustic Models

UMass at TDT Similarity functions 1. BASIC SYSTEM Detection algorithms. set globally and apply to all clusters.

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Data Fusion Models in WSNs: Comparison and Analysis

Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition

Mathematics subject curriculum

Page 1 of 11. Curriculum Map: Grade 4 Math Course: Math 4 Sub-topic: General. Grade(s): None specified

Disambiguation of Thai Personal Name from Online News Articles

Speaker Recognition. Speaker Diarization and Identification

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Automatic Pronunciation Checker

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Evaluation of Systems Engineering Methods, Processes and Tools on Department of Defense and Intelligence Community Programs - Phase II

Math-U-See Correlation with the Common Core State Standards for Mathematical Content for Third Grade

BODY LANGUAGE ANIMATION SYNTHESIS FROM PROSODY AN HONORS THESIS SUBMITTED TO THE DEPARTMENT OF COMPUTER SCIENCE OF STANFORD UNIVERSITY

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

The Talent Development High School Model Context, Components, and Initial Impacts on Ninth-Grade Students Engagement and Performance

Application of Virtual Instruments (VIs) for an enhanced learning environment

Preprint.

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

Lecture 1: Machine Learning Basics

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

Rule Learning With Negation: Issues Regarding Effectiveness

Australian Journal of Basic and Applied Sciences

This scope and sequence assumes 160 days for instruction, divided among 15 units.

First Grade Curriculum Highlights: In alignment with the Common Core Standards

BENCHMARK TREND COMPARISON REPORT:

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Lecture 2: Quantifiers and Approximation

Deep Neural Network Language Models

Progress Monitoring & Response to Intervention in an Outcome Driven Model

South Carolina College- and Career-Ready Standards for Mathematics. Standards Unpacking Documents Grade 5

Grade Dropping, Strategic Behavior, and Student Satisficing

The Smart/Empire TIPSTER IR System

Semi-Supervised Face Detection

Transcription:

SPEAKER INDEXING IN LARGE AUDIO DATABASES USING ANCHOR MODELS D. E. Sturim 1 D. A. Reynolds, E. Singer 1 and J. P. Campbell 3 1 MIT Lincoln Laboratory, Lexington, MA Nuance Communications, Menlo Park, CA 3 Department of Defense fsturim,dar,esg@sst.ll.mit.edu, j.campbell@ieee.org ABSTRACT This paper introduces the technique of anchor modeling in the applications of speaker detection and speaker indexing. The anchor modeling algorithm is refined by pruning the number of models needed. The system is applied to the speaker detection problem where its performance is shown to fall short of the state-of-the-art Gaussian Mixture Model with Universal Background Model (-UBM) system. However, it is further shown that its computational efficiency lends itself to speaker indexing for searching large audio databases for desired speakers. Here, excessive computation may prohibit the use of the -UBM recognition system. Finally, the paper presents a method for cascading anchor model and -UBM detectors for speaker indexing. This approach benefits from the efficiency of anchor modeling and high accuracy of -UBM recognition. 1. INTRODUCTION This paper describes a method of representing and characterizing a target utterance with information gained from a set of anchor models derived from a predetermined set of speakers. Since the speakers of the target utterances are not members of the model training set, the system is capable of characterizing the target speaker with no prior knowledge of that speaker. Previous research [1, ] suggests that the target speaker will be projected into a talker space defined by the anchor models. Since the models are created only once in the training phase, it is unnecessary to train a model for a new target speaker. Applications of the approach include speaker recognition, speaker detection, and speaker clustering for very large speaker populations where it is undesirable or infeasible to train models for every member of the target population. Another application of anchor modeling discussed in this paper is speaker indexing; that is, the use of speaker detection for the retrospective searching of large speech archives. For large archives, current stateof-the-art speaker recognition systems may be too computationally inefficient for large searches. The efficiency of the anchor system lends itself to the application of large speech archive retrieval. It is shown that although the detection performance of the anchor model system falls short This work was sponsored by the Department of Defense under Air Force Contract F1968-00-C-000. Opinions, interpretations, conclusions, and recommendations are those of the authors and not necessarily endorsed by the United States Air Force. of state-of-the-art Gaussian Mixture Model with Universal Background Model (-UBM) speaker detection systems [3, 4], the efficiency of anchor modeling can be effectively exploited by embedding it in a two-stage cascaded system, where the role of the anchor system is to reduce the data load of the more accurate but less computationally efficient -UBM.. ANCHOR MODELS The basic concept of anchor modeling is the representation of a target speech utterance with information gained from a set of models pre-trained from a defined set of talkers. In theory, the models could consist of virtually any method of speech representation. Previous work [1, ] used speakerdependent Hidden Markov Models (HMM) as the anchors. This study uses the -UBM as the representation model for forming the anchors. Segments of speech, s, are scored against a set of pretrained anchor models, A i, i =1; :::; N. Each of the N anchor models yields a likelihood score and the collection of scores is used to form the N-dimensional characterization vector. The speech utterance is represented by this characterization vector V, where V = 6 6 4 p(sja 1 ) p(sja ).. p(sja N ) 3 7 7 5 (1) The characterization vector can be considered a projection of the target utterance into a speaker space defined by the anchor models. If an utterance from a single speaker projects into a unique portion of the speaker space, then the speaker representation is unique. Speaker detection is performed by considering the location of the vectors within this speaker space. Speech segments are compared by scoring a speech segment s u from an unknown speaker and a speech segment s t from a target speaker against the same set of anchor models (Figure 1), thereby forming two characterization vectors, V u and V t, to represent the unknown and target segments of speech. A vector distance is then used to compare the speech segments. Preliminary experiments using Euclidean, absolute value or city block, and Kullback - Leibler distance measures showed that Euclidean distance performed best. Unit nor-

Report Documentation Page Form Approved OMB No. 0704-0188 Public reporting burden for the collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden, to Washington Headquarters Services, Directorate for Information Operations and Reports, 115 Jefferson Davis Highway, Suite 14, Arlington VA -430. Respondents should be aware that notwithstanding any other provision of law, no person shall be subject to a penalty for failing to comply with a collection of information if it does not display a currently valid OMB control number. 1. REPORT DATE 01. REPORT TYPE N/A 3. DATES COVERED - 4. TITLE AND SUBTITLE Speaker Indexing In Large Audio Databases Using Models 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT NUMBER 5e. TASK NUMBER 5f. WORK UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) MIT Lincoln Laboratory, Lexington, MA, USA 8. PERFORMING ORGANIZATION REPORT NUMBER 9. SPONSORING/MONITORING AGENCY NAME(S) AND ADDRESS(ES). SPONSOR/MONITOR S ACRONYM(S) 1. DISTRIBUTION/AVAILABILITY STATEMENT Approved for public release, distribution unlimited 13. SUPPLEMENTARY NOTES The original document contains color images. 14. ABSTRACT 15. SUBJECT TERMS 11. SPONSOR/MONITOR S REPORT NUMBER(S) 16. SECURITY CLASSIFICATION OF: 17. LIMITATION OF ABSTRACT UU a. REPORT b. ABSTRACT c. THIS PAGE 18. NUMBER OF PAGES 4 19a. NAME OF RESPONSIBLE PERSON Standard Form 98 (Rev. 8-98) Prescribed by ANSI Std Z39-18

Unknown speech Target speech Compute Compute likelihood likelihood 1 3 Compute Compute likelihood likelihood p(s u A 1 ) p(s u A ) p(s u A 3 ) p(s t A 1 ) p(s t A ) p(s t A 3 ) Each anchor model is -UBM Characterization vector V u Vector Vector distance distance Figure 1: The anchor model system. V t D(V u,v t ) malizing the elements of characterization vectors in the distance calculation did not change performance. The -UBM anchor models described in this paper were trained using speech from 668 talkers in the NIST- 1996 and NIST-1999 speech corpora. 1 The -UBM algorithm used was the same as that developed for the NIST- 00 speaker recognition workshop [5, 6] but without speaker (T-NORM) and handset (H-NORM) normalizations..1. Model Pruning The full anchor model characterization vector is formed by scoring an utterance against all 668 anchor models. Methods of reducing the size of the Euclidean distance comparison were investigated in an effort to increase performance by using only those anchor models that provide good characterizing information. Reducing the size of the distance comparison reduces the dimensionality of the speaker space and increases computational efficiency. Model pruning strategies were motivated by the observation that the vector distance between characterization vectors derived from the same talker should be small while distances between characterization vectors of different speakers should be large. Characterization vectors of two utterances from the same talker were compared and the resulting element distances, d i, were rank ordered by magnitude, where h d i = (V ti V ui ) i () i=1:n and V t and V u are two characterization vectors obtained from two target speech utterances. A percentage of the models with the lowest element distances was then chosen as the anchor model set. In a similar manner, characterization vectors of utterances from different talkers can be evaluated with Equation, where V t and V u are now characterization vectors from different talkers. With this approach, only those models with the largest element distances are chosen for the anchor model set. Using these two methods of pruning, the size of the Euclidean distance comparison was reduced by 60% while the equal error rate was improved. 3. SPEAKER DETECTION WITH ANCHOR MODELS Results presented in this section used speech data from the NIST-00 Speaker Recognition Workshop, sectioning the 1 The data used in the NIST evaluation is a subset of the Switchboard I-II data corpora. Miss probability (in %) 5 1 0.5 -UBM AM with pruning Model 0.5 1 5 False Alarm probability (in %) Figure : DET curves for the -UBM and anchor model system using the primary condition of single speaker detection NIST-00 speech corpus. corpus into test and training sets and performing the evaluation using the protocols stipulated in [7]. Figure presents the Detection Error Tradeoff (DET) curves for the NIST- 00 single speaker detection task primary condition. The equal error rate for the anchor model system using the full characterization vector (N = 668) was 4.% while the equal error rate of the anchor system with model pruning was 1.4%. Pruning of the models provides a relative performance increase of 11.7%. The performance of the anchor system falls well short of the 7.7% equal error rate of the -UBM system. The next section discusses one application of speaker detection where the computational efficiency of the anchor modeling approach is used to advantage. 4. SPEAKER INDEXING Speaker indexing is defined as the application of speaker detection to the retrospective search of large speech archives. Two possible uses of speaker indexing are the clustering of speech messages contained in a speech archive and the retrieval of a list of messages from an archive in response to an external query. This paper focuses on the list retrieval task. Performance in speaker detection evaluations has traditionally been reported using a (prior-independent) DET curve that describes the underlying tradeoff between misses and false alarms for a given detector and corpus. However, performance in information retrieval applications such as speaker indexing is better described using the notions of precision and recall. Detection theory and information retrieval measures are related as follows: Recall is the proportion of relevant material retrieved from the archive and so is equal to the detection probability. Precision is the proportion of retrieved material that is relevant and is given by P t (1 P m ) P recision = (3) P t (1 P m )+(1 P t )P fa The NIST primary condition uses minute training segments and 15-45 second test segments collected with an electret microphone.

0 1 90 80 11 Precision (in %) 70 60 50 30 Gaussian computations 9 8 0 0 30 50 60 70 80 90 0 Recall (in %) Figure 3: Precision versus recall plot for the -UBM, and anchor model, with P t =9%. where P t is the target probability (richness) of the archive, P m is the probability of a miss, and P fa is the probability of false alarm. These relationships can then be used to derive speaker indexing performance (in terms of precision vs. recall) from a DET plot for any given target probability P t. 4.1. Evaluation of the and models for Speaker Indexing Figure 3 shows the precision versus recall tradeoff for the -UBM and anchor model speaker detectors using the DET plots of Figure (NIST-00 speech corpus) and an archive richness P t = 9% (the richness of the NIST-00 corpus). As expected, the -UBM method outperforms the anchor model. It is worth noting that the curves tend to move toward the upper right with increasing P t and toward the lower left with decreasing P t. Another measure of a speaker detector s value for speaker indexing applications is its computational efficiency. Here it is assumed that each item in the archive is represented by a model (trained off-line) against which a query is scored. For the -UBM, each ms frame of the query is first scored against the 48-component universal background model and then against 5 components of each of the archive models [5]. For anchor model based speaker indexing, the query is first converted to a characterization vector by scoring it against the 668 anchor s. The resulting characterization vector is then compared to each archive characterization vector (trained off-line) using a 668-element Euclidean distance. Figure 4 plots the number of 38-dimensional Gaussian computations (or equivalent) required for a 1 minute query. (It is assumed that the computation time for one 38- element Gaussian and 38 Euclidean distances are equal.) The plot for the anchor model system stays flat to about 6 because the computation is dominated by the conversion of the query to a characterization vector. Note that this is true for the pruned anchor system as well. It is apparent that the anchor model speaker indexing system has significant computational advantages for archives containing more than about 00 items. It should be noted that methods exist for speeding up the computation required for the that 7 3 4 5 6 7 8 Size of archive Figure 4: Plot of computational efficiency for the - UBM and anchor model speaker detectors. Large Archive Model Speaker Detection reduced archive UBM Speaker Detection Figure 5: d speaker detection system. putative target list would improve the efficiency of both the -UBM and anchor model systems. 4.. Cascading Figures 3 and 4 show the tradeoff of computational efficiency versus accuracy for speaker indexing. The - UBM has superior detection performance while the anchor system provides the computational efficiency that is essential when searching large archives. In an effort to gain a better tradeoff between computational performance and accuracy, the anchor and -UBM speaker detection systems were combined in a cascade as shown in Figure 5. The objective of cascading is to construct a system containing the positive aspects of both algorithms. The anchor model is employed in the first stage to reduce the amount of computational loading for the -UBM speaker detection system. The -UBM is then used to provide maximum recognition performance. To evaluate the performance of the cascade, it is first necessary to identify the operating point of the anchor system. Define q to be the fraction of the archive processed by the second system of the cascade (i.e., the probability that the first system declares a target). Note that q is the denominator of Equation 3: q = P t (1 P m )+(1 P t )P fa (4) where (1 P m ) is the probability of detection and P fa is the probability of false alarm for the anchor model speaker detector. Given that the richness of the archive (P t )isdefined by the application, choosing a unique value for q identifies a (P fa ;P m ) pair from the DET curve (Figure ) and represents the chosen operating point for the anchor system. The precision versus recall curve for the cascaded system can be calculated in the same manner as in Section 4.1. Figure 6 presents precision versus recall for the cascaded

Precision (in %) 0 90 80 70 60 50 30 0 0 30 50 60 70 80 90 0 Recall (in %) Figure 6: Precision versus recall plot for the -UBM, anchor model, and cascaded system with q = % and P t =9%. Gaussian computations 1 11 9 8 7 q= % q= 1% q= 0.1% 3 4 5 6 7 8 Size of archive Figure 7: Estimated number of Gaussian (or equivalent) computations, 1 minute query. system with q = % and an archive richness of P t =9%. The effect of the cascade is to slightly reduce the performance in operating regions of low recall and to drastically reduce performance in regions of mid-to-high recall, relative to the system. Figure 7 displays a plot of the estimated computational efficiency for the -UBM, anchor model, and cascaded speaker indexing systems. As the amount of reduction in archive size increases (smaller q), the computational efficiency of the cascaded system also increases. 5. SUMMARY This paper presented a method of characterizing a segment of a talker s speech with information gained from a set of pre-trained anchor models. The anchor models were derived from a set of predetermined speakers. Characterization vectors were then formed by scoring the target speech segment against the set of anchor models. A method for refining the anchor modeling system was presented increased recognition performance. modeling was then applied to the speaker detection problem. Detection error tradeoff performance showed that the anchor modeling system fell short of a state-ofthe-art -UBM system. It was further shown that its computational efficiency was superior to that of the - UBM. Comparison of the anchor model and -UBM systems for speaker indexing showed a similar tradeoff between precision versus recall performance and computational efficiency. A cascaded speaker indexing system was proposed that utilized the anchor model system as the first stage and the -UBM as the second stage. In this configuration, the anchor model reduced the data loading on the -UBM while slightly reducing performance in operating regions of low recall. The effect of the cascaded system was to combine the advantages of both systems at the expense of some loss in both computational performance and detection accuracy. For large archives, the recognition performance of the anchor system and the lack of computational efficiency of the -UBM system could preclude their application to speaker indexing. The cascaded system may offer a viable solution to the speaker indexing application. 6. REFERENCES [1] Douglas E. Sturim, Tracking and Characterization of Talkers Using a Speech Processing System with a Microphone Array as Input, Ph.D. thesis, Brown University, 1999. [] Teva Merlin, Jean-François Bonastre, and Corinne Fredouille, Non directly acoustic process for costless speaker recognition and indexation, International Workshop on Intelligent Communication Technologies and Applications, 1999. [3] Douglas Reynolds, Thomas Quatieri, and Robert Dunn, Speaker verification using adapted gaussian mixture models, Digital Signal Processing, vol., pp. 19 41, 00. [4] Roland Auckenthaler, Michael Carey, and Harvey Lloyd-Thomas, Score normalization for textindependent speaker verification systems, Digital Signal Processing, vol., pp. 4 54, 00. [5] D. A. Reynolds, Comparison of background normalization methods for text-independent speaker verification, in Proceedings of the European Conference on Speech Communication and Technology, 1997. [6] D. A. Reynolds, The effects of handset variability on speaker recognition performance: Experiments on the switchboard corpus, in IEEE International Conference on Acoustics, Speech and Signal Processing, 1996. [7] NIST, The 00 NIST Speaker Recognition Evaluation Plan, Linthicum, MD, June 00, http://www.nist.gov/speech/tests.