MODELING PRONUNCIATION VARIATION FOR CANTONESE SPEECH RECOGNITION

Similar documents
Learning Methods in Multilingual Speech Recognition

STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH

Modeling function word errors in DNN-HMM based LVCSR systems

Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition

Speech Recognition at ICSI: Broadcast News and beyond

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Modeling function word errors in DNN-HMM based LVCSR systems

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Investigation on Mandarin Broadcast News Speech Recognition

Speech Recognition using Acoustic Landmarks and Binary Phonetic Feature Classifiers

A Neural Network GUI Tested on Text-To-Phoneme Mapping

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

Mandarin Lexical Tone Recognition: The Gating Paradigm

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition

Calibration of Confidence Measures in Speech Recognition

A study of speaker adaptation for DNN-based speech synthesis

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

Probabilistic Latent Semantic Analysis

WHEN THERE IS A mismatch between the acoustic

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

English Language and Applied Linguistics. Module Descriptions 2017/18

Segmental Conditional Random Fields with Deep Neural Networks as Acoustic Models for First-Pass Word Recognition

Letter-based speech synthesis

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

DIRECT ADAPTATION OF HYBRID DNN/HMM MODEL FOR FAST SPEAKER ADAPTATION IN LVCSR BASED ON SPEAKER CODE

The NICT/ATR speech synthesis system for the Blizzard Challenge 2008

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Linking Task: Identifying authors and book titles in verbose queries

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Automatic Pronunciation Checker

Improvements to the Pruning Behavior of DNN Acoustic Models

Unsupervised Acoustic Model Training for Simultaneous Lecture Translation in Incremental and Batch Mode

Speech Emotion Recognition Using Support Vector Machine

Disambiguation of Thai Personal Name from Online News Articles

Edinburgh Research Explorer

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

Noisy Channel Models for Corrupted Chinese Text Restoration and GB-to-Big5 Conversion

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

Universal contrastive analysis as a learning principle in CAPT

Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology

A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language

Books Effective Literacy Y5-8 Learning Through Talk Y4-8 Switch onto Spelling Spelling Under Scrutiny

On the Formation of Phoneme Categories in DNN Acoustic Models

Distributed Learning of Multilingual DNN Feature Extractors using GPUs

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Phonological and Phonetic Representations: The Case of Neutralization

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

CEFR Overall Illustrative English Proficiency Scales

Consonants: articulation and transcription

LEARNING A SEMANTIC PARSER FROM SPOKEN UTTERANCES. Judith Gaspers and Philipp Cimiano

DOWNSTEP IN SUPYIRE* Robert Carlson Societe Internationale de Linguistique, Mali

Parallel Evaluation in Stratal OT * Adam Baker University of Arizona

SEMI-SUPERVISED ENSEMBLE DNN ACOUSTIC MODEL TRAINING

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Assignment 1: Predicting Amazon Review Ratings

Large vocabulary off-line handwriting recognition: A survey

An Introduction to the Minimalist Program

Prediction of Maximal Projection for Semantic Role Labeling

Phonological Processing for Urdu Text to Speech System

DNN ACOUSTIC MODELING WITH MODULAR MULTI-LINGUAL FEATURE EXTRACTION NETWORKS

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Beyond the Pipeline: Discrete Optimization in NLP

A Case Study: News Classification Based on Term Frequency

The Strong Minimalist Thesis and Bounded Optimality

Speaker recognition using universal background model on YOHO database

Deep Neural Network Language Models

Taking into Account the Oral-Written Dichotomy of the Chinese language :

Detecting English-French Cognates Using Orthographic Edit Distance

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Generating Test Cases From Use Cases

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Device Independence and Extensibility in Gesture Recognition

Characterizing and Processing Robot-Directed Speech

Building Text Corpus for Unit Selection Synthesis

Improved Hindi Broadcast ASR by Adapting the Language Model and Pronunciation Model Using A Priori Syntactic and Morphophonemic Knowledge

Contrastiveness and diachronic variation in Chinese nasal codas. Tsz-Him Tsui The Ohio State University

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Rule Learning With Negation: Issues Regarding Effectiveness

PHONETIC DISTANCE BASED ACCENT CLASSIFIER TO IDENTIFY PRONUNCIATION VARIANTS AND OOV WORDS

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

ADDIS ABABA UNIVERSITY SCHOOL OF GRADUATE STUDIES MODELING IMPROVED AMHARIC SYLLBIFICATION ALGORITHM

SARDNET: A Self-Organizing Feature Map for Sequences

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Artificial Neural Networks written examination

Proceedings of Meetings on Acoustics

Human Emotion Recognition From Speech

COPING WITH LANGUAGE DATA SPARSITY: SEMANTIC HEAD MAPPING OF COMPOUND WORDS

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

On Developing Acoustic Models Using HTK. M.A. Spaans BSc.

The Perception of Nasalized Vowels in American English: An Investigation of On-line Use of Vowel Nasalization in Lexical Access

SEGMENTAL FEATURES IN SPONTANEOUS AND READ-ALOUD FINNISH

Transcription:

MODELIG PROUCIATIO VARIATIO FOR CATOESE SPEECH RECOGITIO Patgi KAM and Tan LEE Department of Electronic Engineering The Chinese University of Hong Kong, Hong Kong {pgkam, tanlee}@ee.cuhk.edu.hk ABSTRACT Due to the large variability of pronunciation in spontaneous speech, pronunciation modeling becomes a more challenging and essential part in speech recognition. In this paper, we describe two different approaches of pronunciation modeling by using decision tree. At lexical level, a pronunciation variation dictionary is built to obtain alternative pronunciations for each word, in which each entry is associated with a variation probability. At decoding level, decision tree pronunciation models are applied to expand the search space to include alternative pronunciations. Relative error reduction of 7.21% and 4.81% could be achieved at lexical level and decoding level respectively. The results at the two different levels are compared and contrasted. 1. ITRODUCTIO The primary goal of speech recognition is to produce a textual transcription for spoken input. This can be done by establishing a mapping between the extracted acoustic features and the underlying linguistic representations. Given the high variability of human speech, such mapping is not one-to-one. Different linguistic symbols can give rise to similar speech sounds while each symbol may have multiple pronunciations. The variability is due to co-articulation, regional accent, speaking rate, speaking style, etc. Pronunciation modeling (PM) for automatic speech recognition (ASR) is aimed at providing a mechanism by which speech recognition systems can be adapted to pronunciation variability. In a large vocabulary continuous speech recognition (LVCSR) system, three knowledge sources are involved: pronunciation lexicon, acoustic model (AM) and language model (LM). They are used to form a search space from which the most likely sentence(s) or word string(s) is decoded. Within this framework, modeling of pronunciation variations can be done by explicitly modifying the knowledge sources and/or improving the decoding technique. Pronunciation lexicon provides constraints on the combination of speech sounds at the lowest linguistic level. Conventionally, the lexicon contains a baseform transcription for each word in the form of a phoneme sequence. The baseform transcription, also known as canonical transcription, is assumed to be the standard pronunciation of word that the speaker is supposed to use. If there exist alternative pronunciations of the word, they need to be included in the lexicon. These additional items are commonly referred to as surfaceform transcriptions, which are the actual pronunciations that different speakers may use [1][2]. The existence of alternative pronunciations implies that the acoustic models may not be accurate enough to represent the variations of speech sounds. Indeed, in most cases, acoustic models are trained with the assumption that only baseform pronunciations are used. Thus, it would be useful to retrain or refine the acoustic models according to more realistic pronunciations [3][4]. Pronunciation modeling can also be done by expanding the search space for sentence decoding. Being augmented with pronunciation variants, the search space is expected to contain more useful information for the search. In this paper, we focus on the use of decision tree based techniques for automatic prediction of pronunciation variability. The pronunciation modeling techniques are developed and evaluated for continuous Cantonese speech recognition. We investigate the effectiveness of two methods in which pronunciation modeling is applied at lexical level and decoding level respectively. 2. BACKGROUD 2.1. The Cantonese dialect Mandarin and Cantonese are two important dialects of Chinese. The former is the official standard of spoken Chinese while the latter is the most influential dialect in South China, Hong Kong and overseas. Like Mandarin, Cantonese is monosyllabic and tonal. Each Chinese character is pronounced as a monosyllable [5]. A Chinese word is composed of one or more characters. Most characters can also be a meaningful word by themselves. A Cantonese syllable can be divided into an Initial (I) and a Final (F) [6]. There are totally 20 Initials and 53 Finals. Initials and Finals are combined under certain phonological constraints and as a result, there are over 600 legitimate I-F combinations, referred to as base syllables. Table 1 shows the structure of a Chinese word. The Chinese word (we) is a two-syllable word. The base syllable ngo is formed by the Initial I_ng and the Final F_o. The syllable mun is formed by the Initial I_m and the Final F_un. Chinese Chinese Base Sub-syllable units word character syllable ngo I_ng F_o mun I_m F_un Table 1. The structure of a Chinese word.

2.2. LVCSR for Cantonese For Cantonese LVCSR, context-dependent Initials and Finals are usually used as the basic units for acoustic modeling by Hidden Markov Models (HMM). In this research, the acoustic models being used are cross-word bi-if HMMs trained with 20 hours of continuous speech from the CUSET corpus developed by the Chinese University of Hong Kong [7]. The acoustic models are used with a class-based bi-gram language model. The target application deals with domain-specific spoken queries, i.e. stock information inquiry. Pronunciation models are used to derive or predict surfaceform transcriptions from baseform transcription. LetB and S denote respectively the baseform and the surfaceform transcriptions at Initial-Final level. Table 2 shows an example of baseform and surfaceform transcriptions for the word. Chinese word B I_ng F_o I_m F_un S I_ng F_o I_m F_un I_null F_o I_m F_un I_ng F_o I_w F_un I_null F_o I_w F_un Table 2. Baseform and surfaceform transcriptions of the word Two different decoders are under investigation. The first one is a one-pass decoder, in which the knowledge sources are used all at a time to construct the search space. The second decoder performs search in two stages. In stage 1, acoustic models are used to generate a lattice of Initials and Finals. The ultimate sentence output is generated by stage 2 with the assistance of language models. For the one-pass decoder, pronunciation variants can be introduced by either explicitly including the surfaceform pronunciations in the lexicon or dynamically expanding the search space during the decoding process. In the case of twopass decoding, pronunciation models can be used to augment the intermediate search space between the two search stages. The data used in our research includes 1200 utterances from CUSET corpus test set, named as CUTEST, and 1300 utterances of spoken queries on stock information, named as STOCKTEST. The former is used to build 3 sets of decision tree PMs for 3 different experiments. The latter is used as testing data for the 3 experiments. 3. USE OF PROUCIATIO VARIATIO DICTIOAR To incorporate pronunciation modeling at lexical level, one of the methods is to use a pronunciation model to build an augmented dictionary to include alternative pronunciations. The resultant lexicon is referred to as pronunciation variation dictionary (PVD). To use the PVD, the recognition process needs to be modified to take care of the newly added pronunciation variants. This is done by incorporating the variation probabilities (VP) into the decoding process. Given an acoustic observation O, the goal of recognition is to find the word sequence W that maximizes the probability P(W O). According to the Bayes Rule, we have W* = arg max P( W ) P( O W ).(1) W where P(W) is given by language model and P(O W) is computed from acoustic model and pronunciation lexicon. If pronunciation variations are taken into account, equation (1) is modified to: W = arg max P( W ) P( O S ) P( S ).(2) * w, k w, k W W where S w,k is the k th pronunciation variant for the word W. The modified equation essentially searches for a particular pronunciation variant that maximizes the probability. P(O S w,k ) is the acoustic likelihood of pronunciation S w,k.p(s w,k W) gives the probability that W is pronounced as S w,k. 4. COSTRUCTIG THE PVD 4.1. How to obtain the pronunciation variations? PVD is the conventional lexicon with additional alternative pronunciations such that pronunciation variations can be handled. To build a PVD, we have to find out what are the variants (surfaceform transcriptions) to be included in the lexicon. One way is to derive them from a set of speech data by using a proper PM. The speech data we used is the 1200 utterances of CUTEST mentioned in Section 2.2. PM is used to make predictions from baseform transcription of CUTEST. The predicted alternatives provide constraints to phone recognition for CUTEST to obtain the most likely surfaceform [8]. A confusion matrix, which shows the possible variants of a particular IF unit, can be obtained by aligning the baseform transcription with the surfaceform transcription of CUTEST and tabulating the frequency of each surfaceform. Table 3 shows part of a confusion matrix in table form for the Initial I_m and I_ng, and Final F_o and F_un. Baseform B Surfaceform S Variation Probability (VP)% I_m I_m 80 I_m I_w 20 I_ng I_ng 30 I_ng I_null 70 F_o F_o 100 F_un F_un 100 Table 3. Confusion matrix in table form for the Initial I_m and I_ng, and Final F_o and F_un with the corresponding variation probabilities. Too many variations added to the dictionary will introduce excessive confusion during the searching process. ormally, threshold is set to filter rarely occurred surfaceforms. By using the confusion matrix, PVD, which contains pronunciation alternatives for each word, is built. Table 4 shows part of the PVD containing the surfaceform transcriptions and the corresponding variation probabilities of the word. P(S w,k W) for each surfaceform is obtained by multiplying the VPs of all individual surfaceform IFs composing the word as givenintable3.thepvdcanthenbeusedinthedecoding process to find a particular pronunciation variant that maximizes the probability P(W O).

W B S w,k P(S w,k W) I_ng F_o I_m F_un 0.24 I_ng F_o I_m F_un I_null F_o I_m F_un 0.56 I_ng F_o I_w F_un 0.06 I_null F_o I_w F_un 0.14 Table 4. A part of the PVD showing the surfaceforms of the word. The alignment between the baseform and the surfaceform transcription of CUTEST can be used to train another set of PM. Repeating the steps above can obtain a reweight and augmented dictionary, named as retrained dictionary. 4.2. Decision tree pronunciation models Decision tree is essentially a context dependent PM used to predict the surfaceform phones given the baseform phone. A decision tree, which is shown in Figure 1, contains a binary question (yes/no answer) about the phonetic features at each node in the tree. The leaves of the tree contain the best predictions (surfaceform phones) based on training data. In our study, the data used to build the decision tree PMs includes 1200 utterances from CUTEST. The baseform Initial/Final transcription of CUTEST, which is the manually verified version, together with the surfaceform transcription, obtained from phone recognition, form a set of training vectors. The tree context concerns the baseform unit under consideration (C b ), left baseform unit (L b ) and the right baseform unit (R b ). The stopping criterion requires a minimal number of samples in the parent node and child node [9]. One decision tree is built for each Initial and Final. 20 Initials and 53 Finals will result in 73 different trees. The decision tree PMs are applied to the baseform transcription of the training corpus to construct a lattice of pronunciation alternatives for phone recognition, so as to obtain the most likely surfaceform. Lb=I_s? Rb=I_s? Lb=I_c? Lb=I_k? F_an 0.07, F_oeng 0.93 < F_oeng I_s I_s > Rb=I_s? <Cb Lb Rb > F_eon 0.03, F_oeng 0.97 < F_oeng I_c I_s > Figure 1. An example of decision tree generated for the Final F_oeng. 5. PROUCIATIO MODELIG AT DECODIG LEVEL Pronunciation modeling at lexical level can only handle intraword pronunciation variations. To deal with inter-word pronunciation variations, some researchers suggested defining a group of multi-words to be added into the lexicon. But, this method can only handle a limited number of inter-word pronunciation variations. Another way to cope with inter-word pronunciation variations is to incorporate PM at decoding level. When incorporating PM at decoding level, it is not necessary to derive the surfaceform pronunciation dictionary. The search works all the way with the baseform lexicon, which is the lexicon built from the baseform transcription. Moreover, pronunciation variations due to inter-word context can also be handled at the decoding level. The decoding process is to find an optimal sequence of words, given the pronunciation lexicon, acoustic model and language model. Decoding algorithms are generally categorized as one-pass versus multi-pass search. In a one-pass search, all knowledge sources are used at a time to decode an utterance, whilst in a multi-pass search, different knowledge sources are applied at different stages during decoding. The ways to incorporate PM at decoding level for one-pass search and multipass search are very different. They will be discussed in Section 5.1 and 5.2 respectively. 5.1. PM in one-pass search In this research, we use a one-pass decoder for continuous Cantonese speech recognition [10]. It works with a treestructured lexicon that is constructed based on the baseform lexicon. The lexical tree specifies all legitimate connections for the baseform bi-if HMMs. Each node in the lexical tree corresponds to a base phone (IF phone), which carries all the bi- IF HMM corresponding to the same base phone. The search algorithm is forward Viterbi search. It is a token-based search process. A token is defined with the identities: node ID, path score and one of the HMM corresponding to the base phone. Bigram language model is applied whenever a search path reaches a word-end node. The most probable word sequence is obtained when the search reaches the end of an utterance. The advantage of integrating PM into one-pass search is that more knowledge sources are added to direct the search process. However, this would enlarge the search space, and require higher computation. Decision tree PMs are obtained in the same way as described in Section 4.2. It should be noted that the right context for an IF model in the search space is not known in the forward Viterbi search. Therefore, we take the current baseform, the baseform and surfaceform left context into account in the prediction of surfaceform IF models. The incorporation of PM in token-based searching process does not change the original search space but only increase the number of alive tokens to carry the information of pronunciation variations. Each bi-if connection is expanded to the predicted surfaceform dynamically during the search. Thus, the path leading to alternative pronunciations is also allowed to propagate in the search process.

HMM: F_aang+I_z HMM: F_aang+I_s Root ode I_k I_h I_c I_z I_s F_aang F_aan C b=f_aang,l b=i_h, L s=i_k Prediction C s=f_aan F_i F_aang SF: F_aang HMM: F_aang+I_z SF: F_aang HMM: F_aang+I_c SF: F_aang HMM: F_aang+I_s Token (no PM) Token (PM) BF SF SF: F_aan HMM: F_aan+I_z SF: F_aan HMM: F_aan+I_c SF: F_aan HMM: F_aan+I_z Figure 2. Token expansion with the incorporation of PM. As shown in Figure 2, without incorporating PM, there are 2 nodes (I_z and I_s) connected to the node F_aang Therefore, 2 bi-if HMMs are stored in this node and 2 tokens are alive in the node F_aang. The incorporation of PM increases the number of alive tokens. By using the decision trees, predictions can be made with the prior knowledge of current baseform context (C b ), left baseform context (L b ) and left surfaceform context (L s ). For example in Figure 2, given the contextual information (C b =F_aang, L b =I_h, L s =I_k), a predicted surfaceform, F_aan, is obtained from the baseform node F_aang. In Figure 2, the nodes I_h, F_aang and I_z have the surfaceforms I_k, F_aan and I_c respectively. Apart from the original tokens carrying the baseform information, additional tokens are created to carry the surfaceform information, e.g. 2 tokens at node F_aang are expanded to 6 tokens. With these additional tokens carrying surfaceform information, each bi-if connection is modified to allow the path propagates to alternative pronunciations in the search process. 5.2. PM in multi-pass search We also attempt to apply pronunciation modeling to a two-pass decoder for Cantonese speech recognition. In this case, PM is used between the two search stages. An IF lattice is generated in stage 1 using only acoustic models. The IFs inside the lattice are in surfaceform. PM is then applied to expand each node in the IF lattice to all the possible baseform IFs that may result in this particular surfaceform. In stage 2, baseform lexicon and language model are applied to search for the most probable word sequence from the expanded IF lattice. The advantages of adding PM into a multi-pass search are its simplicity and ease of manipulation. The modification does not touch the existing searching algorithms in the two stages. It operates on the intermediate results only. Moreover, contextdependency can take into account of the right context which is not available in a one-pass search. The main drawback of a multi-pass search is that the error from each stage would propagate to the next stage. Thus, the performance depends greatly on stage 1. Decision tree PMs are obtained as described in Section 4.2. The difference is that instead of predicting the surfaceforms from a baseform IF, all the possible baseform IFs that would be realized as a particular surfaceform are predicted. The process would therefore be creating a decision tree for each surfaceform IF having baseform IFs in the leaves. 6. EXPERIMETS 6.1. Experiment Setting The methods described above are evaluated in a domainspecific application of continuous Cantonese speech recognition. The application deals with naturally spoken queries on stock information. The test set, STOCKTEST, contains 1300 sentences (about 65 minutes) recorded from 13 speakers. The acoustic models are cross-word bi-ifs trained by 20 hours CUSET corpus. The number of Gaussian mixtures at each state is 16. Each speech frame is represented by a 39 dimensional feature vector with 12 MFCCs and its energy, as well as their first and second order derivatives. In Experiment 1, the search engine is a one-pass search based on tree-structure lexicon [10]. The effectiveness of a decision tree predicted PVD is evaluated. In Experiment 2, the lexicon is the original baseform lexicon but the search engine is the modified one-pass search with the incorporation of PM. In Experiment 3, the search engine is a two-pass search in which IF lattice is generatedinstage1,andwordsequenceisobtainedinstage2. 6.2. Experiment 1 : PM at lexical level Baseline 1 st Tree PVD 2 nd Tree PVD Th_cnt=0 Th_VP=5% 12.06 11.91 11.82 Th_cnt=5 Th_VP=5% 12.06 11.89 11.82 Th_cnt=5 Th_VP=20% 12.06 11.21 11.19 Th_cnt=5 Th_VP=25% 12.06 11.26 11.25 Table 5. WER(%) of using different PVDs at lexical level. TheresultsofExperiment1areshownasinTable5.1 st tree PVD means the PVD built by the first set of decision trees. 2 nd tree PVD is the PVD built by the retrained decision trees. Different thresholds of the frequency count (Th_cnt) and the variation probability (Th_VP) are evaluated. It can be seen that the incorporation of pronunciation modeling in PVD achieves a better performance of recognition. If the threshold is too small, a large number of variations are included in the lexicon. This causes confusion in the searching process and degradation of recognition performance. If the threshold is too large, some frequently pronounced variations would be missed. It is found that the optimal threshold of variation probability is about 20%. The average number of variations per IF unit for this threshold is 1.15. The average number of variations per word is 1.22. By using this threshold, the WER can be reduced by 0.85%. The WER can be further reduced by 0.02% when the retrained PVD is used. The retrained PVD is better than the first tree PVD because the recognized surfaceform is more accurate than the first tree as more information is added. The 7.21% relative error reduction for the retrained tree PVD is due to 9% reduction in number of words substituted and 18% reduction in number of words inserted.

By analyzing the recognition results in detail, we observe that there are three Initials that are always confused. The liprounded velar /gw/ is usually confused with /g/. asal /n/ is always confused with the tongue rolled /l/. asal /ng/ is always deleted. It seems that Cantonese speakers tend not to pronounce a nasal and to round their lips. It is found that pronunciation variations for the Finals occur mainly in codas. asal codas, for example, /ng/, /n/ and /m/ are always confused. Unvoiced stops, for example, /k/, /t/ and /p/ are also always confused. 6.3. Experiment 2 : PM at decoding level using one-pass search Baseline One-pass PM Th_VP=20% 12.06 11.68 Table 6. WER(%) of using PM in one-pass search at decoding level. With the variation probability threshold setting as 20%, the result in Table 6 shows that the incorporation of pronunciation modeling in one-pass search gives a better performance of recognition. WER can be reduced by 0.38%. 3.15% relative error reduction is achieved. 1 st Tree 2 nd Tree One-pass Baseline PVD PVD PM Th_VP=20% 12.06 11.21 11.19 11.68 Table 7. Comparison of WER(%) by using different PMs It is observed in Table 7 that incorporating PM in one-pass search is not as good as using PVD at lexical level and this does not match our expectation. We think that incorporating PM in the decoding level should perform better, since the inter-word variations are better handled. This contradicting result may be due to the fact that we use different decision trees for the two experiments. At the lexical level, the surfaceforms are predicted from the baseform and both the baseform left and right contexts. While at the decoding level, the prediction of surfaceforms depends on the baseform and surfaceform left context. The surfaceform left context (L s ) is obtained from the partial recognition result. The partial recognition result is not perfectly accurate, thus introducing errors in the surfaceform prediction. 1 st Tree 2 nd Tree One-pass PM One-pass PM Baseline PVD PVD with L s without L s 12.06 11.21 11.19 11.68 11.48 Table 8. Comparison of WER(%) of using different PMs (Th_VP=20%). In order to eliminate the error introduced by mis-recognized left surfaceform, we conduct another experiment that uses only the baseform and the left baseform for surfaceform prediction. The result in Table 8 shows that if we only use the baseform and left baseform to build the decision trees, WER can be reduced by 0.2% comparing with that in the last experiment, which gives an overall relative error reduction of 4.81%. This suggests that partial recognition result might not be suitable to surfaceform prediction. evertheless, the result is still not as good as that at the lexical level. This may be due to the fact that the right context is also considered in the surfaceform prediction at the lexical level but not at the decoding level, as the right context is not yet known during the search. The information could be used for surfaceform prediction at the decoding level is therefore less than that at the lexical level. 6.4. Experiment 3 : PM at decoding level using two-pass search Baseline Two-pass PM Th_VP=20% 24.49 23.42 Table 9. WER(%) of using PM in two-pass search at decoding level With the variation probability threshold setting as 20%, the result in Table 9 shows that the incorporation of pronunciation modeling in two-pass search also gives a better performance of recognition. WER can be reduced by 1.07%. 4.37% relative error reduction is achieved. The lattice is expanded by a factor of about 1.5 to contain more pronunciation variations. As stated earlier, one-pass search generally performs better than a two-pass one. This also agrees with our result. The WER for the one-pass search is 11.74% lower than that for two-pass. This experiment is aimed at showing that decision tree PMs also work in a multi-pass decoding process. 7. COCLUSIO This paper describes various approaches of dealing with pronunciation variations in ASR for Cantonese. At the lexical level, a pronunciation variation dictionary is built to obtain alternative pronunciations for each word. Then, the variation probabilities are incorporated in the searching process. This method gives a better recognition performance. The optimal threshold of variation probability is tuned to be 20%. The application of PVD built by the first set of decision trees reduces the WER by 0.85%. The WER can be further reduced by 0.02% when the retrained PVD is used. At the decoding level, decision tree PMs are applied to expand the search space to include alternative pronunciations. In a one-pass search, the search space is dynamically expanded to allow search paths to contain surfaceform phones. Incorporation of pronunciation modeling in one-pass search gives a reduction of WER by 0.38%. In order to eliminate the error introduced by mis-recognized left surfaceform, the left surfacefrom is not used in the construction of decision tree PMs. WER can further be reducedby0.2%.inordertoverifytheapplicabilityofdecision tree PMs in a multi-pass decoding process, experiment using a two-pass search is done in which IF lattice is generated in stage 1. In stage 2, PM is applied to expand the IF lattice to include alternative pronunciations. WER can be reduced by 1.07%.

8. ACKOWLEDGEMET The project is partially supported by a Research Grant for the Hong Kong Research Grant Council (Ref. CUHK 4206/01E). The first author receives a grant from the CUHK Postgraduate Student Grants for Overseas Academic Activities 08/02. The authors would like to give sincere thanks to Mr..W. Wong, Mr. W.K. Lo, Ms. K.. Kwan and Mr. W.. Choi of DSP lab., CUHK, for their help and invaluable advice. 9. REFERECES [1] M.K. Liu et al, Mandarin Accent Adaptation Based on Context-Independent/Context-Dependent Pronunciation Modeling, in Proceedings of ICASSP-00, Vol.2, pp.1025-1028, Istanbul, 2000. [2] C. Huang et al, Accent Modeling Based on Pronunciation Dictionary Adaptation for Large Vocabulary Mandarin Speech Recognition, in Proceedings of ICSLP-00, Vol.3, pp.818-821, Beijing, 2000. [3] M. Saraclar and S. Khudanpur, Pronunciation Ambiguity VS Pronunciation Variability in Speech Recognition, in Proceedings of ICASSP-00, Vol.3, pp.1679-1682, Istanbul, 2000. [4] V. Venkataramani and W. Byrne, MLLR Adaptation Techniques for Pronunciation Modeling, ASRU-01, CDROM, Trento, 2001. [5] M.. Tsai et al, Pronunciation Variation Analysis with respect to Various Linguistic Levels and Contextual Conditions for Mandarin Chinese, in Proceedings of Eurospeech-01, Vol.2, pp.1445-1448, Alborg, 2001. [6] W.K. Lo, Cantonese Phonology and Phonetics: an Engineering Introduction, Internal Document, Speech Processing Laboratory, Department of Electronic Engineering, the Chinese University of Hong Kong, 1999. [7] W.K. Lo, T. Lee and P.C. Ching, Development of Cantonese Spoken Language Corpora For Speech Applications, in Proceedings of ISCSLP-98, pp.102-107, Singapore, 1998. [8] W. Byrne, et al. Automatic Generation of Pronunciation Lexicon for Mandarin Spontaneous Speech, in Proceedings of ICASSP-01, Vol.1, pp.569-572, Salt Lake City, 2001 [9] http://festvox.org/docs/speech_tools-1.2.0/x3475.htm [10] W.. Choi, An Efficient Decoding Method for Continuous Speech Recognition Based on a Tree-Structured Lexicon, Masters thesis, The Chinese University of Hong Kong, 2001.