Language Modeling in the Era of Abundant Data

Similar documents
Deep Neural Network Language Models

Investigation on Mandarin Broadcast News Speech Recognition

Calibration of Confidence Measures in Speech Recognition

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

Lecture 1: Machine Learning Basics

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Improvements to the Pruning Behavior of DNN Acoustic Models

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

Noisy SMS Machine Translation in Low-Density Languages

COPING WITH LANGUAGE DATA SPARSITY: SEMANTIC HEAD MAPPING OF COMPOUND WORDS

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Learning Methods in Multilingual Speech Recognition

arxiv: v1 [cs.cl] 27 Apr 2016

Multi-Lingual Text Leveling

Segmental Conditional Random Fields with Deep Neural Networks as Acoustic Models for First-Pass Word Recognition

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition

Modeling function word errors in DNN-HMM based LVCSR systems

Speech Recognition at ICSI: Broadcast News and beyond

Modeling function word errors in DNN-HMM based LVCSR systems

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH

Detecting English-French Cognates Using Orthographic Edit Distance

A study of speaker adaptation for DNN-based speech synthesis

Lecture 9: Speech Recognition

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition

Toward a Unified Approach to Statistical Language Modeling for Chinese

Greedy Decoding for Statistical Machine Translation in Almost Linear Time

Mandarin Lexical Tone Recognition: The Gating Paradigm

Reducing Features to Improve Bug Prediction

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

CS Machine Learning

Abbreviated text input. The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters.

Language Model and Grammar Extraction Variation in Machine Translation

The A2iA Multi-lingual Text Recognition System at the second Maurdor Evaluation

Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

Bridging Lexical Gaps between Queries and Questions on Large Online Q&A Collections with Compact Translation Models

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

Using dialogue context to improve parsing performance in dialogue systems

Letter-based speech synthesis

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

A Quantitative Method for Machine Translation Evaluation

Probabilistic Latent Semantic Analysis

Speech Emotion Recognition Using Support Vector Machine

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Dropout improves Recurrent Neural Networks for Handwriting Recognition

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

arxiv:cmp-lg/ v1 7 Jun 1997 Abstract

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Using Synonyms for Author Recognition

The NICT/ATR speech synthesis system for the Blizzard Challenge 2008

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation

Introduction on Lean, six sigma and Lean game. Remco Paulussen, Statistics Netherlands Anne S. Trolie, Statistics Norway

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Malicious User Suppression for Cooperative Spectrum Sensing in Cognitive Radio Networks using Dixon s Outlier Detection Method

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

Evolutive Neural Net Fuzzy Filtering: Basic Description

Improved Hindi Broadcast ASR by Adapting the Language Model and Pronunciation Model Using A Priori Syntactic and Morphophonemic Knowledge

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

MYCIN. The MYCIN Task

Atypical Prosodic Structure as an Indicator of Reading Level and Text Difficulty

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

University of Waterloo School of Accountancy. AFM 102: Introductory Management Accounting. Fall Term 2004: Section 4

Agent-Based Software Engineering

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Clickthrough-Based Translation Models for Web Search: from Word Models to Phrase Models

The NICT Translation System for IWSLT 2012

Massachusetts Institute of Technology Tel: Massachusetts Avenue Room 32-D558 MA 02139

Plattsburgh City School District SIP Building Goals

Dialog Act Classification Using N-Gram Algorithms

WiggleWorks Software Manual PDF0049 (PDF) Houghton Mifflin Harcourt Publishing Company

arxiv: v1 [cs.cl] 2 Apr 2017

Unsupervised Acoustic Model Training for Simultaneous Lecture Translation in Incremental and Batch Mode

Pair Programming: When and Why it Works

Theory of Probability

cmp-lg/ Jan 1998

Georgetown University at TREC 2017 Dynamic Domain Track

Data Fusion Models in WSNs: Comparison and Analysis

Residual Stacking of RNNs for Neural Machine Translation

An Online Handwriting Recognition System For Turkish

Semi-Supervised Face Detection

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

Re-evaluating the Role of Bleu in Machine Translation Research

Naviance / Family Connection

WHEN THERE IS A mismatch between the acoustic

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

SEMI-SUPERVISED ENSEMBLE DNN ACOUSTIC MODEL TRAINING

Online Updating of Word Representations for Part-of-Speech Tagging

The Good Judgment Project: A large scale test of different methods of combining expert predictions

A Comparison of Two Text Representations for Sentiment Analysis

On the Formation of Phoneme Categories in DNN Acoustic Models

Transcription:

Language Modeling in the Era of Abundant Data Ciprian Chelba Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 1

Statistical Modeling in Automatic Speech Recognition Speaker s Speech Acoustic Linguistic Mind W Producer Speech Processor A Decoder W^ Speaker Acoustic Channel Speech Recognizer Ŵ = argmax W P(W A) = argmax W P(A W) P(W) P(A W) acoustic model (AM, Hidden Markov Model); varies depending on problem (machine translation, spelling correction, soft keyboard input) P(W) language model (LM, usually Markov chain) search for the most likely word string Ŵ Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 2

Language Modeling Usual Assumptions we have a word level tokenization of the text (not true in all languages, e.g. Chinese) some vocabulary is given to us (usually also estimated from data); out-of-vocabulary (OoV) words are mapped to <UNK> ( open vocabulary LM) sentences are assumed to be independent and of finite length; LM needs to predict end-of-sentence symbol </S> On my second day, I managed the uphill walk to a waterfall called <UNK> Skok. </S> Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 3

Language Model Evaluation (1) Word Error Rate (WER) TRN: UP UPSTATE NEW YORK SOMEWHERE UH OVER HYP: UPSTATE NEW YORK SOMEWHERE UH ALL ALL D 0 0 0 0 0 I S :3 errors/7 words in transcript; WER = 43% Perplexity (PPL) ( (Jelinek, 1997) ) PPL(M) = exp 1 N N i=1 ln[p M(w i w 1...w i 1 )] good models are smoothed ML estimates: P M (w i w 1...w i 1 ) > ǫ; also guarantees a proper probability model over sentences other metrics: out-of-vocabulary rate/n-gram hit ratios Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 4

Language Model Smoothing Markov assumption leads to N-gram model: P θ (w i w 1...w i 1 ) = P θ (w i w i N+1...w i 1 ),θ Θ,w i V Smoothing using Deleted Interpolation: where: P n (w h) = λ(h) P n 1 (w h )+(1 λ(h)) f n (w h) P 1 (w) = uniform(v) h = (w i n+1...w i 1 ) is the n-gram context, and h = (w i n+2...w i 1 ) is the back-off context weights λ(h) must be estimated on held-out (cross-validation) data. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 5

Language Model Smoothing: Katz Katz Smoothing (Katz, 1987) uses Good-Turing discounting: P n (w h) = where: f n (w h),c(h,w) > K (r +1) t r+1 t r f n (w h),0 < C(h,w) K β(h)p n 1 (w h ) t r represents the number of n-grams (types) that occur r times: t r = (w i n+1...w i ),C(w i n+1...w i ) = r β(h) is the back-off weight ensuring proper normalization Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 6

Language Model Smoothing: Kneser-Ney Kneser-Ney Smoothing (Kneser & Ney, 1995): P n (w h) = { C(h,w) D1 C(h) LeftDivC(h,w) D 2 +λ(h)p n 1 (w h ),n = N w LeftDivC(h,w) +λ(h)p n 1(w h ),0 n < N where: LeftDivC(h,w) = v,c(v,h,w) > 0 is the left diversity count for an n-gram (h, w) See (Goodman, 2001) for a detailed presentation on LM smoothing. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 7

Language Model Representation: ARPA Back-off p(wd3 wd1,wd2)= if(trigram exists) p_3(wd1,wd2,wd3) else if(w1,w2 exists) bo_2(w1,w2)*p(wd3 wd2) else p(wd3 w2) p(wd2 wd1)= if(w1,w2 exists) p_2(wd1,wd2) else bo_1(wd1)*p_1(wd2) \1-grams: p_1 wd bo_1 \2-grams: p_2 wd1 wd2 bo_2 \3-grams: p_3 wd1 wd2 wd3 Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 8

Language Model Size Control: Entropy Pruning Entropy pruning (Stolcke, 1998) is required for use in 1st pass: should one remove n-gram (h, w)? D[q(h)p( h) q(h) p ( h)] = q(h) w p(w h)log p(w h) p (w h) D[q(h)p( h) q(h) p ( h)] < pruning threshold lower order estimates: q(h) = p(h 1 )...p(h n h 1...h n 1 ) or relative frequency: q(h) = f(h) greedily reduces LM size at min cost in PPL Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 9

On Smoothing and Pruning 8.4 8.2 Perplexity Increase with Pruned LM Size Katz (Good Turing) Kneser Ney Interpolated Kneser Ney 8 7.8 PPL (log2) 7.6 7.4 7.2 7 6.8 18 19 20 21 22 23 24 25 Model Size in Number of N grams (log2) KN degrades very fast with aggressive pruning (< 10% of original size) (Ciprian Chelba, 2010) switch from KN to Katz smoothing: 10% WER gain for voice-search Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 10

Voice Search LM Training Setup (Chelba & Schalkwyk, 2013) spelling corrected google.com queries, normalized for ASR, e.g. 5th -> fifth vocabulary size: 1M words, OoV rate 0.57% (!), excellent n-gram hit ratios training data: 230B words Order no. n-grams pruning PPL n-gram hit-ratios 3 15M entropy 190 47/93/100 3 7.7B none 132 97/99/100 5 12.7B 1-1-2-2-2 108 77/88/97/99/100 Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 11

Is Bigger Better? YES! 260 Perplexity (left) and Word Error Rate (right) as a function of LM size 20.5 240 20 220 19.5 200 19 180 18.5 160 18 140 17.5 120 10 3 10 2 10 1 10 0 10 117 LM size: # n grams(b, log scale) PPL is really well correlated with WER when controlling for vocabulary and training set. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 12

Better Language Models: More Smarts 1-billion word benchmark (Chelba et al., 2013) results Model Num. Params PPL Katz 5-gram 1.74 B 79.9 Kneser-Ney 5-gram 1.76 B 67.6 SNM skip-gram 33 B 52.9 RNN 20 B 51.3 ALL, linear interpolation 41.0 there are LMs that handily beat the N-gram by leveraging longer context (when available) how about increasing the amount of data, when we have it? Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 13

Better Language Models: More Smarts, More Data? Ideally Both 10/100 billion word query data benchmark results a Model Data Amount Num. Params PPL Katz 6-gram 10B 3.2 B 123.9 Kneser-Ney 6-gram 10B 4.1 B 114.5 SNM skip-gram 10B 25 B 111.0 RNN 10B 4.1 B 111.1 Katz 6-gram 100B 19.6 B 92.7 Kneser-Ney 6-gram 100B 24.5 B 87.9 RNN 100B 4.1 B 101.0 more data and model is an easy way to get solid gains complex models better scale up gracefully KN smoothing loses its edge over Katz a Thanks Babak Damavandi Ciprian Chelba, for Language themodeling RNNin thexperimental Era of Abundant Data, Information results. Theory Forum, Stanford, 01/09/2015 p. 14

More Data Is Not Always a Winner: Query Stream Non-stationarity (1) USA training data: XX months X months test data: 10k, Sept-Dec 2008 very little impact in OoV rate for 1M wds vocabulary: 0.77% (X months vocabulary) vs. 0.73% (XX months vocabulary) Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 15

More Data Is Not Always a Winner: Query Stream Non-stationarity (2) 3-gram LM Training Set Test Set PPL unpruned X months 121 unpruned XX months 132 entropy pruned X months 205 entropy pruned XX months 209 bigger is not always better a 10% rel reduction in PPL when using the most recent X months instead of XX months no significant difference after pruning, in either PPL or WER a The vocabularies are mismatched, so the PPL comparison is troublesome. The difference would be higher if we used a fixed vocabulary. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 16

More Locales training data across 3 locales: USA, GBR, AUS, spanning same amount of time ending in Aug 2008 test data: 10k/locale, Sept-Dec 2008 Out of Vocabulary Rate: Training Test Locale Locale USA GBR AUS USA 0.7 1.3 1.6 GBR 1.3 0.7 1.3 AUS 1.3 1.1 0.7 locale specific vocabulary halves the OoV rate Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 17

Locale Matters (2) Perplexity of unpruned LM: Training Test Locale Locale USA GBR AUS USA 132 234 251 GBR 260 110 224 AUS 276 210 124 locale specific LM halves the PPL of the unpruned LM Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 18

Open Problems Entropy of text from a given source: how much are we leaving on the table? How much data/model is enough for a given source: does such a bound exist for N-gram models? More data, relevance, transfer learning: not all data is created equal. Conditional ML estimation: LM estimation should take into account the channel model. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 19

Entropy of English High variance, depending on estimate, source of data; 0.1-0.2 bits/char is a significant difference in PPL at word level! (Cover & King, 1978): 1.3 bits/char (Brown, Pietra, Mercer, Pietra, & Lai, 1992): 1.75 bits/char 1-billion corpus: a 1.17 bits/char for KN, 1.03 bits/char for the best reported LM mixing skip-gram SNM with RNN 10, 100 -billion query corpus: 1.43, 1.35 bits/char for KN, respectively. a Modulo OoV word modeling Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 20

Abundant Data: How Much is Enough for Modeling a Given Source? A couple of observations: one can prune an LM to about 10% of unpruned size without significant impact on PPL increasing the amount of data and model size becomes unproductive after a while For a given source, and N-gram order, is there a data size beyond which there is no benefit to the model quality? Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 21

Abundant Data: Not All Data is Created Equal It is not always possible to find very large amounts of data that is well matched to a given application/test set E.g. when building an LM for SMS text we may have very little such data, quite a bit more from posts on social networks, and a lot of text from a web crawl. LM adaptation: leveraging data in different amounts, and of various degrees of relevance a to a given test set. a Relevance of data to a given test set is hard to describe, but you know it when you see it. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 22

References Brown, P. F., Pietra, V. J. D., Mercer, R. L., Pietra, S. A. D., & Lai, J. C. (1992, March). An estimate of an upper bound for the entropy of english. Comput. Linguist., 18(1), 31 40. Available from http://dl.acm.org/citation.cfm?id=146680.146685 Chelba, C., Mikolov, T., Schuster, M., Ge, Q., Brants, T., Koehn, P., et al. (2013). One billion word benchmark for measuring progress in statistical language modeling. Chelba, C., & Schalkwyk, J. (2013). Empirical exploration of language modeling for the google.com query stream as applied to mobile voice search. In Mobile speech and advanced natural language solutions (pp. 197 229). New York: Springer. Available from http://www.springer.com/engineering/signals/book/978-1-4614-6017-6 Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 23

References Ciprian Chelba, Will Neveitt, Peng Xu, Thorsten Brants. (2010). Study on Interaction between Entropy Pruning and Kneser-Ney Smoothing. In Proc. interspeech (pp. 2242 2245). Makuhari, Japan. Cover, T., & King, R. (1978, September). A convergent gambling estimate of the entropy of english. IEEE Trans. Inf. Theor., 24(4), 413 421. Available from http://dx.doi.org/10.1109/tit.1978.1055912 Goodman, J. (2001). A bit of progress in language modeling, extended version (Tech. Rep.). Microsoft Research. Jelinek, F. (1997). Statistical methods for speech recognition. Cambridge, MA, USA: MIT Press. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 24

References Katz, S. (1987). Estimation of probabilities from sparse data for the language model component of a speech recognizer. In Ieee transactions on acoustics, speech and signal processing (Vol. 35, p. 400-01). Kneser, R., & Ney, H. (1995). Improved backing-off for m-gram language modeling. In Proceedings of the ieee international conference on acoustics, speech and signal processing (Vol. 1, pp. 181 184). Stolcke, A. (1998). Entropy-based pruning of back-off language models. In Proceedings of news transcription and understanding workshop (pp. 270 274). Lansdowne, VA: DARPA. Ciprian Chelba, Language Modeling in the Era of Abundant Data, Information Theory Forum, Stanford, 01/09/2015 p. 25