The UKA/CMU translation system for IWSLT 2006

Similar documents
The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment

Noisy SMS Machine Translation in Low-Density Languages

Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

Language Model and Grammar Extraction Variation in Machine Translation

Speech Recognition at ICSI: Broadcast News and beyond

Japanese Language Course 2017/18

The KIT-LIMSI Translation System for WMT 2014

3 Character-based KJ Translation

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation

Clickthrough-Based Translation Models for Web Search: from Word Models to Phrase Models

Greedy Decoding for Statistical Machine Translation in Almost Linear Time

Calibration of Confidence Measures in Speech Recognition

Improvements to the Pruning Behavior of DNN Acoustic Models

Re-evaluating the Role of Bleu in Machine Translation Research

JAPELAS: Supporting Japanese Polite Expressions Learning Using PDA(s) Towards Ubiquitous Learning

The 2014 KIT IWSLT Speech-to-Text Systems for English, German and Italian

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

The NICT Translation System for IWSLT 2012

Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

What is the status of task repetition in English oral communication

Improved Reordering for Shallow-n Grammar based Hierarchical Phrase-based Translation

arxiv: v1 [cs.cl] 2 Apr 2017

Cross Language Information Retrieval

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

Teaching intellectual property (IP) English creatively

Assignment 1: Predicting Amazon Review Ratings

Learning Methods in Multilingual Speech Recognition

Speech Translation for Triage of Emergency Phonecalls in Minority Languages

Investigation on Mandarin Broadcast News Speech Recognition

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Cross-lingual Text Fragment Alignment using Divergence from Randomness

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

The Interplay of Text Cohesion and L2 Reading Proficiency in Different Levels of Text Comprehension Among EFL Readers

Modeling function word errors in DNN-HMM based LVCSR systems

Overview of the 3rd Workshop on Asian Translation

Modeling function word errors in DNN-HMM based LVCSR systems

Detecting English-French Cognates Using Orthographic Edit Distance

CS Machine Learning

Regression for Sentence-Level MT Evaluation with Pseudo References

Add -reru to the negative base, that is to the "-a" syllable of any Godan Verb. e.g. becomes becomes

Segmental Conditional Random Fields with Deep Neural Networks as Acoustic Models for First-Pass Word Recognition

Challenging Assumptions

Probabilistic Latent Semantic Analysis

Constructing Parallel Corpus from Movie Subtitles

Deep Neural Network Language Models

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Impact of Controlled Language on Translation Quality and Post-editing in a Statistical Machine Translation Environment

PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Performance Analysis of Optimized Content Extraction for Cyrillic Mongolian Learning Text Materials in the Database

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

My Japanese Coach: Lesson I, Basic Words

Adding Japanese language synthesis support to the espeak system

Multi-Lingual Text Leveling

The stages of event extraction

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

Switchboard Language Model Improvement with Conversational Data from Gigaword

The Smart/Empire TIPSTER IR System

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Carnegie Mellon University Department of Computer Science /615 - Database Applications C. Faloutsos & A. Pavlo, Spring 2014.

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Lecture 1: Machine Learning Basics

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

(Sub)Gradient Descent

SPEECH RECOGNITION CHALLENGE IN THE WILD: ARABIC MGB-3

THE PERCEPTIONS OF THE JAPANESE IMPERFECTIVE ASPECT MARKER TEIRU AMONG NATIVE SPEAKERS AND L2 LEARNERS OF JAPANESE

A Quantitative Method for Machine Translation Evaluation

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH

Unsupervised Acoustic Model Training for Simultaneous Lecture Translation in Incremental and Batch Mode

Mandarin Lexical Tone Recognition: The Gating Paradigm

A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Fluency is a largely ignored area of study in the years leading up to university entrance

My First Spanish Phrases (Speak Another Language!) By Jill Kalz

A Case Study: News Classification Based on Term Frequency

Machine Translation on the Medical Domain: The Role of BLEU/NIST and METEOR in a Controlled Vocabulary Setting

arxiv: v1 [cs.lg] 7 Apr 2015

Small-Vocabulary Speech Recognition for Resource- Scarce Languages

Language Independent Passage Retrieval for Question Answering

Problems of the Arabic OCR: New Attitudes

The RWTH Aachen University English-German and German-English Machine Translation System for WMT 2017

Statewide Framework Document for:

Speech Recognition using Acoustic Landmarks and Binary Phonetic Feature Classifiers

Atypical Prosodic Structure as an Indicator of Reading Level and Text Difficulty

<September 2017 and April 2018 Admission>

A Neural Network GUI Tested on Text-To-Phoneme Mapping

MOUNT LAWLEY SENIOR HIGH SCHOOL SENIOR SCHOOL YEAR

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Disambiguation of Thai Personal Name from Online News Articles

INVESTIGATION OF UNSUPERVISED ADAPTATION OF DNN ACOUSTIC MODELS WITH FILTER BANK INPUT

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

WHEN THERE IS A mismatch between the acoustic

Transcription:

The UKA/CMU translation system for IWSLT 2006 Matthias Eck, Ian Lane, Nguyen Bach, Sanjika Hewavitharana, Muntsin Kolss, Bing Zhao, Almut Silja Hildebrand, Stephan Vogel, and Alex Waibel InterACT Research Laboratories: University of Karlsruhe, Karlsruhe, Germany Carnegie Mellon University, Pittsburgh, USA

Overview SMT System components Phrase Alignment Models PESA Log-Linear Phrase Alignment (LogLin) Language Model Decoder Experimental Results Analysis Conclusions

PESA Alignment Given source phrase source sentence target sentence

PESA Alignment What is the translation of the source phrase? source sentence target sentence

PESA Alignment Back to IBM-1 probabilities source sentence target sentence

PESA Alignment Probability for this split: f J f j2 f j1 f 1 e 1 e i1 e i2 e I

PESA Alignment Probability for this split: f J j 2 ( j= j i= i 1 i 2 1 p( f j e i )) f j2 Inside Alignment Probability f j1 f 1 e 1 e i1 e i2 e I

Word Alignment Matrix Probability for this split: f J j 2 i ( j= j i= i 1 2 1 p( f j e i )) f j2 * * j 1 1 ( j= 1 i ( i... i ) J ( p( f j= j + 1 i ( i... i ) 1 2 2 1 2 j p( f e)) j i e)) i f j1 f 1 e 1 e i1 e i2 e I Outside Alignment Probability

PESA Alignment Optimize over target boundaries to find optimal split Look from both directions p f ) pe i f ) ( j ei ( j Online phrase extraction Phrases are extracted as needed during decoding process No restriction on phrase length

LogLin Alignment General idea: LogLin extends idea of PESA by adding multiple features e.g. - Word alignment - Fertility (Phrase length) - Relative position in sentence pair - Lexical features (IBM-1) Some feature functions might overlap Framework of Log Linear Model is applied

LogLin Alignment Log-linear model to combine more feature functions Feature functions Model parameters: weights of the feature functions (e,f) is a sentence pair X is a phrase pair extracted from (e,f)

LogLin Alignment 2 Step approach: 1. Find candidates using simple heuristics 2. Score candidates using feature functions

LogLin Alignment Find projected center of target phrase For each source word: Find center of gravity of IBM1 probabilities Projected center for this source word

LogLin Alignment Find projected center of target phrase For each source word: Find center of gravity of IBM1 probabilities Projected center for this source word

LogLin Alignment Find projected center of target phrase For each source word: Find center of gravity of IBM1 probabilities Projected center for this source word

LogLin Alignment Find projected center of target phrase For each source word: Find center of gravity of IBM1 probabilities Projected center for this source word

LogLin Alignment Find projected center of target phrase Average of centers to get projected target center for source phrase

LogLin Alignment Predict target length using IBM-4 fertilities

LogLin Alignment Predict target length using IBM-4 fertilities

LogLin Alignment Predict target length using IBM-4 fertilities Generate candidates using the predictions for center and target length Target phrase does not have to have the projected center in the middle but it has to contain it First step generates a (relatively small) number of phrase translation candidates

13 Features for candidate scoring 4: Phrase-level length relevance Source phrase generates target phrase of this length Rest of sentence generates Rest of sentence of this length + reverse direction 4: IBM Model-1 scores similar to PESA Source phrase generates target phrase Rest of sentence generates Rest of sentence + reverse direction

13 Features for candidate scoring 4: Bracket the sentence pair diagonal and inverse diagonal (both directions) f J f J f j2 f j2 f j1 f j1 f 1 e 1 e i1 f 1 e i2 e I e 1 e i1 e i2 e I 1: average alignment links per source word Every block should contain at least one word alignment from the Viterbi path

Feature weights Weights for each feature function are learned using human aligned gold standard phrase pairs Weights are adjusted to optimize accuracy on these phrases Problems: For BTEC data no human word alignment available to extract gold-standard phrase pairs Used previously trained weights (Chinese English newswire data) Should work reasonably well on Chinese BTEC Questionable on other language pairs Overfitting possible due to overlapping features

Language Model 2 Options: 3-gram SRI language model (Kneser-Ney discounting) 6-gram Suffix Array language model (Good-Turing discounting) 6-gram consistently gave better results Only used 6-gram LM 24.0 23.0 3gram 6gram BLEU 22.0 21.0 20.0 Supplied Data Supplied Data + Free Data Data Full BTEC Full BTEC + any data

Decoding 2 stage decoding process Build translation lattice using the extracted phrase pairs Search for best path through lattice Word reordering possible within reordering window (best results at ~4-5) ASR output translation: Only translated 1best

Italian English results Open Track - 20k lines supplied data C-STAR Track - 55k lines Full BTEC - 3k lines web data (travel phrases) Open Track C-STAR Track BLEU NIST BLEU NIST PESA 0.2388 6.20 0.2630 6.66 LogLin 0.2719 6.61 0.2912 7.08

Arabic English results Open Track - 20k lines supplied data C-STAR Track - 20k lines supplied data - 20k lines additional translated BTEC - 31k lines typed travel books (English) Open Track C-STAR Track BLEU NIST BLEU NIST PESA 0.1908 5.38 0.1989 5.62 LogLin 0.1995 5.34 0.2123 5.87

Chinese English results Open Track - 40k lines supplied data Open Track C-STAR Track - 163k lines Full BTEC - 106k lines newswire data (gathered with IR technique) - 31k lines typed travel books (English) C-STAR Track PESA LogLin BLEU NIST BLEU NIST read 0.1501 4.87 0.1622 5.19 spont 0.1654 5.08 0.1645 5.24 read 0.1630 4.97 - - spont 0.1710 5.08 - -

Japanese English results Open Track - 40k lines supplied data C-STAR Track - 163k lines Full BTEC - 4k medical dialogs Open Track C-STAR Track BLEU NIST BLEU NIST PESA 0.1868 5.63 0.1841 5.40 LogLin 0.1830 5.93 - -

Chinese English Influence of additional data tested with PESA alignment: Supplied Data Supplied Data + IR data spont. 0.1393 0.1501 +7.8% read 0.1539 0.1654 +7.5% Full BTEC Full BTEC + IR data + travel books spont. 0.1388 0.1622 +16.9% read 0.1439 0.1645 +14.9%

Analysis Chinese and Japanese: No improvements Open Data Track C-STAR Data track Alignment problem with Full BTEC data for Chinese - English Word segmentation problems: Provided segmentation could not be used for the C-STAR Data track Re-segmentation was necessary Worse word segmentation quality especially on ASR output

Word segmentation - Japanese Provided-segmentation ASR: 御荷物はに持つ引き取りとにございます (3-errors) REF: 御荷物は荷物引き取り所にございます 3-ASR errors 3 segmentation errors MeCab-segmentation (used on C-STAR track) ASR: 御荷物はに持つ引き取りとにございます (5-errors) REF: 御荷物は荷物引き取り所にございます 3-ASR errors 5 segmentation errors BLEU (% degradation) Word Segmentation Provided MeCab Transcriptions ASR Output 24.3 21.1 (13%) 23.5 19.6 (17%)

Analysis: Phrase alignments LogLin outperforms PESA on Chinese, Arabic Best improvements on Italian (+0.03 BLEU) Slight drop on Japanese BLEU 0.28 0.26 0.24 0.22 0.20 0.18 0.16 PESA LogLin 0.14 Arabic Italian Chinese Japanese Source Language

Analysis BLEU - WER CRR ASR (read) WER Correlation BLEU degradation CRR ASR with WER of ASR output Japanese Arabic Chinese 0.2030 0.2208 0.1996 0.1868 0.1995 0.1710-8.0% -9.6% -14.3% 14.9% 26.1% 26.4% Italian 0.3353 0.2719-18.9% 29.1% relative BLEU loss (%) 20 15 10 5 Japanese Chinese Arabic Italian 0 10 15 20 25 30 35 word error rate (%)

Future Work Use lattice/nbest information for translation of ASR output Provide LogLin with better hand-aligned data (in-domain) in different languages Limit influence of overfitting