Data Selection for Statistical Machine Translation

Similar documents
Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

Language Model and Grammar Extraction Variation in Machine Translation

Cross Language Information Retrieval

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

arxiv: v1 [cs.cl] 2 Apr 2017

Noisy SMS Machine Translation in Low-Density Languages

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment

The NICT Translation System for IWSLT 2012

A heuristic framework for pivot-based bilingual dictionary induction

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Australian Journal of Basic and Applied Sciences

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

Prediction of Maximal Projection for Semantic Role Labeling

Parsing of part-of-speech tagged Assamese Texts

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

The KIT-LIMSI Translation System for WMT 2014

Assignment 1: Predicting Amazon Review Ratings

Using dialogue context to improve parsing performance in dialogue systems

Bridging Lexical Gaps between Queries and Questions on Large Online Q&A Collections with Compact Translation Models

On the Combined Behavior of Autonomous Resource Management Agents

Re-evaluating the Role of Bleu in Machine Translation Research

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Cross-lingual Text Fragment Alignment using Divergence from Randomness

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Reducing Features to Improve Bug Prediction

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

METHODS FOR EXTRACTING AND CLASSIFYING PAIRS OF COGNATES AND FALSE FRIENDS

Python Machine Learning

Multi-Lingual Text Leveling

The stages of event extraction

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

Online Updating of Word Representations for Part-of-Speech Tagging

Probabilistic Latent Semantic Analysis

Constructing Parallel Corpus from Movie Subtitles

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Fragment Analysis and Test Case Generation using F- Measure for Adaptive Random Testing and Partitioned Block based Adaptive Random Testing

Identification of Opinion Leaders Using Text Mining Technique in Virtual Community

Greedy Decoding for Statistical Machine Translation in Almost Linear Time

Impact of Controlled Language on Translation Quality and Post-editing in a Statistical Machine Translation Environment

Rule Learning With Negation: Issues Regarding Effectiveness

Lecture 1: Machine Learning Basics

Deep Neural Network Language Models

Overview of the 3rd Workshop on Asian Translation

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Multilingual Document Clustering: an Heuristic Approach Based on Cognate Named Entities

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Linking Task: Identifying authors and book titles in verbose queries

CS Machine Learning

Task Tolerance of MT Output in Integrated Text Processes

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Multimedia Application Effective Support of Education

Improved Reordering for Shallow-n Grammar based Hierarchical Phrase-based Translation

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Alignment of Australian Curriculum Year Levels to the Scope and Sequence of Math-U-See Program

Learning Methods in Multilingual Speech Recognition

Regression for Sentence-Level MT Evaluation with Pseudo References

EXECUTIVE SUMMARY. TIMSS 1999 International Mathematics Report

Syntactic Patterns versus Word Alignment: Extracting Opinion Targets from Online Reviews

Machine Translation on the Medical Domain: The Role of BLEU/NIST and METEOR in a Controlled Vocabulary Setting

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Circuit Simulators: A Revolutionary E-Learning Platform

The RWTH Aachen University English-German and German-English Machine Translation System for WMT 2017

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Yoshida Honmachi, Sakyo-ku, Kyoto, Japan 1 Although the label set contains verb phrases, they

Grade 6: Correlated to AGS Basic Math Skills

Distant Supervised Relation Extraction with Wikipedia and Freebase

The Good Judgment Project: A large scale test of different methods of combining expert predictions

An Efficient Implementation of a New POP Model

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Mandarin Lexical Tone Recognition: The Gating Paradigm

Learning Computational Grammars

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Word Segmentation of Off-line Handwritten Documents

Speech Emotion Recognition Using Support Vector Machine

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

A Case Study: News Classification Based on Term Frequency

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Mathematics process categories

Universiteit Leiden ICT in Business

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Measurement. When Smaller Is Better. Activity:

arxiv: v1 [cs.lg] 3 May 2013

Radius STEM Readiness TM

Memory-based grammatical error correction

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

Rule Learning with Negation: Issues Regarding Effectiveness

Top US Tech Talent for the Top China Tech Company

Transcription:

Data Selection for Statistical Machine Translation eng LIU, Yu ZHOU and Chengqing ZONG National Laboratory of attern Recognition, Institute of Automation, Chinese Academy of Sciences Beijing, China {pliu, yzhou, cqzong}@nlpr.ia.ac.cn Abstract: The bilingual language corpus has a great effect on the performance of a statistical machine translation system. More data will lead to better performance. However, more data also increase the computational load. In this paper, we propose methods to estimate the sentence weight and select more informative sentences from the training corpus and the development corpus based on the sentence weight. The translation system is built and tuned on the compact corpus. The experimental results show that we can obtain a competitive performance with much less data. Keywords: Data selection; corpus optimization; statistical machine translation 1. Introduction Statistical machine translation model heavily relies on bilingual corpus. Typically, the more data is used in the training and tuning processes, the probabilities and parameters we get will be more accurate and lead to a better performance. However, massive data will cost more computational resources. In some specific applications such as the translation system running on a smart mobile phone, the computational resource is limited and a compact and efficient corpus is expected. Normally, we extract probability information from the training data and tune the translation parameters on the development data. These corpora are the most important resources to build an effective translation system. For the training data, we need to know how many sentences are adequate for the translation system. Too many data will increase the computational load and reduce the translation speed. We have to keep balance between the speed and the performance. For the development data, the main problem is how to select the most informative sentences to tune the translation parameters. Typically, we run the minimum error rate training (MERT) on the development data [1]. The MERT will search for the optimal parameters by maximizing the BLEU score. But what kind of sentence pairs is suitable for the MERT is still uncertain. In this paper, we describe approaches to select more 978-1-4244-6899-7/10/$26.00 2010 IEEE informative sentences from the corpora. For the training data, we estimate weight of sentence based on the phrase it contained. The compact training corpus is build according to the sentence weight. For the development data, we select sentences based on surface feature and deep feature, phrase and structure. For both corpora, we verify the relationship between size and translation performance. The remainder of this paper is organized as follows. Related work is presented in Section 2. The data selection methods for training corpus and development corpus are described in Section 3 and Section 4. We give the experimental results in Section 5. At last we come to the conclusions in Section 6. 2. Related work The previous researches on bilingual corpus mainly focused on how to collect more data to construct the translation system. Resnik and Smith mined parallel text on the web [2]; Snover et al. used the comparable texts to improve the translation performance of a MT system [3]. Some researchers did data selection work on the training corpus. Eck et al. selected informative sentences based on n-gram coverage [4]. They used the quantity of previously unseen n-grams to measure the sentence importance. The corpus was sorted according to the sentence importance. But they didn t take the weight of n-gram into account. Lüet al. selected sentences similar to the test text using TF-IDF [5]. The limitation of this method was that the test text must be known first. Yasuda et al. used the perplexity to select the parallel translation pairs from the out-of-domain corpus. They integrated the translation model by linear interpolation [6]. Liu et al. selected sentences for development set according to the phrase weight estimated from the test set, but they have to know the test text first [7]. As mentioned above, most previous work focused on the training data, a little of work focused on the development set. However, we select data for translation system both in training set and development set. High quality sentence pairs are chosen to construct the translation model and tune the translation parameters. And we don t have to know the test text first.

3. Data selection for training data 3.1. Framework In order to keep balance between the performance and the speed, we have to select the more informative sentences from the corpus. First we select a feature from the data and take it as the basic unit. We assign weight to the basic unit according to the information it contained. Second, we estimate the sentence weight based on the basic unit s weight, and we select sentences which can cover more information of the entire original corpus to build a compact corpus. The translation system is built on the compact corpus. The framework of the data selection is shown in Figure 1. Basic Unit Weighting Original Corpus Sentence Weighting Compact Corpus Figure 1. The framework of data selection 3.2. Data selection method In information theory, the information contained in a statement is measured by the negative logarithm of the probability of the statement [8]. And in the phrase-based translation model (BTM), the phrase is the basic translation unit [9]. It is a natural idea to take the phrase as the basic unit. First, we need to estimate the weight of each phrase, and then we estimate the weight of sentence based on those phrases. According to the information theory, the information contained in a phrase should be calculated by formula (1): (1) where is a phrase, and is the probability of the phrase in the corpus. We also take the length of phrase into account to construct the weight because longer phrase always lead to better performance. The weight of phrase is calculated by formula (2): (2) where is the length of the phrase. We use the square root of the length because of the data smoothing. In order to cover more phrases of the original corpus, we assign higher weight to the sentence which has more unseen phrases. The weight of sentence is defined by the following formula: i W 1 (s) = w(f i)e(f i ) (3) where is a sentence, its length is, and is the phrase contained in the sentence. And is defined as follows: when the phrase has occurred in the new corpus, the is 0, otherwise the is 1. If we only consider the new phrases, the longer sentence will tend to get higher score because it contains more unseen phrases. So we divide the score by the sentence length to overcome this problem. This method tends to select sentence which contains more unseen phrase. We also define another formula to estimate the weight of sentence 8 as following: >< i w(f i)e(f i ) W 2 (s) = i >: E(f if i i) E(f i) 6= 0 (5) 0 else (4) In this formula, the sum weights of phrases divided by the total number of unseen phrases. This method tends to select the sentences which contain the rare unseen phrases and ignore the sentence length. 4. Data selection for development data The development corpus is used to tune the translation parameters. Normally, we employ the MERT on the development set to obtain the optimized parameters. But when the development set is in a large scale, the MERT often consumes too long time and too much computer resources until it converges. It is a practical requirement to select appropriate size of development set for the MERT. The scale of the development corpus is often much smaller than the training corpus, so we can extract effective features to measure the sentence weight. An intuitive idea is if the extracted sentences can cover more information of the original development set, the new development set will perform better. In our work, we mainly focus on two features: phrase and structure. 4.1. hrase-based data selection method As mentioned above, the phrase is an important feature for BTM. So we take the phrase as the basic unit and try to cover more phrases. We call this method as phrase-based data selection method (BDS). We take two aspects into account to estimate the weight of phrase: the information it contains and the length of the phrase. The definition of the phrase weight is just as same as the data selection method for training data, see formula (2). And the sentence weight is defined as follows. f2f W (s) = w(f) (6) where f is a phrase and F is the set of all phrases contained in the sentence s. In this method, we estimate the sentence weight by all the phrases contained in the sentence, not just

consider the unseen phrases. We only use the phrase whose length is not longer than four to avoid the data sparseness problem. The new development data is selected according to the scores of the sentences. Higher score sentence contains more unseen phrases. 4.2. Structure-based data selection method Because the BDS method only uses the phrase, a surface feature, to estimate the sentence weight, we also try to use some deep features, such as the sentence structure to select the development set. We name this method as structure-base data selection method (SBDS). In this method, we want to extract sentences which can cover the majority structures of the development set. First we parse the entire development corpus into phrase-structure trees. Then we extract the subtrees contained in the phrase-structure tree, and select sentences which can cover more subtrees. In order to avoid the data sparseness problem, we use the subtree whose depth is between two and four. In order to estimate the weight of the subtrees, we consider two aspects: depth and information. For a subtree, its depth is and its probability is, which is estimated from the development set. The information is calculated by formula (7) and the weight of subtree is estimated by formula (8). (7) (8) Then we can estimate the weight of each sentence as follows: W S (s) = t2t w(t) (9) where is the set of subtrees contained in the sentence, and W S (s) is the score of the sentence estimated by the SBDS method. We can select the sentences according to their scores. 5. Experimental results In our experiments, we use MOSES as our translation engine, and we use the BLEU metrics to evaluate the translation results [10]. 5.1. Results of training data selection We did the training data selection experiments on Chinese-to-English translation task of the CWMT 2008 (China Workshop on Machine Translation) 2 corpus. We randomly extract 20 million words as the original training corpus and we randomly select 400 sentences from the development set as the test set. We tried four methods: a) We select sentences randomly as the 2 http://nlpr-web.ia.ac.cn/cwmt-2008/ baseline; b) We estimate the sentence weight only considering the quantity of the unseen phrases. This method is called unw; c) We consider both the quantity and weight of the unseen phrase, using formula (3), this method is called W1; d) We calculate the weight of sentence using formula (5), this method is called W2. The BLEU score, recall of words and percentage of sentences are presented in Table 1 to Table 3. From the results, it is clear that the data selected using our methods could cover more phrases and get a better performance using small size of data. For example, when we use 12 million words as the training corpus, the baseline only covers 51.1% words, while the unw method could cover 94.4% words, the W1 method could cover 94.6% words and the W2 method could cover 91.5% words. The word coverage is much higher than the baseline. The BLEU score of W2 method is 0.2171, 5.62% higher than the baseline 0.1609, even higher than the system that using all the available data with its score 0.2132. And the corresponding training corpora have almost the same quantity of sentences. The sentence percentages are 55.7%, 55.2%, 55.8% and 60.2%, respectively. We use about 60% data to get a competitive performance compared to using all the data. When we consider the weight of phrase, the system can reach a higher performance, especially when the training data in small size. And the W2 performs better than the W1 method. The W2 method prefers shorter sentences and the phrase coverage is litter lower than the Table 1. BLEU score of translation results Words(M) Baseline unw W1 W2 2 0.1357 0.1614 0.1726 0.1673 4 0.1384 0.1842 0.1918 0.1863 6 0.1468 0.1887 0.1955 0.1896 8 0.1511 0.1947 0.2010 0.2026 10 0.1532 0.2033 0.2060 0.2114 12 0.1609 0.2059 0.2071 0.2171 14 0.1724 0.2055 0.2098 0.2124 16 0.1990 0.2100 0.2118 0.2124 18 0.2095 0.2046 0.2121 0.2127 20 0.2132 0.2132 0.2132 0.2132 Table 2. Recall of words Words(M) Baseline unw W1 W2 2 24.8% 55.7% 67.0% 63.5% 4 30.6% 74.9% 78.7% 75.1% 6 36.8% 83.1% 85.1% 81.0% 8 42.6% 88.3% 89.2% 85.4% 10 45.9% 91.8% 92.3% 88.7% 12 51.1% 94.4% 94.6% 91.5% 14 60.4% 96.3% 96.4% 94.2% 16 82.8% 97.9% 97.9% 96.7% 18 95.7% 99.2% 99.2% 98.8% 20 100.0% 100.0% 100.0% 100.0%

Table 3. ercentage of sentences Words(M) Baseline unw W1 W2 2 9.3% 7.5% 8.2% 11.4% 4 18.7% 16.5% 17.2% 21.7% 6 27.9% 25.7% 26.5% 31.5% 8 37.1% 35.3% 36.0% 41.1% 10 46.5% 45.2% 45.8% 50.7% 12 55.7% 55.2% 55.8% 60.2% 14 65.3% 65.5% 66.0% 69.7% 16 77.7% 76.3% 76.7% 79.3% 18 89.1% 87.3% 87.6% 88.8% 20 100.0% 100.0% 100.0% 100.0% Figure 2. CWMT 2009 Chinese-to-English W1 method. But it could provide more accurate probability information for the translation system and obtain a better performance. 5.2. Results of development data selection We did the data selection experiments for development corpus on CWMT 2009 3 and IWSLT 2009 4 translation tasks, both in bidirectional translation for Chinese and English. The former is in news domain and the latter is in travel domain. For CWMT 2009 task, we randomly select 400 sentences from the development set as the test set, and take the left as the development set. For IWSLT 2009 tasks, we employ BTEC Chinese-to-English task and Challenge English-to-Chinese task. The Table 4 shows the information of the corpora. Table 4. Corpus for development data selection Task Development set Test set Sen Words Sen CWMT 2009 C-E 2,876 57,010 400 E-C 3,081 55,815 400 IWSLT 2009 C-E 2,508 17,940 469 E-C 1,465 12,210 393 On each task, we select sentences randomly to build the baseline. Then we selected the different scale of development data for the MERT using the approaches we proposed. For the BDS method, we consider the phrases from the Chinese sentences (Ch), the English sentences (En) and both of them (Ch+En). For the SBDS method, we only use the Chinese sentences and parse them using the Stanford parser [11]. The results are shown in Figure 2 to Figure 5. In these figures, the horizontal axis is the scale of the development corpus, the unit is thousand words. The vertical axis is the BLEU score of the test set using the parameters trained on the corresponding development 3 http://www.icip.org.cn/cwmt2009 4 http://mastarpj.nict.go.jp/iwslt2009/ Figure 3. CWMT 2009 English-to-Chinese Figure 4. IWSLT 2009 Chinese-to-English Figure 5. IWSLT 2009 English-to-Chinese data. Comparing to the baseline system, the development corpus selected using our methods can get higher performance with the same quantity of data. Our method can select more informative sentences for MERT. For the BDS, when we consider both the Chinese phrase and the English phrase, the performance is better and more

robust comparing to the methods which only consider monolingual phrase. This is because the sentences extracted using this method could cover the information both in source language and target language, and make the translation parameters more robust. The SBDS performs not as good as the BDS, though it is better than the baseline. This is because that the precision of the parser is not good enough. The parser will import many errors into the parsing results and decrease the performance of the translation system. For this reason, we didn t combine these two methods. syntax structures; the other reason is the translation engine is phrase-based translation system; it could not make full use of the information contained in the parsing results. Acknowledgements The research work has been partially funded by the Natural Science Foundation of China under Grant No. 60975053, the National Key Technology R&D rogram under Grant No. 2006BAH03B02, and also supported by the China-Singapore Institute of Digital Media (CSIDM) project under grant No. CSIDM-200804. References Figure 6. Recall of words for IWSLT 2009 E-C Another interesting phenomenon is that we can get even higher score using a part of the development data than using all the data. For example, in Figure 5, when we using 10 thousand words for MERT, the performance is better than using 12 thousand words. We present the recall of words for the baseline method and the BDS method which considers bilingual phrases in Figure 6. From this figure, the baseline s recall is only 77.0% while the BDS s recall is 99.9% when the development data has 10 thousand words; almost all the words have been covered. Adding more data to the development set brings little improvement to the recall of words, but imports much redundancy sentences and reduces the performance of the translations. 6. Conclusions In this paper, we propose approaches to select more informative sentences from the bilingual corpus. For the training corpus, we select sentences to build a compact training corpus using two kinds of weighted-phrase method. For the new compact training corpus, we can get a competitive performance compared to the baseline system using all training data. The data selection for development corpus using two kinds of features: the phrase and the structure. Both methods perform better than the baseline. When consider the bilingual phrases, the performance is better and more robust. The BDS is better than the SBDS. One reason is that the parser could import errors to the phrase-structure tree and there exists serious data sparseness problem in [1] F. J. Och, "Minimum Error Rate Training in Statistical Machine Translation," in roc. of the 41st ACL, Sapporo, 2003, pp. 160-167. [2]. Resnik and N. A. Smith, "The Web as a arallel Corpus," Computational Linguistics, vol. 29, pp. 349-380, 2003. [3] M. Snover, B. Dorr, and R. Schwartz, "Language and Translation Model Adaptation using Comparable Corpora," in roc. of the EMNL 2008, Honolulu, Hawaii, 2008, pp. 857-866. [4] M. Eck, S. Vogel, and A. Waibel, "Low Cost ortability for Statistical Machine Translation based on N-gram Coverage," in roc. of the 10th MT Summit, huket, Thailand, 2005, pp. 227-234. [5] Y. Lü, J. Huang, and Q. Liu, "Improving Statistical Machine Translation erformance by Training Data Selection and Optimization," in roc. of the EMNL-CoNLL 2007, rague, Czech Republic, 2007, pp. 343-350. [6] K. Yasuda, R. Zhang, H. Yamamoto, and E. Sumita, "Method of Selecting Training Data to Build a Compact and Efficient Translation Model," in roc. of the 3 rd IJCNL, Hyderabad, India, 2008. [7]. Liu, Y. Zhou, and C. Zong, "Approach to Selecting Best Development Set for hrase-based Statistical Machine Translation," in roc. of the 23rd ACLIC, Hongkong, 2009. [8] T. M. Cover and J. A. Thomas, Elements of Information Theory. New York: Wiley, 1991. [9]. Koehn, F. J. Och, and D. Marcu, "Statistical hrase-based Translation," in roc. of the NAACL 2003, Edmonton, Canada, 2003. [10] K. apineni, S. Roukos, T. Ward, and W.-J. Zhu, "Bleu: a Method for Automatic Evaluation of Machine Translation," in roc. of 40th ACL, hiladelphia, ennsylvania, USA, 2002, pp. 311-318. [11] R. Levy and C. D. Manning, "Is it Harder to arse Chinese, or the Chinese Treebank?," in roc. of the 41st ACL, Sapporo, Japan, 2003.