Improving SMT for Baltic Languages with Factored Models

Similar documents
Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

arxiv: v1 [cs.cl] 2 Apr 2017

Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

Noisy SMS Machine Translation in Low-Density Languages

Cross Language Information Retrieval

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

Constructing Parallel Corpus from Movie Subtitles

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

The NICT Translation System for IWSLT 2012

Language Model and Grammar Extraction Variation in Machine Translation

The KIT-LIMSI Translation System for WMT 2014

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Training and evaluation of POS taggers on the French MULTITAG corpus

A heuristic framework for pivot-based bilingual dictionary induction

Re-evaluating the Role of Bleu in Machine Translation Research

Impact of Controlled Language on Translation Quality and Post-editing in a Statistical Machine Translation Environment

Parsing of part-of-speech tagged Assamese Texts

Linking Task: Identifying authors and book titles in verbose queries

What the National Curriculum requires in reading at Y5 and Y6

Initial approaches on Cross-Lingual Information Retrieval using Statistical Machine Translation on User Queries

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Multilingual Document Clustering: an Heuristic Approach Based on Cognate Named Entities

Using dialogue context to improve parsing performance in dialogue systems

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Regression for Sentence-Level MT Evaluation with Pseudo References

The stages of event extraction

A hybrid approach to translate Moroccan Arabic dialect

Cross-lingual Text Fragment Alignment using Divergence from Randomness

Memory-based grammatical error correction

Emmaus Lutheran School English Language Arts Curriculum

Inteligencia Artificial. Revista Iberoamericana de Inteligencia Artificial ISSN:

ELD CELDT 5 EDGE Level C Curriculum Guide LANGUAGE DEVELOPMENT VOCABULARY COMMON WRITING PROJECT. ToolKit

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

Opportunities for Writing Title Key Stage 1 Key Stage 2 Narrative

EUROPEAN DAY OF LANGUAGES

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

A Quantitative Method for Machine Translation Evaluation

ScienceDirect. Malayalam question answering system

Grade 4. Common Core Adoption Process. (Unpacked Standards)

Finding Translations in Scanned Book Collections

Comprehension Recognize plot features of fairy tales, folk tales, fables, and myths.

The Acquisition of Person and Number Morphology Within the Verbal Domain in Early Greek

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

A High-Quality Web Corpus of Czech

Derivational and Inflectional Morphemes in Pak-Pak Language

BULATS A2 WORDLIST 2

PAGE(S) WHERE TAUGHT If sub mission ins not a book, cite appropriate location(s))

TINE: A Metric to Assess MT Adequacy

Stefan Engelberg (IDS Mannheim), Workshop Corpora in Lexical Research, Bucharest, Nov [Folie 1] 6.1 Type-token ratio

Writing a composition

Machine Translation on the Medical Domain: The Role of BLEU/NIST and METEOR in a Controlled Vocabulary Setting

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

Online Updating of Word Representations for Part-of-Speech Tagging

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

Written by: YULI AMRIA (RRA1B210085) ABSTRACT. Key words: ability, possessive pronouns, and possessive adjectives INTRODUCTION

Detecting English-French Cognates Using Orthographic Edit Distance

Greedy Decoding for Statistical Machine Translation in Almost Linear Time

Sample Goals and Benchmarks

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Ch VI- SENTENCE PATTERNS.

Improved Reordering for Shallow-n Grammar based Hierarchical Phrase-based Translation

ELA/ELD Standards Correlation Matrix for ELD Materials Grade 1 Reading

1. Introduction. 2. The OMBI database editor

The taming of the data:

Multilingual Sentiment and Subjectivity Analysis

The RWTH Aachen University English-German and German-English Machine Translation System for WMT 2017

Multi-Lingual Text Leveling

Learning Methods in Multilingual Speech Recognition

Introduction. Beáta B. Megyesi. Uppsala University Department of Linguistics and Philology Introduction 1(48)

Presentation Exercise: Chapter 32

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

knarrator: A Model For Authors To Simplify Authoring Process Using Natural Language Processing To Portuguese

Mercer County Schools

Correspondence between the DRDP (2015) and the California Preschool Learning Foundations. Foundations (PLF) in Language and Literacy

Adding syntactic structure to bilingual terminology for improved domain adaptation

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

Probabilistic Latent Semantic Analysis

CS Machine Learning

Morphosyntactic and Referential Cues to the Identification of Generic Statements

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

A Case Study: News Classification Based on Term Frequency

Interactive Corpus Annotation of Anaphor Using NLP Algorithms

CS 598 Natural Language Processing

LEXICAL COHESION ANALYSIS OF THE ARTICLE WHAT IS A GOOD RESEARCH PROJECT? BY BRIAN PALTRIDGE A JOURNAL ARTICLE

The Smart/Empire TIPSTER IR System

TIMSS Highlights from the Primary Grades

TABE 9&10. Revised 8/2013- with reference to College and Career Readiness Standards

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Character Stream Parsing of Mixed-lingual Text

Transcription:

Improving SMT for Baltic Languages with Factored Models Raivis SKADIŅŠ a,b, Kārlis GOBA a a and Valters ŠICS a Tilde SIA, Latvia b University of Latvia, Latvia Abstract. This paper reports on implementation and evaluation of English-Latvian and Lithuanian-English statistical machine translation systems. It also gives brief introduction of project scope Baltic languages, prior implementations of MT and evaluation of MT systems. In this paper we report on results of both automatic and human evaluation. Results of human evaluation show that factored SMT gives significant improvement of translation quality compared to baseline SMT. Keywords. Statistical Machine Translation, Factored models Introduction Besides Google machine translation engines and research experiments with statistical MT for Latvian [1] and Lithuanian, there are both English-Latvian [2] and English- Lithuanian [3] rule based MT systems available. Both Latvian and Lithuanian are morphologically rich languages with quite free phrase order in a sentence and with very limited parallel corpora available. All mentioned aspects are challenging for SMT systems. We used Moses SMT toolkit [4] for SMT system training and decoding. The aim of the project was not to build yet another SMT using publicly available parallel corpora and tools, but also to add language specific knowledge to assess the possible improvement of translation quality. Another important aim of this project was the evaluation of available MT systems; we wanted to understand whether we can build SMT systems outperforming other existing statistical and rule based MT systems. 1. Training resources For training the SMT systems, both monolingual and bilingual sentence-aligned parallel corpora of substantial size are required. The corpus size largely determines the quality of translation, as has been shown both in case of multilingual SMT [5] and English-Latvian SMT [1]. For all of our trained SMT systems the parallel training corpus includes DGT-TM, OPUS and localization corpora. The DGT-TM corpus is a publicly available collection of legislative texts available in 22 languages of European Union. The OPUS translated text collection [6][7] contains publicly available texts from web in different domains. For Latvian we chose the EMEA (European Medicines Agency) sentence-aligned

corpus. For Lithuanian we chose the EMEA and the KDE4 sentence-aligned corpus. Localization parallel corpus was obtained from translation memories that were created during localization of software content, appliance user manuals and software help content. We additionally included word and phrase translations from bilingual dictionaries to increase word coverage. Both parallel and monolingual corpora were filtered according to different criteria. Suspicious sentences containing too much non-alphanumeric symbols and repeated sentences were removed. Monolingual corpora were prepared from the corresponding monolingual part of parallel corpora, as well as news articles from Web for Latvian and LCC (Leipzig Corpora Collection) corpus for English. Table 1. Bilingual corpora for English-Latvian system Bilingual corpus Localization TM DGT-TM OPUS EMEA Fiction Dictionary data Total Table 2. Bilingual corpora for Lithuanian-English system Bilingual corpus Localization TM DGT-TM OPUS EMEA Dictionary data OPUS KDE4 Total Table 3. Monolingual corpora Monolingual corpus Latvian side of parallel corpus News (web) Fiction Total, Latvian Parallel units ~1.29 mil. ~1.06 mil. ~0.97 mil. ~0.66 mil. ~0.51 mil. 4.49 mil. (3.23 mil. filtered) Parallel units ~1.56 mil. ~0.99 mil. ~0.84 mil. ~0.38 mil. ~0.05 mil. 3.82 mil. (2.71 mil. filtered) 60M 250M 9M 319M Words English side of parallel corpus 60M News (WMT09) 440M LCC 21M Total, English 521M The evaluation and development corpora were prepared separately. For both corpora we used the same mixture of different domains and topics (Table 4) representing the expected translation needs of a typical user. The development corpus contains 1000 sentences, while the evaluation set is 500 sentences long.

Table 4. Topic breakdown of evaluation and development sets Topic Percentage General information about European Union 12% Specifications, instructions and manuals 12% Popular scientific and educational 12% Official and legal documents 12% News and magazine articles 24% Information technology 18% Letters 5% Fiction 5% 2. SMT training The baseline SMT models were trained on lowercased surface forms for source and target languages only. The SMT baseline models were trained for reference point to assess the relative improvement of additional data manipulation, factors, corpus size and language models. The phrase-based approach allows translating source words differently depending on their context by translating whole phrases, whereas target language model allows matching target phrases at their boundaries. However, most phrases in inflectionally rich languages can be inflected in gender, case, number, tense, mood and other morphosyntactic properties, producing considerable amount of variations. Both Latvian and Lithuanian belong to the class of inflected languages which are the most complex from the point of view of morphology. Latvian nouns are divided into 6 declensions. Nouns and pronouns have 6 cases in both singular and plural. Adjectives, numerals and participles have 6 cases in singular and plural, 2 genders, and the definite and indefinite form. The rules of case generation differ for each group. There are two numbers, three persons and three tenses (present, future and past tenses), both simple and compound, and 5 moods in the Latvian conjugation system. Latvian is quite regular in the sense of forming inflected forms however the form endings in Latvian are highly ambiguous. Nouns in Latvian have 29 graphically different endings and only 13 of them are unambiguous, adjectives have 24 graphically different endings and half of them are ambiguous, verbs have 28 graphically different endings and only 17 of them are unambiguous. Lithuanian has even more morphological variation and ambiguity. Another significant feature of both languages is the relatively free word order in the sentence which makes parsing and translation complicated. The inflectional variation increases data sparseness at the boundaries of translated phrases, where a language model over surface forms might be inadequate to estimate the probability of target sentence reliably. The baseline SMT system was particularly weak at adjective-noun and subject-object agreement. To address that, we introduced an additional language model over morphologic tags in the English-Latvian system. The tags contain relevant morphologic properties (case, number, gender, etc.) that are generated by a morphologic tagger. The order of the tag LM was increased to 7, as the tag data has significantly smaller vocabulary. When translating from morphologically rich language, the SMT baseline system will not give translation for all forms of word that is not fully represented in the training data. The solution addressing this problem would be to separate richness of morphology from the words and translate lemmas instead. Morphology tags could be

used as additional factor to improve quality of translation. However, as we do not have a morphologic tagger for Lithuanian we used a simplified approach, splitting each token into two separate tokens containing the stem and an optional suffix. The stems and suffixes were treated in the same way in the training process. Suffixes were marked (prefixed by a special symbol) to avoid overlapping with stems. The suffixes we used correspond to inflectional endings of nouns, adjectives and verbs, however, they are not supposed to be linguistically accurate, but rather as a way to reduce data sparsity. Moreover, the processing always splits the longest matching suffix, which produces errors with certain words. We trained another English-Latvian system with a similar approach, using the suffixes instead of morphologic tags for the additional LM. Although the suffixes are often ambiguous (e.g. the ending -a is used in several noun, adjective and verb forms), our goal was to check whether we can get improvement in quality by using knowledge about morphology in case we do not have morphological tagger, and to assess how big is this improvement compared with using the tagger. Table 5 gives an overview of SMT systems trained and the structure of factored models. Table 5. Structure of Translation and Language Models System Translation Models Language Models EN-LV SMT baseline 1: Surface Surface 1: Surface form EN-LV SMT suffix 1: Surface Surface, suffix 1: Surface form 2: Suffix EN-LV SMT tag 1: Surface Surface, morphology tag 1: Surface form 2: Morphology tag LT-EN SMT baseline 1: Surface Surface 1: Surface form LT-EN SMT Stem/suffix 1: Stem/suffix Surface 1: Surface form LT-EN SMT Stem 1: Stem Surface 1: Surface form 3. Results and Evaluation 3.1. Automated evaluation We used BLEU [8] and NIST [9] metric for automatic evaluation. The summary of automatic evaluation results is presented in Table 6. Table 6. Automatic evaluation BLEU scores System Language pair BLEU Tilde rule-based MT English-Latvian 8.1% Google 1 English-Latvian 32.9% Pragma 2 English-Latvian 5.3% SMT baseline English-Latvian 24.8% SMT suffix English-Latvian 25.3% SMT tag English-Latvian 25.6% Google Lithuanian-English 29.5% SMT baseline Lithuanian-English 28.3% SMT stem/suffix Lithuanian-English 28.0% 1 Google Translate (http://translate.google.com/) as of July 2010 2 Pragma translation system (http://www.trident.com.ua/eng/produkt.html)

For Lithuanian-English system we also measured the out of vocabulary (OOV) rate on both per-word and per-sentence basis (Table 7). The per-word OOV rate is the percentage of untranslated words in the output text, and the per-sentence OOV rate is the percentage of sentences that contain at least one untranslated word. It was not possible to automatically determine the OOV rates for other translation systems (e.g. Google), as the OOV rates were calculated by analyzing the output of Moses decoder. Table 7. OOV rates for Lithuanian-English System Language pair OOV, Words OOV, Sentences SMT baseline Lithuanian-English 3.31% 39.8% SMT stem/suffix Lithuanian-English 2.17% 27.3% 3.2. Human evaluation We used a ranking of translated sentences relative to each other for manual evaluation of systems. This was the official determinant of translation quality used in the 2009 Workshop on Statistical Machine Translation shared tasks [10]. The same test corpus was used as in automatic evaluation. The summary of manual evaluation results is presented in Table 8. Table 8. Manual evaluation results for 3 systems, balanced test corpus System Language pair BLEU NIST Average rank in manual evaluation Tilde Rule-Based MT English-Latvian 8.1% 3.82 1.98 ± 0.08 SMT Baseline English-Latvian 21.7% 5.32 2.06 ± 0.07 SMT F1 English-Latvian 23.0% 5.40 1.59 ± 0.07 We did evaluation both ranking several systems simultaneously and ranking only two systems (ties were allowed). We discovered that it is more convenient for evaluators to evaluate only two systems and results of such evaluations are easier to interpret as well. We developed a web based evaluation environment where we can upload sources sentences and outputs of two MT systems as simple txt files. Once evaluation of two systems is set up we can send a link of evaluation survey to evaluators. Evaluators are evaluating systems sentence by sentence. Evaluators see source sentence and output of two MT systems. The order of MT system outputs in evaluation differs; sometimes evaluator gets the output of the first system in a first position, sometimes he gets the output of the second system in a first position. Evaluators are encouraged to evaluate at least 25 sentences, we allow evaluator to perform evaluation is small portions. Evaluator can open the evaluation survey and evaluate few sentences and go away and come back later to continue. Each evaluator never gets the same sentence to evaluate. We are calculating how often users prefer each system based on all answers and based on comparison of sentences. When we calculate evaluation results based on all answers we just count how many times users chose one system to be better than other. In a result we get percentage showing how in many percents of answers users preferred one system over the other. To be sure about the statistical relevance of results we also calculate confidence interval of the results. If we have A users preferring the first system and B users preferring the second system, then we calculate percentage using Eq. (1) and confidence interval using Eq. (2).

(1) (2) where z for a 95% confidence interval is 1.96. When we have calculated p and ci, then we can say that users prefer the first system over the second in p±ci percents of individual evaluations. We say that evaluation results are weakly sufficient to say that with a 95% confidence the first system is better than the second if Eq. (3) is true. Such evaluation results are weakly sufficient because they are based on all evaluations but they do not represent system output variation from sentence to sentence. We can perform system evaluation using just one test sentence and get such weakly sufficient evaluation results. It is obvious that such evaluation is not reliable. To get more reliable results we have to base evaluation on sentences instead of all answers. We can calculate how evaluators have evaluated systems on a sentence level; if we have A evaluators preferring the particular sentence from the first system and B evaluators preferring sentence from the second system, then we can calculate percentage using Eq. (1) and confidence interval using Eq. (2). We say that particular sentence is translated better by the first system than by other system if Eq. (3) is true. To get more reliable evaluation results we are not asking evaluators to evaluate sentences which have sufficient confidence that they are translated better by one system than by other. When we have A sentences evaluated to be better translated by the first system and B sentences evaluated to be better translated by the second system or systems are in tie, then we can calculate evaluation results on sentence level using Eqs. (1) and (2) again. And we can say that evaluation results are strongly sufficient to say that the first system is better than the second in the sentence level if Eq. (3) is true. We can say that evaluation is just sufficient if we ignore ties. Table 9. Manual evaluation results. Comparison of two systems System1 System2 Language pair p ci SMT F1 SMT baseline English-Latvian 58.67 % ±4.98 % Google SMT F1 English-Latvian 55.73 % ±6.01 % SMT stem/suffix SMT baseline Lithuanian-English 52.32 % ±4.14 % Best factored systems where compared to baseline systems and best English- Latvian factored system to Google SMT system using system for manual comparison of two systems described above. Results of manual evaluation are given in Table 9. Manual comparison of English-Latvian factored and baseline SMT systems shows that evaluation results are sufficient to say that factored system is better than baseline system, because in 58.67% (± 4.98%) of cases users judged its output to be better than the output of baseline system. Manual comparison of English-Latvian factored and Google systems shows that Google system is slightly better, but evaluation results are not sufficient to say that it is really better, because the difference between systems is not statistically significant (55.73 6.01% < 50%). Manual comparison of our best Lithuanian-English and the baseline systems shows that system with stems and suffixes is slightly better, but evaluation results are not sufficient to say that with a strong (3)

confidence, because difference between systems also is not statistically significant (52.32 4.14% < 50%). 4. Conclusions The MT system evaluation shows that used automatic metrics are unreliable for comparing rule-based and statistical systems, strongly favoring the latter. Both Pragma and Tilde rule-based systems have received very low BLEU score. This behavior of automated metrics has been shown before [11]. By development of factored EN-LV SMT models we expected to improve human assessment of quality by targeting local word agreement and inter-phrase consistency. Human evaluation shows a clear preference for factored SMT over the baseline SMT, which operates only with the surface forms. However, automated metric scores show only slight improvement on balanced test corpus (BLEU 21.7% vs 23.8%). By developing of the LT-EN SMT Stem/suffix model we expected to increase overall translation quality by reduction of untranslated words. The BLEU score slightly decreased (BLEU 28.0% vs 28.3%), however the OOV rate differs significantly. Human evaluation results suggest that users prefer lower OOV rate despite slight reduction in overall translation quality in terms of BLEU score. Acknowledgements The research within the project LetsMT! leading to these results has received funding from the ICT Policy Support Programme (ICT PSP), Theme 5 Multilingual web, grant agreement n o 250456 and this research was partially supported by the European Social Fund (ESF) activity No. 1.1.2.1.2 Support to doctor s studies, project No. 2009/0138/1DP/1.1.2.1.2/09/IPIA/VIAA/004. References [1] I. Skadiņa, E. Brālītis, English-Latvian SMT: knowledge or data, in Proceedings of the 17th Nordic Conference on Computational Linguistics NODALIDA, Odense, Denmark, NEALT Proceedings Series, Vol. 4 (2009), 242 245., 2009. [2] R. Skadiņš, I. Skadiņa, D. Deksne, T. Gornostay, English/Russian-Latvian Machine Translation System, in Proceedings of HLT 2007, Kaunas, Lithuania, 2007. [3] E. Rimkute, J. Kovalevskaite, Linguistic Evaluation of the First English-Lithuanian Machine Translation System, in Proceedings of HLT 2007, Kaunas, Lithuania, 2007. [4] P. Koehn, M. Federico, B. Cowan, R. Zens, C. Duer, O. Bojar, A. Constantin, E. Herbst, Moses: Open Source Toolkit for Statistical Machine Translation, in Proceedings of the ACL 2007 Demo and Poster Sessions, pages 177-180, Prague, 2007. [5] P. Koehn, J.F. Och, D. Marcu, Statistical Phrase-Based Translation, in Proceedings of HLT/NAACL 2003, 2003. [6] J. Tiedemann, L. Nygaard, The OPUS corpus - parallel & free, in Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC'04). Lisbon, Portugal, May 26-28., 2004. [7] J. Tiedemann, News from OPUS - A Collection of Multilingual Parallel Corpora with Tools and Interfaces, in N. Nicolov, K. Bontcheva, G. Angelova, R. Mitkov (eds.) Recent Advances in Natural Language Processing (vol V), 237-248, John Benjamins, Amsterdam/Philadelphia, 2009.

[8] K. Papineni, S. Roukos, T. Ward, W. Zhu, BLEU: a method for automatic evaluation of machine translation, in Proceedings of the 40th Annual Meeting of the Association of Computational Linguistics (ACL), 2002 [9] G. Doddington, Automatic evaluation of machine translation quality using n-gram co-occurrence statistics, in Proceedings of HLT-02, 2002. [10] C. Callison-Burch, P. Koehn, C. Monz, J. Schroeder, Findings of the 2009 Workshop on Statistical Machine Translation, in Proceedings of the Fourth Workshop on Statistical Machine Translation, 1-28, Athens, Greece, 2009. [11] C. Callison-Burch, M. Osborne, P. Koehn, Re-evaluating the role of BLEU in machine translation research, in Proceedings of EACL, 2006.