Transfer Learning across Low-Resource, Related Languages for Neural Machine Translation

Similar documents
Residual Stacking of RNNs for Neural Machine Translation

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

The RWTH Aachen University English-German and German-English Machine Translation System for WMT 2017

The KIT-LIMSI Translation System for WMT 2014

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

arxiv: v1 [cs.cl] 2 Apr 2017

A heuristic framework for pivot-based bilingual dictionary induction

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

The NICT Translation System for IWSLT 2012

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Language Model and Grammar Extraction Variation in Machine Translation

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

arxiv: v4 [cs.cl] 28 Mar 2016

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

Second Exam: Natural Language Parsing with Neural Networks

Noisy SMS Machine Translation in Low-Density Languages

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

arxiv: v5 [cs.ai] 18 Aug 2015

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

arxiv: v1 [cs.lg] 15 Jun 2015

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Artificial Neural Networks written examination

Learning Methods in Multilingual Speech Recognition

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

Deep Neural Network Language Models

Lecture 1: Machine Learning Basics

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Python Machine Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Detecting English-French Cognates Using Orthographic Edit Distance

Lip Reading in Profile

Florida Reading Endorsement Alignment Matrix Competency 1

What Can Neural Networks Teach us about Language? Graham Neubig a2-dlearn 11/18/2017

Speech Recognition at ICSI: Broadcast News and beyond

Modeling function word errors in DNN-HMM based LVCSR systems

(Sub)Gradient Descent

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

arxiv: v3 [cs.cl] 7 Feb 2017

3 Character-based KJ Translation

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

Georgetown University at TREC 2017 Dynamic Domain Track

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

A study of speaker adaptation for DNN-based speech synthesis

The A2iA Multi-lingual Text Recognition System at the second Maurdor Evaluation

Modeling function word errors in DNN-HMM based LVCSR systems

Cross Language Information Retrieval

Knowledge Transfer in Deep Convolutional Neural Nets

Linking Task: Identifying authors and book titles in verbose queries

Word Segmentation of Off-line Handwritten Documents

Learning Methods for Fuzzy Systems

Overview of the 3rd Workshop on Asian Translation

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

English Language and Applied Linguistics. Module Descriptions 2017/18

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Exploration. CS : Deep Reinforcement Learning Sergey Levine

arxiv: v1 [cs.lg] 7 Apr 2015

A Comparison of Two Text Representations for Sentiment Analysis

Calibration of Confidence Measures in Speech Recognition

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

The Oregon Literacy Framework of September 2009 as it Applies to grades K-3

METHODS FOR EXTRACTING AND CLASSIFYING PAIRS OF COGNATES AND FALSE FRIENDS

Problems of the Arabic OCR: New Attitudes

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

arxiv: v2 [cs.ir] 22 Aug 2016

arxiv: v1 [cs.cv] 10 May 2017

Semantic and Context-aware Linguistic Model for Bias Detection

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

A Reinforcement Learning Variant for Control Scheduling

Modeling full form lexica for Arabic

Improvements to the Pruning Behavior of DNN Acoustic Models

Generative models and adversarial training

Program Matrix - Reading English 6-12 (DOE Code 398) University of Florida. Reading

An Evaluation of the Interactive-Activation Model Using Masked Partial-Word Priming. Jason R. Perry. University of Western Ontario. Stephen J.

An Online Handwriting Recognition System For Turkish

Testing A Moving Target: How Do We Test Machine Learning Systems? Peter Varhol Technology Strategy Research, USA

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

The Strong Minimalist Thesis and Bounded Optimality

Individual Differences & Item Effects: How to test them, & how to test them well

On document relevance and lexical cohesion between query terms

Experiments with a Higher-Order Projective Dependency Parser

ON THE USE OF WORD EMBEDDINGS ALONE TO

A JOINT MANY-TASK MODEL: GROWING A NEURAL NETWORK FOR MULTIPLE NLP TASKS

Finding Translations in Scanned Book Collections

Probabilistic Latent Semantic Analysis

Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

arxiv: v3 [cs.cl] 24 Apr 2017

Radius STEM Readiness TM

Investigation on Mandarin Broadcast News Speech Recognition

Identification of Opinion Leaders Using Text Mining Technique in Virtual Community

Axiom 2013 Team Description Paper

Accessing Higher Education in Developing Countries: panel data analysis from India, Peru and Vietnam

Reinforcement Learning by Comparing Immediate Reward

Transcription:

Transfer Learning across Low-Resource, Related Languages for Neural Machine Translation Toan Q. Nguyen and David Chiang Department of Computer Science and Engineeering University of Notre Dame {tnguye28,dchiang}@nd.edu Abstract We present a simple method to improve neural translation of a low-resource language pair using parallel data from a related, also low-resource, language pair. The method is based on the method of Zoph et al., but whereas their method ignores any source vocabulary overlap, ours exploits it. First, we split words using Byte Pair Encoding (BPE) to increase vocabulary overlap. Then, we train a model on the first language pair and its parameters, including its source word embeddings, to another model and continue training on the second language pair. Our experiments show that learning helps word-based translation only slightly, but when used on top of a much stronger BPE, it yields larger improvements of up to 4.3 BLEU. 1 Introduction Neural machine translation (NMT) (Sutskever et al., 14; Bahdanau et al., 15) is rapidly proving itself to be a strong competitor to other statistical machine translation methods. However, it still lags behind other statistical methods on very lowresource language pairs (Zoph et al., 16; Koehn and Knowles, 17). A common strategy to improve learning of lowresource languages is to use resources from related languages (Nakov and Ng, 09). However, adapting these resources is not trivial. NMT offers some simple ways of doing this. For example, Zoph et al. (16) train a parent model on a (possibly unrelated) high-resource language pair, then use this model to initialize a child model which is further trained on a low-resource language pair. In particular, they showed that a French-English model could be used to improve translation on a wide range of low-resource language pairs such as Hausa-, Turkish-, and Uzbek-English. In this paper, we explore the opposite scenario, where the parent language pair is also lowresource, but related to the child language pair. We show that, at least in the case of three Turkic languages (Turkish, Uzbek, and Uyghur), the original method of Zoph et al. (16) does not always work, but it is still possible to use the parent model to considerably improve the child model. The basic idea is to exploit the relationship between the parent and child language lexicons. Zoph et al. s original method makes no assumption about the relatedness of the parent and child languages, so it effectively makes a random assignment of the parent-language word embeddings to child-language words. But if we assume that the parent and child lexicons are related, it should be beneficial to source word embeddings from parent-language words to their child-language equivalents. Thus, the problem amounts to finding a representation of the data that ensures a sufficient overlap between the vocabularies of the languages. To do this, we map the source languages to a common alphabet and use Byte Pair Encoding (BPE) (Sennrich et al., 16) on the union of the vocabularies to increase the number of common subwords. In our experiments, we show that learning helps word-based translation, but not always significantly. But when used on top of a much stronger BPE, it yields larger and statistically significant improvements. Using Uzbek as a parent language and Turkish and Uyghur as child languages, we obtain improvements over BPE of 0.8 and 4.3 BLEU, respectively. 296 Proceedings of the The 8th International Joint Conference on Natural Language Processing, pages 296 301, Taipei, Taiwan, November 27 December 1, 17 c 17 AFNLP

English Turkish Uzbek clinic poliklinikte, poliklinie, polikliniine poliklinikasi, poliklinikaga, poliklinikalar parliament meningococcus parlamentosuna, parlamentosundan, parlamentosu meningokokuna, meningokosemi, meningokoklar parlamentning, parlamentini, parlamentiga meningokokk, meningokokkli, meningokokkning Table 1: Some examples of similar words in Turkish and Uzbek that share the same root. 2 Background 2.1 Attentional Model We use the 2-layer, 512-hidden-unit global attentional model with general scoring function and input feeding by Luong et al. (15). For the purposes of this paper, the most important detail of the model is that (as in many other models) the word types of both the source and target languages are mapped to vector representations called word embeddings, which are learned automatically with the rest of the model. 2.2 Language We follow the learning approach proposed by Zoph et al. (16). In their work, a parent model is first trained on a high-resource language pair. Then the child model s parameter values are copied from the parent s and are fine-tuned on its low-resource data. The source word embeddings are copied with the rest of the model, with the ith parent-language word embedding being assigned to the ith childlanguage word. Because the parent and child source languages have different vocabularies, this amounts to randomly assigning parent source word embeddings to child source words. In other words, even if a word exists in both the parent and child vocabularies, it s extremely unlikely that it will be assigned the same embedding in both models. By contrast, because the target language is the same in both the parent and child models, the target word embeddings are frozen during finetuning. 2.3 Related languages The experiments described below are on translation from three Turkic languages to English. The Turkic language family is a group of related lanword-based BPE 5k BPE 60k model toks sents toks sents toks sents 10 6 10 3 10 6 10 3 10 6 10 3 Uzb par 1.5 102 2.4 92 1.9 103 Tur chd 0.9 56 1.5 50 1.2 57 Uzb par 1.5 102 2.4 90 2.0 103 Uyg chd 1.7 82 2.1 77 2.0 88 Table 2: Number of tokens and sentences in our training data. guages with a very wide geographic distribution, from Turkey to northeastern Siberia. Turkic languages are morphologically rich, and have similarities in phonology, morphology, and syntax. For instance, in our analysis of the training data, we find many Turkish and Uzbek words sharing the same root and meaning. Some examples are shown in Table 1. 2.4 Byte Pair Encoding BPE (Sennrich et al., 16) is an efficient word segmentation algorithm. It initially treats the words as sequences of character tokens, then iteratively finds and merges the most common token pair into one. The algorithm stops after a controllable number of operations, or when no token pair appears more than once. At test time, the learned merge operations are applied to merge the character sequences in test data into larger symbols. 3 Method The basic idea of our method is to extend the method of Zoph et al. (16) to share the parent and child s source vocabularies, so that when source word embeddings are red, a word that appears in both vocabularies keeps its embedding. In order for this to work, it must be the case 297

that the parent and child languages have considerable vocabulary overlap, and that when a word occurs in both languages, it often has a similar meaning in both languages. Thus, we need to process the data to make these two assumptions hold as much as possible. 3.1 Transliteration If the parent and child language have different orthographies, it should help to map them into a common orthography. Even if the two use the same script, some transformation could be applied; for example, we might change French -eur endings to Spanish -or. Here, we take a minimalist approach. Turkish and Uzbek are both written using Latin script, and we did not apply any transformations to them. Our Uyghur data is written in Arabic script, so we transliterated it to Latin script using an off-the-shelf transliterator. 1 The transliteration is a string homomorphism, replacing Arabic letters with English letters or consonant clusters independent of context. 3.2 Segmentation To increase the overlap between the parent and child vocabularies, we use BPE to break words into subwords. For the BPE merge rules to not only find the common subwords between two source languages but also ensure consistency between source and target segmentation among each language pair, we learn the rules from the union of source and target data of both the parent and child models. The rules are then used to segment the corpora. It is important to note that this results in a single vocabulary, used for both the source and target languages in both models. 4 Experiments We used Turkish-, Uzbek-, and Uyghur-English parallel texts from the LORELEI program. We tokenized all data using the Moses toolkit (Koehn et al., 07); for Turkish-English experiments, we also truecased the data. For Uyghur-English, the word-based models were trained in the original Uyghur data written in Arabic script; for BPEbased systems, we transliterated it to Latin script as described above. For the word-based systems, we fixed the vocabulary size and replaced out-of-vocabulary 1 https://cis.temple.edu/ anwar/code/ latin2uyghur.html words with UNK. We tried different sizes for each language pair; however, each word-based system s target vocabulary size is limited by that of the child, so we could only use up to 45,000 word types for Turkish-English and,000 for Uyghur- English. The BPE-based systems could make use of bigger vocabulary size thanks to the combination of both parent and child source and target vocabularies. We varied the number of BPE merge operations from 5,000 to 60,000. Instead of using a fixed vocabulary cutoff, we used the full vocabulary; to ensure the model still learns how to deal with unknown words, we trained on two copies of the training data: one unchanged, and one in which all rare words (whose frequency is less than 5) were replaced with UNK. Accordingly, the number of epochs was halved. Following common practice, we fixed an upper limit on training sentence length (discarding longer sentences). Because the BPE-based systems have shorter tokens and therefore longer sentences, we set this upper limit differently for the word-based and BPE-based systems to approximately equalize the total size of the training data. This led to a limit of 50 tokens for word-based models and 60 tokens for BPE-based models. See Table 2 for statistics of the resulting datasets. We trained using Adadelta (Zeiler, 12), with a minibatch size of 32 and dropout with a dropout rate of 0.2. We rescaled the gradient when its norm exceeded 5. For the Uzbek-English to Turkish- English experiment, the parent and child models were trained for 100 and 50 epochs, respectively. For the Uzbek-English to Uyghur-English experiment, the parent and child models were trained for 50 and 0, respectively. As mentioned above, the BPE models were trained for half as many epochs because their data is duplicated. We used beam search for translation on the dev and test sets. Since NMT tends to favor short translations (Cho et al., 14), we use the length normalization approach of Wu et al. (16) which uses a different score s(e f ) instead of logprobability: s(e f ) = lp(e) = log p(e f ) lp(e) (5 + e )α (5 + 1) α. We set α = 0.8 for all of our experiments. 298

Tur-Eng Uyg-Eng +freeze BLEU size BLEU size BLEU size word-based 8.1 30k 8.5 30k 8.6 30k BPE 12.4 10k 13.2 k word-based 8.5 15k 10.6 15k 8.8 15k BPE 11.1 10k 15.4 8k Table 3: Whereas learning at word-level does not always help, our method consistently yields a significant improvement over the stronger BPE systems. Scores are case-sensitive test BLEU. Key: size = vocabulary size (word-based) or number of BPE operations (BPE). The symbols and indicate statistically significant improvements with p < 0.05 and p < 0.01, respectively, while indicates a statistically insignificant improvement (p > 0.05). Word-based BPE-based Tur-Eng BLEU 15 10 5 +freeze 0 0 10 30 40 50 5 0 40 60 0 15 10 Uyg-Eng BLEU 30 25 +freeze 15 0 5 10 15 Vocabulary size ( 1000) 30 25 0 40 60 15 Number of BPE operations ( 1000) Figure 1: Tokenized dev BLEU scores for various settings as a function of the number of word/subword types. Key: = train child model only; = train parent, then child model; +freeze = freeze target word embeddings in child model. 299

task settings train dev Tur-Eng Uyg-Eng word-based 30k 3.9% 3.6% BPE k 58.8% 25.0% word-based 15k 0.5% 1.7% BPE 8k 57.2% 48.5% Table 4: Amount of child s source types that appear in parent. We stopped training when the tokenized BLEU score was maximized on the development set. We also optimized the vocabulary size and the number of BPE operations for the word-based and BPEbased systems, respectively, to maximize the tokenized BLEU on the development set. After translation at test time, we rejoined BPE segments, recased, and detokenized. Finally, we evaluated using case-sensitive BLEU. As a, we trained a child model using BPE but without (that is, with weights randomly initialized). We also compared against a word-based (without ) and two word-based systems using without vocabulary-sharing, corresponding with the method of Zoph et al. (16) ( 2.2): one where the target word embeddings are fine-tuned, and one where they are frozen. 5 Results and Analysis Our results are shown in Table 3. In this lowresource setting, we find that ring wordbased models does not consistently help. On Turkish-English, both methods give only a statistically insignificant improvement (p > 0.05); on Uyghur-English, without freezing target embeddings helps somewhat, but with freezing helps only insignificantly. In both language pairs, the models that use BPE perform much better than their word-based counterparts. When we apply learning to this much stronger, we find that the relative improvements actually increase; that is, the combined effect of BPE and learning is more than additive. On Turkish-English, the improvement is +0.8 BLEU over the BPE ; on Uyghur-English, a healthy +4.3 over the BPE. A similar pattern emerges when we examine the best BLEU scores on the development set (Figure 1). Whereas word-based methods help very little for Turkish-English, and help or hurt slightly for Uyghur-English, our BPE-based approach consistently improves over both the and word-based models. We surmise that the improvement is primarily due to the vocabulary overlap created by BPE (see Table 4). 6 Conclusion In this paper, we have shown that the learning method of Zoph et al. (16), while appealing, might not always work in a low-resource context. However, by combining it with BPE, we can improve NMT performance on a low-resource language pair by exploiting its lexical similarity with another related, low-resource language. Our results show consistent improvement in two Turkic languages. Our approach, which relies on segmenting words into subwords, seems well suited to agglutinative languages; further investigation would be needed to confirm whether our method works on other types of languages. 7 Acknowledgements This research was supported in part by University of Southern California subcontract 67108176 under DARPA contract HR0011-15-C-0115. Nguyen was supported by a fellowship from the Vietnam Education Foundation. We would like to express our great appreciation to Dr. Sharon Hu for letting us use her group s GPU cluster (supported by NSF award 1629914), and to NVIDIA corporation for the donation of a Titan X GPU. References Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 15. Neural machine translation by jointly learning to align and translate. In Proc. ICLR. Kyunghyun Cho, Bart van Merriënboer, Dzmitry Bahdanau, and Yoshua Bengio. 14. On the properties of neural machine translation: Encoder decoder approaches. In Proc. Workshop on Syntax, Semantics and Structure in Statistical Translation. Philipp Koehn, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Richard Zens, et al. 07. Moses: Open source toolkit for statistical machine translation. In Proc. ACL. Philipp Koehn and Rebecca Knowles. 17. Six challenges for neural machine translation. In Proc. First Workshop on Neural Machine Translation. Association for Computational Linguistics. 300

Thang Luong, Hieu Pham, and Christopher D. Manning. 15. Effective approaches to attention-based neural machine translation. In Proc. EMNLP. Preslav Nakov and Hwee Tou Ng. 09. Improved statistical machine translation for resource-poor languages using related resource-rich languages. In Proc. EMNLP. Rico Sennrich, Barry Haddow, and Alexandra Birch. 16. Neural machine translation of rare words with subword units. In Proc. ACL. Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 14. Sequence to sequence learning with neural networks. In Advances in Neural Information Processing Systems 27. Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, et al. 16. Google s neural machine translation system: Bridging the gap between human and machine translation. arxiv:1609.08144. Matthew D. Zeiler. 12. ADADELTA: An adaptive learning rate method. arxiv:1212.5701v1. Barret Zoph, Deniz Yuret, Jonathan May, and Kevin Knight. 16. Transfer learning for low-resource neural machine translation. In Proc. EMNLP. 301