Byte-based Neural Machine Translation

Similar documents
The KIT-LIMSI Translation System for WMT 2014

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

The RWTH Aachen University English-German and German-English Machine Translation System for WMT 2017

Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

arxiv: v1 [cs.cl] 2 Apr 2017

Residual Stacking of RNNs for Neural Machine Translation

The NICT Translation System for IWSLT 2012

Language Model and Grammar Extraction Variation in Machine Translation

Noisy SMS Machine Translation in Low-Density Languages

Initial approaches on Cross-Lingual Information Retrieval using Statistical Machine Translation on User Queries

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

arxiv: v4 [cs.cl] 28 Mar 2016

Constructing Parallel Corpus from Movie Subtitles

A heuristic framework for pivot-based bilingual dictionary induction

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Training and evaluation of POS taggers on the French MULTITAG corpus

Cross Language Information Retrieval

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Python Machine Learning

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Enhancing Morphological Alignment for Translating Highly Inflected Languages

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

arxiv: v3 [cs.cl] 7 Feb 2017

A study of speaker adaptation for DNN-based speech synthesis

Speech Emotion Recognition Using Support Vector Machine

Word Segmentation of Off-line Handwritten Documents

Modeling function word errors in DNN-HMM based LVCSR systems

Finding Translations in Scanned Book Collections

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Overview of the 3rd Workshop on Asian Translation

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Methods for Fuzzy Systems

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Online Updating of Word Representations for Part-of-Speech Tagging

EACL th Conference of the European Chapter of the Association for Computational Linguistics. Proceedings of the 2nd International Workshop on

METHODS FOR EXTRACTING AND CLASSIFYING PAIRS OF COGNATES AND FALSE FRIENDS

Lessons from a Massive Open Online Course (MOOC) on Natural Language Processing for Digital Humanities

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

IB Diploma Program Language Policy San Jose High School

3 Character-based KJ Translation

Multilingual Document Clustering: an Heuristic Approach Based on Cognate Named Entities

Combining Bidirectional Translation and Synonymy for Cross-Language Information Retrieval

Cross-lingual Text Fragment Alignment using Divergence from Randomness

Second Exam: Natural Language Parsing with Neural Networks

ATENEA UPC AND THE NEW "Activity Stream" or "WALL" FEATURE Jesus Alcober 1, Oriol Sánchez 2, Javier Otero 3, Ramon Martí 4

Deep Neural Network Language Models

arxiv: v3 [cs.cl] 24 Apr 2017

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

A Quantitative Method for Machine Translation Evaluation

Lip Reading in Profile

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Georgetown University at TREC 2017 Dynamic Domain Track

Matching Similarity for Keyword-Based Clustering

arxiv: v2 [cs.cl] 18 Nov 2015

A Neural Network GUI Tested on Text-To-Phoneme Mapping

NEURAL DIALOG STATE TRACKER FOR LARGE ONTOLOGIES BY ATTENTION MECHANISM. Youngsoo Jang*, Jiyeon Ham*, Byung-Jun Lee, Youngjae Chang, Kee-Eung Kim

Semantic and Context-aware Linguistic Model for Bias Detection

Matching Meaning for Cross-Language Information Retrieval

TINE: A Metric to Assess MT Adequacy

arxiv: v1 [cs.cv] 10 May 2017

ON THE USE OF WORD EMBEDDINGS ALONE TO

Re-evaluating the Role of Bleu in Machine Translation Research

Regression for Sentence-Level MT Evaluation with Pseudo References

EUROPEAN DAY OF LANGUAGES

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Improved Reordering for Shallow-n Grammar based Hierarchical Phrase-based Translation

the contribution of the European Centre for Modern Languages Frank Heyworth

USER ADAPTATION IN E-LEARNING ENVIRONMENTS

Detecting English-French Cognates Using Orthographic Edit Distance

Postprint.

Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

A High-Quality Web Corpus of Czech

Rule Learning With Negation: Issues Regarding Effectiveness

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

arxiv: v2 [cs.cv] 3 Aug 2017

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Modeling function word errors in DNN-HMM based LVCSR systems

Noisy Channel Models for Corrupted Chinese Text Restoration and GB-to-Big5 Conversion

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Applications of memory-based natural language processing

(English translation)

SARDNET: A Self-Organizing Feature Map for Sequences

Language Center. Course Catalog

An Online Handwriting Recognition System For Turkish

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

1. Introduction. 2. The OMBI database editor

LIM-LIG at SemEval-2017 Task1: Enhancing the Semantic Similarity for Arabic Sentences with Vectors Weighting

Speaker Identification by Comparison of Smart Methods. Abstract

Bibliography Deep Learning Papers

Boosting Named Entity Recognition with Neural Character Embeddings

Transcription:

Byte-based Neural Machine Translation Marta R. Costa-jussà, Carlos Escolano and José A. R. Fonollosa TALP Research Center, Universitat Politècnica de Catalunya, Barcelona marta.ruiz@upc.edu,carlos.escolano@tsc.upc.edu, jose.fonollosa@upc.edu Abstract This paper presents experiments comparing character-based and byte-based neural machine translation systems. The main motivation of the byte-based neural machine translation system is to build multilingual neural machine translation systems that can share the same vocabulary. We compare the performance of both systems in several language pairs and we see that the performance in test is similar for most language pairs while the training time is slightly reduced in the case of byte-based neural machine translation. 1 Introduction Multilingual neural machine translation is raising interest in the community because it re-opens the possibility of an interlingual architecture. the main advantage of the current setting is that interlingua is not manually designed but it seems that it can be automatically extracted (Johnson et al., 2016). In addition, this multilingual environment seems to allow to build translation systems among language pairs that do not have parallel corpus available (Johnson et al., 2016), what is called zeroshot translation. These two motivations (interlingua and zeroshot translation) are strong enough to motivate the entire commmunity to experiment towards multilingual architectures. Recently, there have appeared works in multilingual word representations (Schwenk et al., 2017; España-Bonet et al., 2017) Most multilingual works are at the level of words. As multilingual character research we can find (Lee et al., 2016) which goes from manyto-one languages in translation and achieves improvements for several language pairs. Previous work on character-based neural machine translation includes (Ling et al., 2015; Costa-jussà and Fonollosa, 2016), among others. We want to explore multilingual characterbased neural machine translation with a diversity of languages, including languages as Chinese. In case of using languages with different alphabets, the character dictionary can not be shared or it has to be considerably augmented. In order to keep the dictionary to the order of hundreds, we want to explore how byte-based neural machine translation behaves. In this paper, we propose to use the fullycharacter neural machine translation architecture (Lee et al., 2016) but using bytes instead of characters. We compare the performance of character against byte-based neural machine translation among similar languages (Catalan/Spanish and Portuguese/Brazilian) and relatively far languages (in terms of alphabet) (German/Finnish/Turkish- English). As far as we are concerned, we are not aware of any research work in neural machine translation that has experimented with bytes. Related work can be found int the area of natural language processing. Gillick et al. (2016) propose an neural network that reads text as bytes and use this model in tasks of Part-of-Speech and Named Entity Recognition. The recent investigation of Irie et al (2017) describes the use of a byte-level convolutional layer (instead of character-level) in the neural language model (Irie et al., 2017), which is applied to low resource speech recognition. 2 Character-based Neural Machine Translation Our system uses the architecture from (Lee et al., 2016) where a character-level neural MT model that maps the source character sequence to the target character sequence. The main difference in the encoder architecture of the standard neural 154 Proceedings of the First Workshop on Subword and Character Level Models in NLP, pages 154 158, Copenhagen, Denmark, September 7, 2017. c 2017 Association for Computational Linguistics.

MT model from (Bahdanau et al., 2015) is that instead of using word embeddings, the system uses character embeddings based on previous works like (Kim et al., 2015; Costa-jussà and Fonollosa, 2016). The architecture uses character embeddings include convolution layers, max pooling and highway network layers. The character embeddings from the decoder are the input of the bidirectional recurrent neural network. The main difference in the decoder architecture is that the singlelayer feedforward network computes the attention score of next target character (instead of word) to be generated with every source segment representation. And afterwards, a two-layer character-level decoder takes the source context vector from the attention mechanism and predicts each target character. 3 Byte-based Neural Machine Translation The byte-based Neural Machine Translation changes the character representation of words to the byte representation. Each sentence is represented as the concatenation of bytes that form its characters in utf-8 encoding. No explicit vocabulay is used but we can consider the byte representation as a vocabulary of 256 positions in which every possible byte can be represented. This modifications provides the following improvements over the previously seen architecture. Both languages share the same representation. If a word is identical in the source and in the target language they share the same representation when converted into sequences to be fed in the network. This is an advantage over the character-based representation, which dictionary is language-dependent. This representation uses a limited set of tokens of size 256 independently of the language. Therefore, the system is not affected by the size of character vocabulary. Note that there are languages that have a very rich explicit morphological representation or that have a wide range of characters (e.g. Chinese). However, the byte-based decoding also produces a sequence of correct bytes in a similar way that character level translation works compared to word-based systems. All words are theoretically representable by the system even if they have not been previuosly seen in the training. This is due to the fact that every single character of word can be seen as a concatenation of bytes and the full range of possible bytes is covered by the 4 Experimental Framework In this section we detail experimental corpora, architecture and parameters that we used. 4.1 Data and Preprocessing For Catalan-Spanish, We use a large corpus extracted from ten years of the paper edition of a bilingual Catalan newspaper, El Periódico (Costajussà et al., 2014). The Spanish-Catalan corpus is partially available via ELDA (Evaluations and Language Resources Dis-tribution Agency) in catalog number ELRA-W0053. Development and test sets are extracted from the same corpus. For Portuguese-Brazilian, we used the OPUS corpus (Tiedemann, 2012) which is a growing collection of translated texts from the web. In particular, for Portuguese-Brazilian the source corpus are from Ubuntu and GNOME. We extracted the parallel text from translation memories (TMX format) and from the complete text, we extracted a collection of development and test set. Finally, we used WMT 2017 1 corpus data for German, Finish and Turkish to English. For the three language pairs, we used all data parallel data provided in the evaluation. For German-English, we used: europarl v.7, news commentary v.12, common crawl and rapid corpus of EU press releases. For Finnish-English, we used europarl v.8, wiki headlines and rapid corpus of EU press releases. For Turkish-English, we used setimes2. The German and Finish test set is the news 2015 evaluation set, for Turkish the test set is the news 2016 evaluation set. Preprocessing consisted in cleaning empty sentences, limiting sentences up to 50 words, tokenization and truecasing for each language using tools from Moses (Koehn et al., 2007). Table 1 shows details about the corpus statistics after preprocessing. 4.2 Parameters Both character and byte-based systems share the same parameters. Further research may explore different parameters for the byte-based system, 1 http://www.statmt.org/wmt17/translation-task.html 155

LP L Set S W V Train 6478618 165170227 736873 Es Dev 2244 55478 12237 EsCa Test 2244 55988 12218 Train 6478618 178954335 713445 Ca Dev 2244 60130 11734 Test 2244 60693 11691 Train 4280310 33954616 362592 Pt Dev 2000 16122 4155 PtBr Test 2000 16012 4141 Train 4280310 33600508 320972 Br Dev 2000 15963 3939 Test 2000 15663 3964 Train 9659106 2 03634165 1721113 De Dev 2999 62362 12674 DeEn Test 2169 44085 9895 Train 9659106 210205446 954387 En Dev 2999 64503 9506 Test 2169 46830 7871 Train 2468673 3 7755811 863898 Fi Dev 3000 47779 16236 Test 2870 43069 15748 FiEn Train 2468673 52262051 240625 En Dev 3000 63519 9059 Test 2870 60149 8961 Train 200290 42 48508 158276 Tu Dev 1001 16954 6463 TuEn Test 3000 54128 15898 Train 299290 4713025 73906 En Dev 1001 22136 4318 Test 3000 66394 9503 Table 1: Corpus Statistics. Number of sentences (S),words (W), vocabulary (V). since we are adopting for the byte-based systems the character-based parameters from previous research (Lee et al., 2016). For the embedding of the source sentence, we use set of convolutional layers which number kernels are (200-200-250-250-300-300-300-300) and their lengths are (1-2-3-4-5-6-7-8) respectively. Additionally 4 highway layers are employed. And a bidirectional LSTM layer of 512 units for encoding. The maximum souce sentence s length is 450 during training and 500 for decoding both during training and sampling. 4.3 Byte differences among language pairs Characters may be represented by a single or several bytes. English and its close languages usually have a correspondance of character and byte (a character is represented by a single byte). Therefore, these languages are not really affected by this representation, mainly because the ASCII encoding makes possible to represent all possible characters in a single byte which results in a similar length representation in both baseline and proposed However, in other languages (e.g. Turkish, Finnish...) which contain stressed characters (among other modifications), a single character in utf-8 may be a concatenation of several bytes. For these cases, the performance of the byte-based Language Bytes Spanish 1.026 Catalan 1.040 German 1.015 Finnish 1.044 Turkish 1.087 English 1.000 Portuguese 1.027 Brazilian Portuguese 1.028 Chinese 2.108 Table 2: Mean bytes for character for all the languages tested, Chinese added for comparison. system differs from the character-based Table 2 shows the mean number of bytes for character changes for different languages. As the languages are more similiar to English the difference between bytes and characters changes. For all the languages tested in our experiments using all latin alphabet the differences are small and resultant sentence length is similar to its character counterpart. On the other hand for languages that use a different alphabet such as Chinese we can observe how for each character more than two bytes have to be correctly generated. 5 Results This section compares the performance of the byte-based neural machine translation system with the character-based in terms of translation quality and training time. In order to compare training times all systems have been trained in the same machine using an NVIDIA TITAN X with 12GB of RAM. Table 3 shows BLEU results and number for close languages Catalan-Spanish and Portuguese- Brazilian in both directions. Comparison between character and byte-based models shows that by using a byte-base system comparable results to the ones obtained using a character-based In our experiments, we have observed that the bytebased system tends to converge at least a couple of hundred iterations earlier than the character-based Table 4 shows BLEU results for distant languages German, Finish and Turkish into English. Comparison between character and byte-based models shows how even in distance languages similar or even equal results can be obtained us- 156

Language System BLEU esca char 87.12 byte 86.91 caes char 82.43 byte 81.27 ptbr char 48.54 byte 48.28 brpt char 46.41 byte 47.20 Table 3: Close languages. BLEU results. Language System BLEU tren char 5.87 byte 4.75 deen char 28.63 byte 25.29 fien char 14.75 byte 14.75 Table 4: Distant languages. BLEU results. ing the byte-based For the case of Turkish and English (and training for an equal number of iterations), results are lower than the ones obtained by the char-based system but given the reduced size of the corpus variance in the results can be caused by other factors. For Finnish-English, we achieved the same results with several hundred training iterations less. In the case of the German-English pair, results show that the charbased approach provides quite better results than the proposed byte-based system but also it is worth mentioning that when using bytes, the results converged several hundred training iterations before than when using the baseline 6 Conclusions This paper shows an experimental comparison between char-based neural machine translation and byte-based neural machine translation. Variability has been found in the results for different language pairs ranging for 3.3 to +0.8 BLEU points. Also the main advantage of the proposed system compared to the baseline system is that it tends to require less training iterations which translates in a reduction of training time that can vary from several days to even two weeks. Further work includes the experimentation of multilingual neural machine translation using byte-based In addition, we want to include in our experimentation languages with complete different alphabets such as Chinese and we will investigate if byte-based neural machine translation system proposes malformed characters. In future research, we will test different parameters such as changing the number of convolutional filters. Acknowledgements This work is supported by the Spanish Ministerio de Economía y Competitividad and Fondo Europeo de Desarrollo Regional, through contract TEC2015-69266-P (MINECO/FEDER, UE) and the postdoctoral senior grant Ramón y Cajal. References Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. volume abs/1409.0473. http://arxiv.org/abs/1409.0473. Marta R. Costa-jussà and José A. R. Fonollosa. 2016. Character-based neural machine translation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Association for Computational Linguistics, Berlin, Germany, pages 357 361. http://anthology.aclweb.org/p16-2058. Marta R. Costa-jussà, March Poch, José A.R. Fonollosa, Mireia Farrús, and José B. Mariño. 2014. A large Spanish-Catalna parallel corpus release for Machine Translation. Computing and Informatics Journal 33. Cristina España-Bonet, Ádám Csaba Varga, Alberto Barrón-Cedeño, and Josef van Genabith. 2017. An empirical analysis of nmt-derived interlingual embeddings and their use in parallel sentence identification. CoRR abs/1704.05415. http://arxiv.org/abs/1704.05415. Dan Gillick, Cliff Brunk, Oriol Vinyals, and Amarnag Subramanya. 2016. Multilingual language processing from bytes. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. San Diego, California, pages 1296 1306. Kazuki Irie, Pavel Golik, Ralf Schlter, and Hermann Ney. 2017. Investigations on byte-level convolutional neural networks for language modeling in low resource speech recognition page 5. Melvin Johnson, Mike Schuster, Quoc V. Le, Maxim Krikun, Yonghui Wu, Zhifeng Chen, Nikhil Thorat, Fernanda B. Viégas, Martin Wattenberg, Greg Corrado, Macduff Hughes, and Jeffrey Dean. 2016. Google s multilingual neural machine translation system: Enabling zero-shot translation. CoRR abs/1611.04558. http://arxiv.org/abs/1611.04558. 157

Yoon Kim, Yacine Jernite, David Sontag, and Alexander M. Rush. 2015. Character-aware neural language models. CoRR abs/1508.06615. http://arxiv.org/abs/1508.06615. Philipp Koehn, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Richard Zens, Chris Dyer, Ondřej Bojar, Alexandra Constantin, and Evan Herbst. 2007. Moses: Open source toolkit for statistical machine translation. In Proceedings of the 45th Annual Meeting of the ACL on Interactive Poster and Demonstration Sessions. Association for Computational Linguistics, Stroudsburg, PA, USA, ACL 07, pages 177 180. http://dl.acm.org/citation.cfm?id=1557769.1557821. Jason Lee, Kyunghyun Cho, and Thomas Hofmann. 2016. Fully character-level neural machine translation without explicit segmentation. CoRR abs/1610.03017. http://arxiv.org/abs/1610.03017. Wang Ling, Isabel Trancoso, Chris Dyer, and Alan W. Black. 2015. Character-based neural machine translation. CoRR abs/1511.04586. Holger Schwenk, Ke Tran, Orhan Firat, and Matthijs Douze. 2017. Learning joint multilingual sentence representations with neural machine translation. CoRR abs/1704.04154. http://arxiv.org/abs/1704.04154. Jörg Tiedemann. 2012. Parallel data, tools and interfaces in opus. In Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC-2012). European Language Resources Association (ELRA), Istanbul, Turkey, pages 2214 2218. 158