An English to Xitsonga statistical machine translation system for the government domain

Similar documents
Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

Noisy SMS Machine Translation in Low-Density Languages

arxiv: v1 [cs.cl] 2 Apr 2017

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS

The KIT-LIMSI Translation System for WMT 2014

The NICT Translation System for IWSLT 2012

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

Language Model and Grammar Extraction Variation in Machine Translation

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Linking Task: Identifying authors and book titles in verbose queries

The taming of the data:

Re-evaluating the Role of Bleu in Machine Translation Research

Regression for Sentence-Level MT Evaluation with Pseudo References

A Quantitative Method for Machine Translation Evaluation

Constructing Parallel Corpus from Movie Subtitles

Training and evaluation of POS taggers on the French MULTITAG corpus

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation

A heuristic framework for pivot-based bilingual dictionary induction

Rule Learning With Negation: Issues Regarding Effectiveness

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Cross Language Information Retrieval

Learning Methods in Multilingual Speech Recognition

Lessons from a Massive Open Online Course (MOOC) on Natural Language Processing for Digital Humanities

The role of the first language in foreign language learning. Paul Nation. The role of the first language in foreign language learning

Impact of Controlled Language on Translation Quality and Post-editing in a Statistical Machine Translation Environment

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

Providing student writers with pre-text feedback

Inteligencia Artificial. Revista Iberoamericana de Inteligencia Artificial ISSN:

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

The stages of event extraction

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Distant Supervised Relation Extraction with Wikipedia and Freebase

Cross-lingual Text Fragment Alignment using Divergence from Randomness

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Rule Learning with Negation: Issues Regarding Effectiveness

Combining Bidirectional Translation and Synonymy for Cross-Language Information Retrieval

Yoshida Honmachi, Sakyo-ku, Kyoto, Japan 1 Although the label set contains verb phrases, they

Greedy Decoding for Statistical Machine Translation in Almost Linear Time

Word Segmentation of Off-line Handwritten Documents

Initial approaches on Cross-Lingual Information Retrieval using Statistical Machine Translation on User Queries

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

Introduction. Beáta B. Megyesi. Uppsala University Department of Linguistics and Philology Introduction 1(48)

Indian Institute of Technology, Kanpur

Disambiguation of Thai Personal Name from Online News Articles

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Word Translation Disambiguation without Parallel Texts

Memory-based grammatical error correction

MADERA SCIENCE FAIR 2013 Grades 4 th 6 th Project due date: Tuesday, April 9, 8:15 am Parent Night: Tuesday, April 16, 6:00 8:00 pm

Reducing Features to Improve Bug Prediction

PIRLS. International Achievement in the Processes of Reading Comprehension Results from PIRLS 2001 in 35 Countries

Page 1 of 11. Curriculum Map: Grade 4 Math Course: Math 4 Sub-topic: General. Grade(s): None specified

Dublin City Schools Mathematics Graded Course of Study GRADE 4

A High-Quality Web Corpus of Czech

Finding Translations in Scanned Book Collections

A hybrid approach to translate Moroccan Arabic dialect

What the National Curriculum requires in reading at Y5 and Y6

MOODLE 2.0 GLOSSARY TUTORIALS

The RWTH Aachen University English-German and German-English Machine Translation System for WMT 2017

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Annotation Projection for Discourse Connectives

Semi-supervised Training for the Averaged Perceptron POS Tagger

An Introduction to Simio for Beginners

A Case Study: News Classification Based on Term Frequency

EACL th Conference of the European Chapter of the Association for Computational Linguistics. Proceedings of the 2nd International Workshop on

Ensemble Technique Utilization for Indonesian Dependency Parser

WP 2: Project Quality Assurance. Quality Manual

The Ups and Downs of Preposition Error Detection in ESL Writing

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

CS Machine Learning

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

Modeling function word errors in DNN-HMM based LVCSR systems

On the Combined Behavior of Autonomous Resource Management Agents

HOLIDAY LESSONS.com

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

GENERAL COMMENTS Some students performed well on the 2013 Tamil written examination. However, there were some who did not perform well.

Defragmenting Textual Data by Leveraging the Syntactic Structure of the English Language

Using dialogue context to improve parsing performance in dialogue systems

Matching Meaning for Cross-Language Information Retrieval

LANGUAGE IN INDIA Strength for Today and Bright Hope for Tomorrow Volume 11 : 12 December 2011 ISSN

Adding syntactic structure to bilingual terminology for improved domain adaptation

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Day 1 Note Catcher. Use this page to capture anything you d like to remember. May Public Consulting Group. All rights reserved.

Math Pathways Task Force Recommendations February Background

STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH

Proficiency Illusion

Using LibQUAL+ at Brown University and at the University of Connecticut Libraries

Improved Reordering for Shallow-n Grammar based Hierarchical Phrase-based Translation

The Task. A Guide for Tutors in the Rutgers Writing Centers Written and edited by Michael Goeller and Karen Kalteissen

Odyssey Writer Online Writing Tool for Students

Modeling function word errors in DNN-HMM based LVCSR systems

Transcription:

An English to Xitsonga statistical machine translation system for the government domain Cindy A. McKellar Centre for Text Technology, North-West University, Potchefstroom. Email: cindy.mckellar@nwu.ac.za Abstract Although a straightforward machine translation system can be trained with relatively little effort additional processing can often make a significant difference in the quality of the translated output. This article discusses several different sets of experiments aimed at improving the output of the English- Xitsonga machine translation system. These experiments include data cleaning, adding extra linguistic information to the training data and using placeholders to decrease data sparsity. All experiments were evaluated using the same evaluation set to ensure comparable results. Some of the methods attempted show promising results and it would be worthwhile to apply the same methods to other South African language pairs. Keywords Statistical machine translation; English; Xitsonga; Factored translation models; Language identification; Placeholders. I. INTRODUCTION In a multilingual environment like South Africa, machine translation is playing an increasingly important role in the distribution of information. This is especially the case in the government domain, where as much of the official documentation as possible should be available for all South Africans in the language of their choice. Machine translation can be used to speed up the work of the translators by creating a first draft of the translation which the human translator can then correct. The National Department of Arts and Culture (DAC) therefore launched the Autshumato project in 2007 1. The goal of this project was to create translation aids and resources for all the official languages. The initial project also included the creation of machine translation systems for three language pairs English to isizulu, English to Sepedi and English to Afrikaans within the public administration domain. An extension of the Autshumato project was launched in 2013 which included a machine translation system to translate from English to Xitsonga. The English to Xitsonga machine translation system, like the previous three machine translation systems, is based on statistical, phrase based machine translation. Although a 1 See http://autshumato.sourceforge.net/ for more information about the project straightforward machine translation system can be trained with relatively little effort (provided there is data available for the language pair) additional processing can often make a significant difference in the quality of the translated output. This article contains the results of several different experiments that were done in an attempt to improve the machine translation quality without adding more training data to the system. The first set of experiments made an attempt to "clean" the corpora by removing segments that did not look like useful data. The second set of experiments made use of a built in function of the machine translation toolkit to add additional linguistic information to the training data. The third experiment makes use of another built in function of the machine translation toolkit to replace certain words with placeholders in order to decrease data sparsity. The rest of this article is organised as follows: In Section 2 an overview of the experimental setup is given. This includes the training data, software, evaluation methods and the training of a baseline machine translation system. The following section (Section 3) discusses the data cleaning experiments and their results. Section 4 discusses the use of factored translation models to add additional information to the training data. The experiment on using placeholders to decrease data sparsity is in Section 5. The final section (Section 6) gives the conclusion and discusses some possibilities for future work. II. SETUP This section will give an overview of the setup used to train the English-Xitsonga machine translation systems. This includes available data, data processing, machine translation tools and evaluation methods. The development and evaluation of a baseline system is also discussed. A. Data Most of the available data for this language pair comes from translated documents in the government domain. Some documents were gathered from individual translators but a large portion of the data was gathered by crawling the government domain websites for documents that exist in both English and Xitsonga. English sentences were also selected from the Europarl corpus [1] and translated to Xitsonga to

increase the amount of data available. These translations were done by professional English-Xitsonga translators. Document pairs were first aligned by filename and then sentencised. Each document was then aligned on sentence level using the Hunalign sentence aligner [2]. To increase sentence alignment accuracy a bilingual English-Xitsonga dictionary was added to the Hunalign program. The aligned documents were then combined and tokenized to insert whitespace between words and punctuation. The final bilingual English- Xitsonga corpus contained 374 446 aligned segments. In addition to the aligned, bilingual corpus training a machine translation system also requires a monolingual corpus in the target language (Xitsonga) to use as training data for the language model. This corpus consists of the target language part of the bilingual corpus as well as additional monolingual Xitsonga data form the same sources. This data was also sentencised and tokenized before use. The monolingual corpus consists of 645 877 segments of Xitsonga text. B. Evaluation Machine translation can be evaluated either by humans or by automatic measures. Using human evaluation for a number of different machine translation systems can however be both time consuming and expensive. Therefore automatic evaluation is used in this paper. The Nist [3] and Bleu [4] automatic evaluation metrics are used to evaluate the quality of the machine translated text. Both evaluation metrics are based on the principle that a good machine translated text is one that shows a high degree of similarity to a human translation. The similarity between the translated text and the reference translation can be measured by the number of overlapping n-grams; the higher the number of overlapping n-grams, the better the quality of the machine translated text and the higher the Nist and Bleu scores. Bleu scores can range between 0 and 1. Nist scores do not have a clearly defined range, but a higher score is an indication of a better translation. Since any source language sentence can have multiple correct translations, even human translations seldom achieve the highest possible Nist and Bleu scores. For this reason it is advisable to use multiple, different human translations of the same source text as reference translations. This ensures that the automatic evaluation metrics give the most accurate results. For these experiments a test set of 480 sentences from the government domain is used for evaluation. There are 4 reference translations for the test set, each translated by a different, professional translator. C. Development of the baseline machine translation system In order to see if the experiments cause any improvement on the quality of the translation created by the machine translation system a baseline with no additional processing was trained to compare all the other systems with. The Moses open source statistical machine translation toolkit [5] was used to train the machine translation systems. The language models were trained using the IRST Language Modelling toolkit [6]. All language models used for the experiments discussed in this paper were trained using 4-grams. Past experiments have shown that 4-gram language models tend to achieve the best overall results, however this study did not experiment on other possible n-gram values. Phrase-based machine translation models were used. For the baseline system no addition processing other than what was discussed in the data section was done. The baseline machine translation system got a Nist score of 8.4351 and a Bleu score of 0.3694. III. DATA CLEANING Crawling the internet for data can result in very bad quality corpora. Many webpages contain multiple languages or lists of contact details. These types of data do not add any value to the machine translation system. The language mixing could even cause the output of the machine translation system to contain words in a language other than the target language. Two different cleaning methods were used in attempt to remove the "bad" data from the corpus. A. Language Identification The Language Identifier developed by Pienaar and Snyman [7] was used to check the corpus. The language identifier uses second generation spelling checkers to identify the language of a document or each individual sentence in a document. For this research the sentence level identification was used, as only those sentences that were identified as not belonging to the source or target language were removed. To identify the language of a sentence each word in the sentence is spellchecked by each of the 11 spelling checkers (one for each South African language). The percentage of words in each sentence that were correctly spelled is then calculated for each individual language. The sentence is marked as belonging to the language with the highest score. The language identifier classifies each sentence as certain or uncertain by using a benchmark percentage. All sentences with a chance of being a specific language that is higher than the benchmark are marked as Certain. The benchmark used by the program can be selected by the user. For this experiment the benchmark was set to 80. Therefore the program had to be more than 80% certain before marking a sentence as definitely belonging to a specific language. Both the monolingual and bilingual corpora were checked with the language identifier. Only sentences that were identified as being English in the source part of the bilingual corpus and Xitsonga in the target part and monolingual corpus were kept. Table 1 shows the sentence counts of the monolingual and bilingual corpora before and after the language identification step. B. Named Entity Recognition The web crawled data also contains a lot of named entities, where there is nothing but the name in the line. These segment pairs do not contribute any useful information to the machine translation system and can be removed with little effect on the output quality.

The Autshumato Text Anonymizer [8] was used to identify the named entities in the text. The Anonymizer uses a combination of rules expressed as regular expressions and language specific wordlists to identify named entities. The named entities were tagged and all lines from both corpora containing only a named entity were removed. If there was any additional text in the line with the named entity, the sentence remained in the corpus. There were also some cases were one of the pair of bilingual sentences consisted only of a named entity but the other contained additional text, these sentences were also removed. After the sentences were removed, the remaining tags in the data were stripped as they were not needed during the training of the machine translation system. C. Evaluation Three machine translation systems were trained using the cleaned corpora. One using only the language identified data, one using the data with the named entities removed and one using a combination of both methods. Table 1 gives the segment counts for each of the corpora. The evaluation results are in Table 2. TABLE I. DATA COUNTS Data counts before and after cleaning Type of cleaning applied Monolingual Bilingual Without data cleaning 645 877 374 446 Language Identifier 528 579 312 474 Autshumato Annonimiser 622 193 363 607 Combination of Language Identifier and Annonimiser TABLE II. 528 332 312 394 EVALUATION RESULTS: DATA CLEANING Evaluation results after cleaning data Type of cleaning applied Nist Bleu Language Identifier 8.5033 0.3683 Autshumato Annonimiser 8.4688 0.3699 Combination of Language Identifier and Annonimiser 8.4485 0.3631 The data cleaning caused the Nist score to rise in all three models. The Blue score however only shows a small rise in the model trained on the data with the named entities removed. Although the evaluation does not show a large improvement in the evaluation metrics, the removal of parts of the data did not have a big negative effect either. This could indicate that the data that was removed did not really contribute to the translation of the evaluation data. IV. FACTORED MACHINE TRANSLATION MODELS Factored translation models offer the ability to add additional linguistic information to the data being used to train a machine translation model. This information can include lemmas, part-of-speech tags, gender classes or other morphological information. Adding such information to the training data may enable the resulting machine translation system to do a better job of translating data. This can be especially useful in languages where there isn t a large amount of training data available. In the rest of this section two experiments to add extra information to the training data will be discussed. A. Data annotation To train factored translation models both the monolingual and bilingual data needs to be annotated with the additional information needed. For these experiments lemmas and part-ofspeech tags were used. The evaluation data also needs to be annotated with the same additional factors. The English part of the bilingual corpus, as well as the English evaluation data, was annotated with the Treetagger [9]. The Treetagger is a language independent lemmatiser and partof-speech tagger. The English model distributed with the program was used to annotate the data. The Xitsonga part of the bilingual data, as well as the monolingual Xitsonga corpus, was annotated with the core technologies developed as part of the NCHLT Text project [10]. The Xitsonga lemmatiser developed in this project uses language specific normalization rules to lemmatise text. The part-of-speech tagger was trained using the HunPoS open source Hidden Markov Model tagger. B. Data alignment An important step in the training of a machine translation system is word alignment. During the word alignment phase each pair of aligned sentences are further aligned on word level. The Moses machine translation toolkit uses Giza++ [11] for the word alignment. As Giza++ uses statistics, the more data is available the better the word alignments will be. Better word alignments in turn will lead to more accurate phrase tables and better translations. Unfortunately the language pair English-Xitsonga does not have a huge corpus available. This data sparsity can lead to sub-optimal word alignment. One possible solution for this problem is to do word alignment on the lemmas of words instead of their surface forms. The factored models allow the lemma of each word to be added to the surface form as an additional factor in the training data. During the training of the machine translation system the word alignment is then done on the lemmas instead of the surface forms. But the translation table still uses the surface forms, so any text will be translated normally. To test the effect of using lemmas for word alignment a machine translation system was trained with the annotated corpus. The language model from the baseline system was used, as this particular factored model does not require a special language model. The word alignment was done using the factors containing the lemmas. All other setting were kept the same as during the baseline training. The results of this system can be seen in Table 3.

C. Sentence reordering The machine translation system uses a language model trained on the monolingual, target language corpus to determine how a correct sentence in the target language should look. However, due to the number of possible combinations of words that can exists in any language it is nearly impossible to gather a corpus large enough to contain all the examples needed to correctly order the words in the output. Adding part-of-speech tags to the data and training a language model on these tags can allow for better word reordering in the output of the machine translation system. The machine translation system can use both a language model trained on the surface forms and a language model trained on the part-of-speech data. This ensures that, in the case of insufficient data in the surface form model, the part-of-speech tag model can be used to guess at a good word order for a sentence. Two language models are needed for the part-of-speech reordering model. The first, trained on the surface forms of the corpus, is the same as was used in training the baseline machine translation system. A second version of the monolingual corpus was created using only the part-of-speech tags of the words. A new language model was trained on this data. Both language models were used to train the machine translation system. The machine translation system was trained to translate from the surface form and part-of-speech tag of the source language to the surface form and part-of-speech tag in the target language. Both the surface form of the output and the part-of-speech tag was then used to ensure the best possible word order of the output sentence. D. Evaluation The additional data used in the training of the factored machine translation systems was also added to the evaluation data. Each of the factored translation models were then evaluated with the factored test set. As the output of these models does not contain any factors the normal reference set was used to calculate the Nist and Bleu scores. The results of the evaluation can be seen in Table 3. TABLE III. EVALUATION RESULTS: FACTORED MODELS Evaluation Results of the Factored machine translation systems Factored model type Nist Bleu Word alignment with lemmas 8.5534 0.3720 Word reordering with part of seech tags Combination of lemmas and part of speech tags 8.5126 0.3675 8.5641 0.3731 As can be seen in the results table (Table 3) adding factors to the training data can have a beneficial outcome on the machine translation systems. Adding lemmas to use during word alignment caused both the Nist and Bleu scores to go up. The part of speech tags however only increase the Nist score. Combining both methods led to the greatest increase in both the Nist and Bleu scores. Although these results seem to indicate that adding factors to the training data improves the machine translation output, to ensure that the changes in the automatic evaluation metrics truly reflect a positive change in the translation quality human evaluation will be needed. V. PLACEHOLDERS Many numbers occur only a few times in a corpus. This leads to sparse data and makes it difficult for reliable statistic to be calculated during the training of the machine translation system. Replacing these numbers with a placeholder symbol can reduce the data sparsity. Instead of thousands of unique numbers in the corpus leading to very sparse phrases there is only a single placeholder symbol. One of the advanced features of the Moses toolkit is the ability to train machine translation systems with placeholders. Any type of word can be replaced with a placeholder. The placeholders need to be added in both the bilingual data and the monolingual language model data. For this experiment only numbers were replaced with placeholders. A pre-processing script was used to replace all the numbers (digit format only) with the placeholder in both monolingual and bilingual data. The language model was then trained as usual using the monolingual corpus with the placeholders. The machine translation system training is constrained to only create phrase table entries were placeholders are aligned 1-1. This is to ensure that no extra placeholders will be added during translation and none of the placeholders will fall away. This means that each of the placeholders can be replaced with a number again after the text has been translated. To make the machine translation system usable, the placeholders need to be replaced with the original numbers after translation. To enable this ability the text that needs to be translated must also have placeholders where the numbers are. But to ensure that the correct value of each placeholder is added to the sentence after translation the original numbers are added to the placeholder as xml mark-up. When translating a text containing placeholders the Moses toolkit changes the input into factored input where the placeholder is seen as the surface form that is used during translation and the actual number is stored as an additional factor attached to the placeholder. After each sentence has been translated, the placeholders are automatically replaced with the original numbers. As the numbers are placed back into the sentences there is no need to change anything in the reference translations used to evaluate the placeholder system. The results of the placeholder experiment can be seen in Table 4. TABLE IV. EVALUATION RESULTS: PLACEHOLDERS Evaluation Results of the Placeholder machine translation systems Model type Nist Bleu Placehodlers 8.4884 0.3680 The addition of placeholders to the data seems as though it should improve the translation quality both by ensuring more

reliable statistics and by making sure that all the numbers that were in the input are also in the output. However while the Nist score does get better the Bleu score is worse than the baseline. Further experiments with the placeholder function may yield better results but the initial work discussed here, while not as good as hoped for, is not totally discouraging either. VI. CONCLUSION In this paper we discussed several different experiments that aimed to improve the quality of machine translation output. For most of the South African languages there is little data available, especially the bilingual data needed to train machine translation systems. This is why it is extremely important that we explore methods that could potentially increase the usability of a machine translation system's output without requiring additional training data. None of the methods tested in this paper caused huge increases in the Nist and Bleu scores but all of them did show promise. The next step for the English-Xitsonga machine translation system could be to combine the different methods into one single system. This may yield even better results. Another important step in the continuation of this research is the inclusion of human evaluation. Automatic evaluation metrics are extremely useful to give a quick indication of the effects of a new approach but to get a true view of the usability of the resulting system human evaluation is critical. It would also be really interesting to apply these same methods to other South-African language pairs to see if they also improve. REFERENCES [2] D. Varga, L. Németh, P. Halácsy, A. Kornai, V. Trón, V. Nagy, "Parallel corpora for medium density languages" in Proceedings of the RANLP, Borovets, Bulgaria, September 2005, pp 590-596. [3] G. Doddington, Automatic Evaluation of Machine Translation Quality using N-gram Co-occurence Statistics, Proceedings of the 2nd International Conference on Human Language Technology Research, pp 138-145, San Diego, California, 2002. [4] K. Papineni, S. Roukos, T. Ward and W. J. Zhu, BLEU: A Method for Automatic Evaluation of Machine Translation, Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, pp 311-318, Philadelphia, USA, 2002. [5] P. Koehn et al., Moses: Open source toolkit for statistical machine translation. Proceedings of the ACL demo and poster sessions, Prague, Czech Republic, 2007, pp. 177-180. [6] N. B. M. Federico and M. Cettolo, Irstlm: an open source toolkit for handling large scale language models, in Proceedings of Interspeech, Brisbane, September 2008. [7] W. Pienaar and D. Snyman, Spelling Checker-based Language Identification for the Eleven Official South African Languages Proceedings of the 21 st annual Symposium of the Pattern Recognition Association of South Africa (PRASA), Stellenbosch, South African, 2010. [8] H. Groenewald and L. D. Plooy, Processing parallel text corpora for three South African language pairs in the Autshumato project, in Proceedings of the Second Workshop on African Language Technology, Malta, May 2010, pp. 27 30. [9] H. Schmid, Probabilistic part-of-speech tagging using decision trees, in International Conference on New Methods in Language Processing, Manchester, 1994. [10] R. Eiselen and M. Puttkammer, Developing text resources for ten South African languages, in Proceedings of LREC, Reykjavik: Iceland, May 2014. [11] F. Och and H. Ney, A systematic comparison of various statistical alignment models, Computational Linguistics, vol. 29, no. 1, pp. 19 51, March 2003. [1] P. Koehn, Europarl: a parallel corpus for statistical machine translation, in Proceedings of the Tenth Machine Translation Summit, Phuket: Asia-Pacific, September 2005, pp. 79 86.