Available online at ScienceDirect. Procedia Technology 18 (2014 ) Krzysztof Wołk, Krzysztof Marasek

Similar documents
Linking Task: Identifying authors and book titles in verbose queries

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

arxiv: v1 [cs.cl] 2 Apr 2017

Noisy SMS Machine Translation in Low-Density Languages

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Language Model and Grammar Extraction Variation in Machine Translation

Procedia - Social and Behavioral Sciences 141 ( 2014 ) WCLTA Using Corpus Linguistics in the Development of Writing

ScienceDirect. Malayalam question answering system

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Cross Language Information Retrieval

The NICT Translation System for IWSLT 2012

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

Constructing Parallel Corpus from Movie Subtitles

AQUA: An Ontology-Driven Question Answering System

Procedia - Social and Behavioral Sciences 143 ( 2014 ) CY-ICER Teacher intervention in the process of L2 writing acquisition

Using dialogue context to improve parsing performance in dialogue systems

Domain Adaptation in Statistical Machine Translation of User-Forum Data using Component-Level Mixture Modelling

Language Independent Passage Retrieval for Question Answering

Detecting English-French Cognates Using Orthographic Edit Distance

Word Segmentation of Off-line Handwritten Documents

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Procedia - Social and Behavioral Sciences 154 ( 2014 )

The KIT-LIMSI Translation System for WMT 2014

Cross-lingual Text Fragment Alignment using Divergence from Randomness

Memory-based grammatical error correction

Re-evaluating the Role of Bleu in Machine Translation Research

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment

The role of the first language in foreign language learning. Paul Nation. The role of the first language in foreign language learning

A heuristic framework for pivot-based bilingual dictionary induction

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Matching Similarity for Keyword-Based Clustering

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

Developing Grammar in Context

South Carolina English Language Arts

NCU IISR English-Korean and English-Chinese Named Entity Transliteration Using Different Grapheme Segmentation Approaches

METHODS FOR EXTRACTING AND CLASSIFYING PAIRS OF COGNATES AND FALSE FRIENDS

The RWTH Aachen University English-German and German-English Machine Translation System for WMT 2017

Speech Recognition at ICSI: Broadcast News and beyond

1.11 I Know What Do You Know?

A Case Study: News Classification Based on Term Frequency

Using Web Searches on Important Words to Create Background Sets for LSI Classification

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

Cross-Lingual Text Categorization

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Finding Translations in Scanned Book Collections

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Page 1 of 11. Curriculum Map: Grade 4 Math Course: Math 4 Sub-topic: General. Grade(s): None specified

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

On document relevance and lexical cohesion between query terms

The Effect of Extensive Reading on Developing the Grammatical. Accuracy of the EFL Freshmen at Al Al-Bayt University

The Ups and Downs of Preposition Error Detection in ESL Writing

Procedia - Social and Behavioral Sciences 197 ( 2015 )

Providing student writers with pre-text feedback

Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology. Michael L. Connell University of Houston - Downtown

SIE: Speech Enabled Interface for E-Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Procedia - Social and Behavioral Sciences 146 ( 2014 )

Probabilistic Latent Semantic Analysis

Overview of the 3rd Workshop on Asian Translation

Introduction, Organization Overview of NLP, Main Issues

Task Tolerance of MT Output in Integrated Text Processes

Procedia - Social and Behavioral Sciences 191 ( 2015 ) WCES Why Do Students Choose To Study Information And Communications Technology?

Modeling function word errors in DNN-HMM based LVCSR systems

LANGUAGE IN INDIA Strength for Today and Bright Hope for Tomorrow Volume 11 : 12 December 2011 ISSN

Distant Supervised Relation Extraction with Wikipedia and Freebase

Rule Learning With Negation: Issues Regarding Effectiveness

Taxonomy of the cognitive domain: An example of architectural education program

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

DOES RETELLING TECHNIQUE IMPROVE SPEAKING FLUENCY?

Trend Survey on Japanese Natural Language Processing Studies over the Last Decade

CEFR Overall Illustrative English Proficiency Scales

Using Moodle in ESOL Writing Classes

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Writing a composition

Parallel Evaluation in Stratal OT * Adam Baker University of Arizona

CS Machine Learning

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Multilingual Sentiment and Subjectivity Analysis

International Conference on Education and Educational Psychology (ICEEPSY 2012)

Software Maintenance

Rule Learning with Negation: Issues Regarding Effectiveness

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

ScienceDirect. Noorminshah A Iahad a *, Marva Mirabolghasemi a, Noorfa Haszlinna Mustaffa a, Muhammad Shafie Abd. Latif a, Yahya Buntat b

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Procedia - Social and Behavioral Sciences 209 ( 2015 )

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Procedia - Social and Behavioral Sciences 180 ( 2015 )

Evidence for Reliability, Validity and Learning Effectiveness

How to Judge the Quality of an Objective Classroom Test

Loughton School s curriculum evening. 28 th February 2017

Author: Justyna Kowalczys Stowarzyszenie Angielski w Medycynie (PL) Feb 2015

The taming of the data:

Clickthrough-Based Translation Models for Web Search: from Word Models to Phrase Models

Mandarin Lexical Tone Recognition: The Gating Paradigm

Transcription:

Available online at www.sciencedirect.com ScienceDirect Procedia Technology 18 (2014 ) 126 132 International workshop on Innovations in Information and Communication Science and Technology, IICST 2014, 3-5 September 2014, Warsaw, Poland Building subject-aligned comparable corpora and mining it for truly parallel sentence pairs Krzysztof Wołk, Krzysztof Marasek Polish Japanese Institute of Information Technology, Warsaw, Poland kwolk@pjwstk.edu.pl Abstract Parallel sentences are a relatively scarce but extremely useful resource for many applications including cross-lingual retrieval and statistical machine translation. This research explores our methodology for mining such data from previously obtained comparable corpora. The task is highly practical since non-parallel multilingual data exist in far greater quantities than parallel corpora, but parallel sentences are a much more useful resource. Here we propose a web crawling method for building subjectaligned comparable corpora from Wikipedia articles. We also introduce a method for extracting truly parallel sentences that are filtered out from noisy or just comparable sentence pairs. We describe our implementation of a specialized tool for this task as well as training and adaption of a machine translation system that supplies our filter with additional information about the similarity of comparable sentence pairs. 2014 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/3.0/). Peer-review under responsibility of the Scientific Committee of IICST 2014. Peer-review under responsibility of the Scientific Committee of IICST 2014 Keywords: Comparable corpora, machine translation, NLP 1. Introduction Parallel sentences are an invaluable information resource especially for machine translation systems as well as for other cross-lingual information-dependent tasks. Unfortunately such data is quite rare, especially for the Polish English language pair. On the other hand, monolingual data for those languages is accessible in far greater quantities. We can classify the similarity of data as four main corpora types. Most rare parallel corpora can be defined as corpora that contain translations of the same document into two or more languages. Such data should be 2212-0173 2014 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/3.0/). Peer-review under responsibility of the Scientific Committee of IICST 2014 doi:10.1016/j.protcy.2014.11.024

Krzysztof Wołk and Krzysztof Marasek / Procedia Technology 18 ( 2014 ) 126 132 127 aligned at least at the sentence level. A noisy-parallel corpus contains bilingual sentences that are not perfectly aligned, or has poor quality translations. Nevertheless mostly bilingual translation of a specific document should be present in it. A comparable corpus is built from non-sentence-aligned and not-translated bilingual documents, but the documents should be topic-aligned. A quasi-comparable corpus includes very heterogeneous and very non-parallel bilingual documents that can, but don t have to, be topic-aligned [1]. In this article we present a methodology that allows us to obtain truly parallel corpora from non-sentence-aligned data sources, such as noisy-parallel or comparable corpora. For this purpose we used a set of specialized tools for obtaining, aligning, extracting and filtering text data, combined together into a pipeline that allows us to complete the task. We present the results of our initial experiments based on randomly selected text samples from Wikipedia. We chose Wikipedia as a source of data because of a large number of documents that it provides (1,047,423 articles on PL Wiki and 4,524,017 on EN, at the time of writing this article). Furthermore, Wikipedia contains not only comparable documents, but also some documents that are translations of each other. The quality of our approach is compared to human evaluation *. The solution can be divided into three main steps. First the data is collected, then it is aligned, and lastly the results of the alignment are filtered. The last two steps are not trivial because of the disparities between Wikipedia documents. Based on the Wikipedia statistics we know that an average article on PL Wiki contains about 379 words, whereas on EN Wiki it is 590 words. This is most likely why sentences in the raw Wiki corpus are mostly misaligned, with translation lines whose placement does not correspond to any text lines in the source language. Moreover, some sentences may have no corresponding translation in the corpus at all. The corpus might also contain poor or indirect translations, making the alignment difficult. Thus, alignment is crucial for accuracy. Sentence alignment must also be computationally feasible in order to be of practical use in various applications. The Polish language presents a particular challenge to the application of such tools. It is a complicated West- Slavic language with complex elements and grammatical rules. In addition, the Polish language has a large vocabulary due to many endings and prefixes changed by word declension. These characteristics have a significant impact on the data and data structure requirements. In addition, English is a position-sensitive language. The syntactic order (the order of words in a sentence) plays a very significant role, and the language has very limited inflection of words (due to the lack of declension endings). The word position in an English sentence is often the only indicator of the meaning. The sentence order follows the Subject-Verb-Object (SVO) schema, with the subject phrase preceding the predicate. On the other hand, no specific word order is imposed in Polish, and the word order has little effect on the meaning of a sentence. The same thought can be expressed in several ways. For example, the sentence I bought myself a new car. can be written in Polish as one of the following: Kupiłem sobie nowy samochód ; Nowy samochód sobie kupiłem. ; Sobie kupiłem nowy samochód. ; Samochód nowy sobie kupiłem.. It must be noted that such differences exist in many language pairs and need to be dealt with in some way [2]. 2. The pipeline Our procedure starts with a specialized web crawler. Because PL Wiki contains less data and almost all articles have their correspondence on EN Wiki, the program crawls data starting from the non-english site first. It is a language independent solution. The crawler can obtain and save bilingual articles in any language supported by Wikipedia. Fig. 1. Pipeline * http://en.wikipedia.org/wiki/wikipedia:statistics

128 Krzysztof Wołk and Krzysztof Marasek / Procedia Technology 18 ( 2014 ) 126 132 First the data is saved in HTML files and then it is topic-aligned. In order to narrow the search field to specific indomain documents, it is necessary to give the crawler the first link to the article in the domain and then the program will automatically obtain other topic-related documents. Narrowing the search domain not only helps to adjust the output to the specific needs it also narrows the vocabulary, which makes the aligning task easier. After obtaining HTML documents, the crawler extracts plain text from them and cleans the data. Tables, URL s, figures, pictures, menus, references and other unnecessary data are removed. Finally, bilingual documents are tagged with a unique ID as a topic-aligned comparable corpus. We propose a two-level sentence alignment method that prepares a dictionary for itself. The Hunalign tool is used first to match bilingual sentences. Its input is tokenized and sentence-segmented. In the presence of a dictionary, Hunalign combines the dictionary information with the Gale-Church sentence-length information. In the absence of a dictionary, it first falls back to the sentence-length information, and then builds an automatic dictionary based on this alignment. Then it realigns the text in a second pass, using the automatic dictionary. The option without a dictionary is the one we used [3]. Like most sentence aligners, Hunalign does not deal well with changes in the sentence order. It is unable to come up with crossing alignments, i.e., segments A and B in one language corresponding to segments B A in the other language. In order to cope with this problem, and to filter out bad or poor bilingual sentence pairs, we implemented a special tool [4]. 2.1. Filtering strategy Our strategy is to find a correct translation of each Polish line using any translation engine. We translate all lines of the Polish file (src.pl) with a translator and put each line s translation in an intermediate English translation file (src.trans). This intermediate translation helps us find the correct line in the English translation file (src.en) and put it in the correct position or remove incorrect pairs from the corpora. There are additional complexities that must be addressed. Comparing the src.trans lines with the src.en lines is not easy, and it becomes harder when we want to use the similarity rate to choose the correct, real-world translation. Fig. 2. Filtering There are many strategies to compare two sentences. We can split each sentence into its words and find the number of words in both sentences. However, this approach has some problems. For example, let us compare It is origami. to these sentences: The common theme what makes it origami is folding is how we create the form, and This is origami. With this strategy, the first sentence is more similar because it contains all 3 words. However, it is clear that the second sentence is the correct choice. We can solve this problem by dividing the number of words in both sentences

Krzysztof Wołk and Krzysztof Marasek / Procedia Technology 18 ( 2014 ) 126 132 129 by the number of total words in the sentences. However, counting stop words in the intersection of sentences sometimes causes incorrect results. So, we remove these words before comparing two sentences. Another problem is that sometimes we find stemmed words in sentences, for example boy and boys. Despite the fact that these two words should be counted as similarity of two sentences, with this strategy, these words are not counted. The next comparison problem is the word order in sentences. There are other ways for comparing strings that are better than counting intersection lengths. For example, we can find matching blocks in the strings "abxcd" and "abcd". Our function can count ratio and divide the length of matching blocks by the length of two strings, and return a measure of the sequences similarity as a float value in the range [0, 1]. This measure is 2.0*M / T, where T is the total number of elements in both sequences, and M is the number of matches. Using this function to compare strings instead of counting similar words helps us solve the problem of the similarity of boy and boys. It also solves the problem of considering the position of words in sentences. Another problem in comparing lines is synonyms. For example, these two sentences: I will call you tomorrow, and I would call you tomorrow. We used the NLTK Python module and WordNet to find synonyms for each word and use these synonyms in comparing sentences. Using synonyms for each word, we created multiple sentences from each original sentence and compared them as a many-to-many relation. To obtain the best results, our script provides users with the ability to have multiple functions with multiple acceptance rates. Fast functions with lower quality results are tested first. If they can find results with a very high acceptance rate, we accept their selection. If the acceptance rate is not sufficient, we use slower but higher accuracy functions [5]. 2.2. Wikipedia Machine Translation Engine The filtering tool, which is the most important part of the entire process, is dependent on the translation engine. It is possible to use online engines for general use, but better results can be obtained with specialized translation systems. We obtained all PL-EN parallel data from various domains from the OPUS project and used it for training a specialized machine translation system. To improve its performance, we conducted the system s adaptation to Wikipedia using a dump of all English articles as a language model. The final training corpora counted 36,751,049 sentences and the language model counted 79,424,211 sentences. The unique word forms count was 3,209,295 in the Polish side of the corpora, 1,991,418 in the English side and 37,702,319 in the language model. Implementation of the translation system included many steps. Processing of the corpora was accomplished, including tokenization, cleaning, factorization, lowercasing, splitting, and a final cleaning after splitting. Training data was processed and the language model was developed. Tuning was performed as well [13]. The training was done using the Moses open source SMT toolkit with its Experiment Management System (EMS) [6]. The SRI Language Modeling Toolkit (SRILM) [7] with an interpolated version of the Kneser-Key discounting (interpolate unk kndiscount) was used for the 6-gram language model training. We used the MGIZA++ tool for word and phrase alignment. KenLM [8] was used to binarize the language model, with a lexical reordering set to use the msd-bidirectional-fe model. Reordering probabilities of phrases were conditioned on lexical values of a phrase. It considers three different orientation types on source and target phrases like monotone(m), swap(s) and discontinuous(d). The bidirectional reordering model adds probabilities of possible mutual positions of source counterparts to the current and following phrases. Probability distribution to a foreign phrase is determined by f and to the English phrase by e [9,10]. MGIZA++ is a multi-threaded version of the well-known GIZA++ tool [11]. The symmetrization method was set to grow-diag-final-and for word alignment processing. First, two-way direction alignments obtained from GIZA++ were intersected, so only the alignment points that occurred in both alignments remained. In the second phase, additional alignment points existing in their union were added. The growing step adds potential alignment points of unaligned words and neighbours. Neighbourhood can be set directly to left, right, top or bottom, as well as to diagonal (grow-diag). In the final step, alignment points between words from which at least one is unaligned are added (grow-diag-final). If the grow-diag-final-and method is used, an alignment point between two unaligned words appears [12].

130 Krzysztof Wołk and Krzysztof Marasek / Procedia Technology 18 ( 2014 ) 126 132 2.2.1. MT Evaluation Metrics are necessary to measure the quality of translations produced by the SMT systems. For this purpose, various automated metrics are available to compare SMT translations to high quality human translations. Since each human translator produces a translation with different word choices and orders, the best metrics measure SMT output against multiple reference human translations. Among the commonly used SMT metrics are: Bilingual Evaluation Understudy (BLEU), the U.S. National Institute of Standards & Technology (NIST) metric, the Metric for Evaluation of Translation with Explicit Ordering (METEOR), Translation Error Rate (TER). BLEU was one of the first metrics to demonstrate a high correlation with reference human translations. The general approach for BLEU, as described in [14], is to attempt to match variable length phrases to reference translations. Weighted averages of the matches are then used to calculate the metric. The NIST metric seeks to improve the BLEU metric by valuing information content in several ways. It takes the arithmetic versus geometric mean of the n-gram matches to reward good translation of rare words. The NIST metric also gives heavier weights to rare words. Lastly, it reduces the brevity penalty when there is a smaller variation in the translation length. The METEOR metric, developed by the Language Technologies Institute of Carnegie Mellon University, is also intended to improve the BLEU metric. We used it without synonym and paraphrase matches for Polish. METEOR rewards recall by modifying the BLEU brevity penalty, takes into account higher order n-grams to reward matches in a word order, and uses arithmetic vice geometric averaging. For multiple reference translations, it reports the best score for word-to-word matches. TER is one of the most recent and intuitive SMT metrics developed. This metric determines the minimum number of human edits required for an SMT translation to match a reference translation in meaning and fluency. Required human edits might include inserting words, deleting words, substituting words, and changing the order or words or phrases. For the evaluation, we randomly selected 1000 parallel sentences from Wikipedia documents. None of those sentences were included inside the training data on our system. Table 1 presents the evaluation of translation quality in comparison to general use online translation engines. Table 1. MT Results BLEU NIST METEOR TER Google 18,15 5,22 48,86 70,23 Bing 18,87 5,27 48,80 70,61 Our SMT 20,51 5,31 49,23 69,11 3. Experiments and mining evaluation To evaluate quality and quantity of parallel data, extracted automatically from comparable corpora, we randomly selected 20 bilingual documents from Wikipedia. Some of them differed greatly in respect to vocabulary, text amounts and parallelism. We asked human translators to manually align those articles on the sentence level. The information about the human translators is presented in Table 2. In the Vocab Count column we present the number of distinct words and their forms, in Sentences the number of recognized sentences in each language, and finally the number of sentence pairs aligned by a human. Table 2. Human Alignment No. Vocab.Count Sentences Human Aligned No. Vocab.Count Sentences Human Aligned PL EN PL EN PL EN PL EN 1 2526 1910 324 186 127 11 2861 2064 412 207 8 2 2950 3664 417 596 6 12 2186 1652 345 188 2 3 504 439 45 43 34 13 2799 3418 496 472 124 4 2529 1383 352 218 4 14 1164 1037 184 196 3 5 807 1666 104 275 10 15 2465 1781 365 189 3

Krzysztof Wołk and Krzysztof Marasek / Procedia Technology 18 ( 2014 ) 126 132 131 6 2461 4667 368 618 1 16 1946 1839 282 198 132 7 2701 1374 560 210 16 17 966 782 113 96 7 8 1647 768 274 78 1 18 2005 1253 309 134 1 9 1189 1247 121 120 64 19 2443 2001 251 189 21 10 2933 3047 296 400 150 20 9888 3181 1297 367 2 The same articles were processed with our pipeline. In Table 3 we present how many sentences Hunalign initially aligned as similar, and how many of them remained after filtering with our tool. Both columns YES and NO under the Hunaligned section are aligned sentences, the numbers represent how many of them were aligned correctly and how many by mistake. In the Filtered column we present the number of parallel sentences that remained after filtering, in YES we show properly-aligned sentences and in NO mistaken ones. In this scenario, we also asked a human translator to check which of the remaining sentence pairs were truly parallel and if any pairs were missed out. Table 3. Automatic Alignment No. Hunaligned Filtered No. Hualigned Filtered YES NO YES NO YES NO Yes No 1 109 130 18 0 11 8 325 0 0 2 6 527 25 2 12 2 256 0 0 3 17 24 0 0 13 70 414 1 0 4 4 302 1 0 14 3 182 0 0 5 6 211 1 0 15 3 285 0 0 6 1 498 0 0 16 111 108 0 0 7 16 440 0 0 17 7 98 0 0 8 1 221 0 0 18 1 202 0 0 9 51 62 0 0 19 21 192 0 0 10 127 245 0 0 20 2 1078 0 0 4. Conclusions and future work We introduced a new method for obtaining, mining and filtering very parallel bilingual sentence pairs from noisy-parallel and comparable corpora. Nowadays, the bi-sentence extraction task is becoming more and more popular in unsupervised learning for numerous specific tasks. The method overcomes disparities between English and Polish or any other West-Slavic language. It is a language-independent method that can easily be adjusted to a new environment, and it only requires parallel corpora for initial training. The experiments show that the method provides good accuracy and some correlation with human judgements. That is what should be expected from the task of mining from comparable data. From a practical standpoint, the method neither requires expensive training nor requires language-specific grammatical resources, while producing satisfying results. Nevertheless, there is still some room for improvement in two areas. In the presented experiments the amount of obtained data in comparison with human work is not satisfactory. The first one is Hunalign, which would perform much better if it was provided a good quality dictionary, especially one that contains in-domain vocabulary. The second one is the statistical machine translation system (SMT), which would greatly increase quality by providing better translations. After the initial mining of the corpora, the obtained parallel data can possibly be used for both purposes. Firstly, a phrase-table can be trained from extracted bi-sentences and from it we can easily extract a good in-domain dictionary (also including probabilities of translations). Secondly, the SMT can be retrained with newly mined data and adapted based on it [15]. Lastly, the pipeline can be re-run with new capabilities. The steps can be repeated until the extraction results are fully satisfactory. Acknowledgements This work was supported by the European Community from the European Social Fund within the Interkadra project UDA-POKL-04.01.01-00-014/10-00 and Eu-Bridge 7th FR EU project (Grant Agreement No. 287658).

132 Krzysztof Wołk and Krzysztof Marasek / Procedia Technology 18 ( 2014 ) 126 132 References [1] Wu D., Fung P.; Inversion Transduction Grammar Constraints for Mining Parallel Sentences from Quasi-Comparable Corpora ; Natural Language Processing IJCNLP 2005; Lecture Notes in Computer Science Volume 3651, 2005, pp 257-268 [2] Wołk K., Marasek K., Polish English Speech Statistical Machine Translation Systems for the IWSLT 2013., Proceedings of the 10th International Workshop on Spoken Language Translation, Heidelberg, Germany, p. 113-119, 2013 [3] Varga D., Németh L., Halácsy P., Kornai A., Trón V., Nagy V.; Parallel corpora for medium density languages ; In Proceedings of the RANLP 2005, p. 590-596. [4] Wołk K., Marasek K., A Sentence Meaning Based Alignment Method for Parallel Text Corpora Preparation, Advances in Intelligent Systems and Computing volume 275, p.107-114, Publisher: Springer, ISSN 2194-5357, ISBN 978-3-319-05950-1, Madeira Island, Portugal, 2014 [5] http://korpusy.s16874487.onlinehome-server.info/ [6] Snover, M., Dorr, B., Schwartz, R., Micciulla, L., and Makhoul, J., A Study of Translation Edit Rate with Targeted Human Annotation, Proc. of 7th Conference of the Assoc. for Machine Translation in the Americas, Cambridge, August 2006. [7] Koehn, P. et al., Moses: Open Source Toolkit for Statistical Machine Translation, Annual Meeting of the Association for Computational Linguistics (ACL) demonstration session, Prague, June 2007. [8] http://www.cs.jhu.edu/~ccb/publications/iwslt05- report.pdf [9] Heafield, K. "KenLM: Faster and smaller language model queries", Proc. of Sixth Workshop on Statistical Machine Translation, Association for Computational Linguistics, 2011. [10] Marta R. Costa-jussa, Jose R. Fonollosa, Using linear interpolation and weighted reordering hypotheses in the Moses system, Barcelona, Spain, 2010 [11] Stolcke, A., SRILM An Extensible Language Modeling Toolkit, INTERSPEECH, 2002. [12] Gao, Q. and Vogel, S., Parallel Implementations of Word Alignment Tool, Software Engineering, Testing, and Quality Assurance for Natural Language Processing, pp. 49-57, June 2008. [13] Tiedemann J., Parallel Data, Tools and Interfaces in OPUS. ; In Proceedings of the 8th International Conference on Language Resources and Evaluation (LREC 2012). [14] Wołk K., Marasek K., Real-Time Statistical Speech Translation", Advances in Intelligent Systems and Computing volume 275, p.107-114, Publisher: Springer, ISSN 2194-5357, ISBN 978-3-319-05950-1, Madeira Island, Portugal, 2014 [15] Durrani N., Haddow B., Heafield K., Koehn P.; Edinburgh's Machine Translation Systems for European Language Pairs, Proceedings of the Eighth Workshop on Statistical Machine Translation, 2013