UCSG Shallow Parsing: Optimum Chunk Sequence Selection

Size: px
Start display at page:

Download "UCSG Shallow Parsing: Optimum Chunk Sequence Selection"

Transcription

1 UCSG Shallow Parsing: Optimum Chunk Sequence Selection B Hanumantha Rao and Kavi Narayana Murthy Department of Computer and Information Sciences, University of Hyderabad, Hyderabad, India. hanu 499@yahoo.com,knmuh@yahoo.com Abstract This paper is about syntactic analysis of natural language sentences. The focus is on wide coverage partial parsing architectures. In this work we enhance and enrich the UCSG shallow parsing architecture being developed here over the last many years. UCSG architecture combines linguistic grammars in the form of Finite State Machines for recognising all potential chunks and HMMs to rate and rank the chunks so produced. Here we have explored Mutual Information statistics for rating and ranking chunks as also complete parses (chunk sequences) so as to place the best parses near the top. The main aim of this work is to identify the best word group (chunk) sequence or global parse for a given sentence using a information-theoretic measure called mutual information. This method is based on the hypothesis that the best chunk can be obtained by analysing the mutual information values of the chunk tag n-grams. In the initial version of UCSG, HMMs were local to chunks. Global information such as the probability of a chunk of a given type starting a sentence or the probability of a chunk of a particular type occurring next to a chunk of a given type are also useful. We try to capture this global information in the form of mutual information score and use it in improving the ranks of correct chunks. Combining the two methods, namely HMMs and Mutual Information, to get the full information of a chunk to improve the ranks further, is another important aspect of this work. Later, a Best First Search module uses these ranked chunks to find the final parse. We have also added 1000 sentence manually parsed corpus to the existing 4000 manually parsed data. Experiments on the existing and newly added text corpora are included. 1 Introduction Developing high performance, wide coverage computational grammars and parsers has proved to be a very difficult task. Hence the recent interest in shallow or partial parsing systems. Shallow or Partial Parsing means The task of recovering only a limited amount of syntactic information from natural language sentences. Shallow parsing does not attempt to resolve all semantically significant decisions. Most of the times it is restricted to identifying phrases in the sentence. Partial syntactical information can help to solve many natural language processing tasks, such as information extraction, text summarisation, machine translation and spoken language understanding. Purely linguistic approaches have not proved practicable for developing wide coverage parsers and purely machine learning approaches are also impracticable in many cases due to non-availability of large enough training corpora. Only a judicious combination of the two approaches can lead to wide coverage, robust parsing systems. In this work, we are using the UCSG shallow parser architecture being developed at University of Hyderabad, which combines linguistic constrains expressed in the form of finite state grammars with statistical rating using HMMs built from a large POS-tagged corpus and a Best First search for global optimisation for determining the best parser for the give sentence [3, 2]. The UCSG shallow parsing architecture is depicted in figure1. The input to the system is either a plain or a POS tagged sentence. The aim is to produce all possible parses in ranked order hoping to get the best parse to the top. In this work by parse we mean a sequence of chunks. Chunks are sequence of words. A chunk or a word group as we prefer to call it in UCSG, is a structural unit, a non-over lapping and non-recursive sequence of words, that can as a whole, play a role in some prediction [3, 2]. 1

2 UCSG is a three module architecture. The first module is a Finite State Grammar (FSM) Parser. The goal of this module is to accept all valid word groups but not necessarily the only word groups that are appropriate in context for a given sentence. We do not restrict or prune the various possibilities. Instead we use a separate module to rate and rank these word groups later. The second module of this architecture uses a set of Hidden Markov Models (HMM) for rating and ranking the word groups produced by the FSM module. We can of course limit the number of parses generated if required but the ability to produce all possible parses is fundamental to the architecture. Note that we do not produces all possible parses first and then rate and rank them. Instead we attempt to produce the parses in best first order. An example is given below: input sentence: Savitri is loved by sita. ##<NP0><Savitri>##<VBZ><is>##<VVN_VVD><loved> ##<PRN_AVP_PRP><by>##<NP0><Sita>## Chunks got after FSM : <ng><0-1><np0><savitri> <vg><1-2><vbz><is> <vg><1-3><vbz><is>##<vvn><loved> <vg><1-4><vbz><is>##<vvn><loved>##<avp><by> <vg><2-3><vvd><loved> <vgs><2-3><vvn><loved> <vg><2-4><vvd><loved>##<avp><by> <ajg><2-3><vvn><loved> <vgs><2-4><vvn><loved>##<avp><by> <ng><3-5><prp><by>##<np0><sita> <part><3-4><avp><by> <ng><4-5><np0><sita> Figure 1. UCSG Shallow Parsing Architecture The aim here is to assign the highest rank for the correct chunk and to push down other chunks. Since a final parse is a sequence of chunks that covers the given sentence with no overlaps or gaps, we evaluate the alternatives at each position in the sentence in a left-to-right manner. We evaluate the second module using Mean Score, which is the mean of the distribution of ranks of correct chunks produced for a given training corpus. Ideally, all the correct chunks would be at the top and hence the score would be 1. The aim is to get Mean Score as close to 1 as possible. We have used a manually parsed corpus of 4000 sentences developed by us to evaluate the Mean Score. The third module is for identifying the best chunk sequence or global parse for a given sentence. This module generate all possible parses hopefully in best first order. Words and chunks in a sentence are referred to in terms of the positions they occupy in the sentence. Positions are marked between words, starting from zero to the left of the first word. The very first word is between positions 0 and 1. A word group containing the third and fourth words in the sentence can be referred as W 2,4. The following steps describe how we map a given sentence and word groups present in the sentence into a graph. The positions in the sentence are treated as nodes of the resulting graph. If a sentence contains N words then the graph contains N + 1 nodes corresponding to the N + 1 positions in the sentence. Word group W i,j is represented as an edge form node i to node j. The probability of a word group W i,j given by HMM module and the transition probability from previous word group type to current word group type are combined to estimate the cost of an arc between the nodes i and j. We always start from the initial node 0. Length of the sentence N is the goal node. Now our parse selection problem of a sentence containing N words becomes the task of finding an optimal path from node 0 to node N.

3 Choosing the locally best chunks at each position in a given sentence does not necessarily give us the best parse (chunk sequence) in all cases. The HMMs used in the initial version of UCSG are local to chunks. They use only with-in-chunk information for calculating the score of a chunk, ignoring the global information such as the probability of a chunk of a given type starting a sentence or the probability of a chunk of a particular type occurring next to a chunk of a given type. In this work, we have used Mutual Information statistics to capture this global information. 2 Manually Parsed Corpus Development We have a manually parsed corpus of 4000 sentences already developed in our centre here. In order to add to this effort and to test the performance of the system on new data, we have developed an additional 1000 sentence manually parsed corpus, taking plain sentences from Reuters corpus. This corpus is thus very useful for evaluating the various modules of the parsing architecture and also for bootstrapping. This corpus was developed by parsing the sentences using the UCSG shallow parser for English and then manually checking the top parse and making corrections where required. We felt this was far easier than parsing the sentences entirely by hand. 3 Rating and ing Chunks using Mutual information 3.1 Mutual Information In his treatise on information theory Transmission of Information [4] Fano discusses mutual information as a measure of the interdependence of two signals in a message. In probability theory and information theory, mutual information between two random variables is a quantity that measures the mutual dependence of the two variables. This bigram mutual information is a function of the probabilities of the two events: MI(x, y) = log Pr(x, y) Pr(x) Pr(y) Consider these events not as signals but as a sequence of part-of-speech tags in a sentence. Then an estimate of mutual information of two categories xy is: #xyincorpus total#of bigramsincorpus M I(x, y) log ( ) ( ) #x #y corpussize corpussize We can thus take advantage of context in determining the rank of particular chunk using pairs of tokens or bigrams. 3.2 Rating and ing chunks In UCSG shallow parsing Architecture [1] the first module produces all possible chunks according to the given finite state grammar. Due to lexical ambiguities and inherent ambiguities in the grammar, we often produce many alternatives. By giving the top rank to the correct chunk and pushing down other chunks we can improve the performance of the parser. At each possible branching point or position in a sentence we find all the chunk types on either side of the point, calculate the bigram mutual information score between each pair of chunk types and associate this score with the right side chunk. In the final module, we start form left and select the chunk on the right side with a high score. This goes on till the end to complete one single parse. The following steps describe you how we calculate MI score of any chunk: At each position i in the sentence, find all the chunk groups ending with i and all the chunk groups beginning with i. Let W j1,i, W j2,i, W j3,i...w jn,i are all chunks ending with i and W i,k1, W i,k2, W i,k3...w i,km are all chunks beginning with i. MI(x,y) gives us the mutual information score between x and y. MI score of any chunk group which is beginning with i is calculated by combining all the MI scores of chunk groups which are ending with i. MI(W i,k1 ) = n p=1 MI(W jp,i, W i,k1 ) MI(W i,k2 ) = n p=1 MI(W jp,i, W i,k1 ) MI(W i,k3 ) = n p=1 MI(W jp,i, W i,k1 ) : : MI(W i,km ) = n p=1 MI(W jp,i, W i,km )

4 Initially we calculate the score of a chunk using bigram mutual information, which only checks the chunk types. This way, we are giving equal weight to the chunks of same type even though they were differ in terms of length. Thus, differentiating long and short chunks of same type becomes a difficult task. We have therefore extended the set of chunk types to include the chunk lengths. Now noun groups of 3 words, for example, will be treated as different from a noun group of 4 words. In this work, we have used the term LMI to denote MI applied to this new, refined set of chunk types. Note that the basic mutual information technique itself is not changed, we are only applying MI to a more fine grained set of chunk types. Our observations show that LMI is better than the general MI. LMI gives better Mean Score than general MI or HMM alone. However, it gives less recall compared to the HMM method. This result encourages us to combine the two techniques. By combining the two techniques we will hopefully get the full information of any single chunk - both the local as well as the global information. 3.3 Weighted Voting Technique We find that HMM is giving good recall while LMI improves the Mean score. The best combination of these techniques can be helpful in improving the whole system further. A voting system is a means of choosing between a number of options, based on their performance. Let W1, W2 be the two methods and let f1, f2 be the weights. We maximise W1*f1+W2*f Results Since we need high Recall and a low (close to 1) Mean Score, we can search for optimum weights by defining a score that combines these two requirements: CombinedScore = M R 25 + Recall. Performance for the original version of UCSG which uses only HMMs is given first: Table 1. Performance of the HMM Module on the Manually Parsed Corpus of 4000 sentences - Plain Sentences as Input Cutoff Mean Recall Combined Score Score (%) [MR*-25+Recall*1] We give below the performance of the system using an optimum weighting based on extensive experimentation. We find that it is useful to make the LMI weight a function of the length of the chunks. Best results have been obtained when the weight is half of the chunk length. Table 2. Performance of the Weighted Voting technique(hmm+lmi) Module on the Manually Parsed Corpus of 4000 sentences - Plain Sentences as Input Cutoff Mean Recall Combined Score Score (%) [MR*-25+Recall*1] We see that both the Mean Score and the Recall have improved, as also the Combined Score. 4 Optimum Chunk sequence Selection The parse generation module has been evaluated on the manually parsed corpus in terms of rank of the fully correct parse and also in terms of percentage of correct chunks in

5 the top parse. Plain sentences and POS tagged sentences have been considered separately for input. The results are summarised in tables given below. Here, we have restricted the parsing time taken by the best first search algorithm to 10 epoch seconds for each sentence because the time and space complexity increases exponentially as branching factor (b) and length of the sentence (n) increases. 4.1 Experiments and Results Table 4. Performance of the Best First Search Module after using MI in UCSG - Test Data of 4000 Sentences Plain POS tagged Sentences Sentences % of Correct parses % of Correct chunks Total Recall In first experiment we have taken 4000 plain and POS tagged sentences from the manually parsed corpus developed by us here for evaluating the two systems. We give below the results. The performance of the original system using only HMMs is included for comparison. Table 3. Performance of Best First Search Module in initial version of UCSG - Test Data of 4000 Sentences Plain POS tagged Sentences Sentences % of Correct parses % of Correct chunks Total Recall All these experiments have been done using epoch time limit of 10 seconds. For training MI we have used 4000 manually parsed corpus. We observe that performance has not improved. MI improved Mean Scores. We observe that the ratings of good chunks are improved but the ratings for other chunks are also improved. As a result, the performance of the final chunk sequence generation module does not necessarily improve. We find that this is because of sparse data problem. More experimentation is required to find optimal solutions. In order to check for corpus effects, we have taken 1000 sentences from Reuters News Corpus for testing. For training we have used 4000 manually parsed corpus. The results are given below. Table 5. Performance of the Best First Search Module - Test Data of 1000 Sentences (Initial Version) (UCSG with MI) % of Correct parses % of Correct chunks Total Recall

6 Again we see that we have not been able to obtain better performance in the final parse. It has been shown [2] that the POS tags of the top parse are largely correct. Here we explore the idea of using the UCSG shallow parser as a POS tagger. We run the parser twice, first time only for obtaining the POS tags for words as given the top parse. The results are given below: Table 6. Performance of the Best First Search Module using existed system as POS tagger- Test Data of 4000 Sentences Initial Version UCSG with MI % of Correct parses % of Correct chunks Total Recall Because of time limits, we are able to get only a small portion of the sentences parsed. Most of the POS tags in the top parse are correct. There will be a few errors and hence when we parse the same sentences a second time, the performance in terms of accuracy will remain more or less the same as original with the small reduction explained out by the POS tag errors. Of course the second pass with POS tagged sentences as input will be much faster. 5 Significant Observations The following observations relate to the two versions specified below: System A: Initial version of the UCSG parser which uses HMMs only for rating and ranking the chunks System B: New version of the UCSG parser which uses the optimal combination of LMI and HMM for rating and ranking the chunks. Case1: Correct chunks have been given bad ranks (other than first) in system A. This is the most common case. We give some examples here. All the phrases we show here are in the format: Each entry includes the chunk type, the starting and ending positions, the chunk itself with the POS tags of all the words, log probability given by HMM, rank, number of items in the set, and the serial number of the branching points. The correct word group is will have taken. Phrases got in System B: <vg><6-9><vm0><will>##<vhb><have>##<vvn><taken> < ><1><5><7> <vg><6-8><vm0><will>##<vhb><have> < ><2><5><7> <vg><6-7><vm0><will>< ><3><5><7> <ng><6-7><np0><will>< ><4><5><7> <ng><6-7><nn1><will>< ><5><5><7> <vg><7-9><vhb><have>##<vvn><taken> < ><1><2><8> <vg><7-8><vhb><have>< ><2><2><8> <ajg><8-9><vvn><taken>< ><1><2><9> <vgs><8-9><vvn><taken>< ><2><2><9> Phrases got in System A: <vg><6-7><vm0><will>< ><1><5><7> <vg><6-8><vm0><will>##<vhb><have> < ><2><5><7> <ng><6-7><np0><will>< ><3><5><7> <vg><6-9><vm0><will>##<vhb><have>##<vvn><taken> < ><4><5><7> <ng><6-7><nn1><will> < ><5><5><7> <vg><7-8><vhb><have>< ><1><2><8> <vg><7-9><vhb><have>##<vvn><taken> < ><2><2><8> <vgs><8-9><vvn><taken> < ><1><2><9> <ajg><8-9><vvn><taken> < ><2><2><9> of correct chunk: System B rank : 1 System A rank : 3 Final Parse : In both systems the top parse has the correct chunk, even though system A fails to give rank 1 to the correct chunk Reason: The combined score of will, have and taken is worse than the score for will have taken and hence the correct chunk will be selected even if it is not at the top. Case2: This is the reverse of case 1. Here system A gives the correct rank but not System B. For example, The correct chunk sequence is neither abdul nor karim For the first chunk of the above sequence neither

7 Phrases got in System B: <ng><0-4><dt0><neither>##<np0><abdul> ##<CJC><nor>##<NP0><karim>< ><1><4><1> <ng><0-2><dt0><neither>##<np0><abdul> < ><2><4><1> <avg><0-1><av0><neither>< ><3><4><1> <ng><0-1><pnp><neither>< ><4><4><1> <ng><1-4><np0><abdul>##<cjc><nor>##<np0><karim> < ><1><2><2> Phrases got in System A: <avg><0-1><av0><neither>< ><1><4><1> <ng><0-1><pnp><neither>< ><2><4><1> <ng><0-2><dt0><neither>##<np0><abdul> < ><3><4><1> <ng><0-4><dt0><neither>##<np0><abdul> ##<CJC><nor>##<NP0><karim>< ><4><4><1> <ng><1-2><np0><abdul> < ><1><2><2> <ng><1-4><np0><abdul>##<cjc><nor>## <NP0><karim>< ><2><2><2> System B rank : 3 System A rank : 1 Final Parse: In both systems the top parse has the correct chunk, even though System B fails to give rank 1 Reason: Similar to case 1 Case 3: This is the case Where all the correct chunks have been given 1st rank. Yet the top parse is not the correct parse The Correct Chunk Sequence is have you seen ganguli the artists drawings Phrases got in System B: <vg><0-1><vhb><have>< ><1><1><1> <ng><1-2><pnc><you><3.6783><1><1><2> <ng><2-4><vvn><seen>##<np0><ganguli> < ><1><3><3> <vgs><2-3><vvn><seen><-3.082><2><3><3> <ajg><2-3><vvn><seen>< ><3><3><3> <ng><3-4><np0><ganguli><2.363><1><1><4> <ng><4-7><at0><the>##<nps><artist s>## <NN2><drawings>< ><1><1><5> Phrases got in System A: <vg><0-1><vhb><have>< ><1><1><1> <ng><1-2><pnc><you>< ><1><1><2> <vgs><2-3><vvn><seen>< ><1><3><3> <ng><2-4><vvn><seen>##<np0><ganguli> < ><2><3><3> <ajg><2-3><vvn><seen>< ><3><3><3> <ng><3-4><np0><ganguli>< ><1><1><4> <ng><4-7><at0><the>##<nps><artist s> ##<NN2><drawings> < ><1><1><5> System B rank: Except one chunk (seen) all the other correct chunks are given rank1 System A rank: All the correct chunks have been given rank 1 Final Parse: Top parse is not the correct parse Reason: Sum of correct chunk (seenand ganguli) scores is less than the wrong chunk s (seen ganguli) score 6 Conclusions and Future work In this work we have explored the use of mutual information between chunk types for improving the UCSG shallow parsing architecture. Initial results are encouraging. More experimentation is required to get optimal results. We have also added 1000 more sentences to the manually parsed corpus. Using MI we can improve the ranks of correct chunks. We can use some threshold point to push the bad chunks further down. A variety of machine learning strategies as also sentence level linguistic constraints can be incorporated to further improve the chunk sequence selection. The partial parsing system is currently producing only chunk sequences as parse output. Thematic role assignment (such as subject and object) should be included. Further enhancements to handle clause structure would take it closer to a full syntactic parsing system. It is clear that if we reduce the sparse data effect we will get even better results. We can use linguists features with MI in selection of correct chunk. We can extend the bigram mutual information to generalised mutual information, where we take a window of chunk types in calculating MI between two chunks. We have worked with chunk type only. We can extend it to include the lexical information also. References [1] G. B. Kumar and K. N. Murthy. Ucsg shallow parser. Proceedings of CICLING 2006, LNCS, 3878: , [2] D. M. Magerman and M. P. Marcus. Parsing natural language using mutual information statistics. pages , [3] K. N. Murthy. Universal Clause Structure Grammar. PhD Thesis, University of Hyderabad, [4] F. R. Transmission of information

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

The Internet as a Normative Corpus: Grammar Checking with a Search Engine The Internet as a Normative Corpus: Grammar Checking with a Search Engine Jonas Sjöbergh KTH Nada SE-100 44 Stockholm, Sweden jsh@nada.kth.se Abstract In this paper some methods using the Internet as a

More information

Using dialogue context to improve parsing performance in dialogue systems

Using dialogue context to improve parsing performance in dialogue systems Using dialogue context to improve parsing performance in dialogue systems Ivan Meza-Ruiz and Oliver Lemon School of Informatics, Edinburgh University 2 Buccleuch Place, Edinburgh I.V.Meza-Ruiz@sms.ed.ac.uk,

More information

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17. Semi-supervised methods of text processing, and an application to medical concept extraction Yacine Jernite Text-as-Data series September 17. 2015 What do we want from text? 1. Extract information 2. Link

More information

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar Chung-Chi Huang Mei-Hua Chen Shih-Ting Huang Jason S. Chang Institute of Information Systems and Applications, National Tsing Hua University,

More information

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Chinese Language Parsing with Maximum-Entropy-Inspired Parser Chinese Language Parsing with Maximum-Entropy-Inspired Parser Heng Lian Brown University Abstract The Chinese language has many special characteristics that make parsing difficult. The performance of state-of-the-art

More information

Parsing of part-of-speech tagged Assamese Texts

Parsing of part-of-speech tagged Assamese Texts IJCSI International Journal of Computer Science Issues, Vol. 6, No. 1, 2009 ISSN (Online): 1694-0784 ISSN (Print): 1694-0814 28 Parsing of part-of-speech tagged Assamese Texts Mirzanur Rahman 1, Sufal

More information

The Smart/Empire TIPSTER IR System

The Smart/Empire TIPSTER IR System The Smart/Empire TIPSTER IR System Chris Buckley, Janet Walz Sabir Research, Gaithersburg, MD chrisb,walz@sabir.com Claire Cardie, Scott Mardis, Mandar Mitra, David Pierce, Kiri Wagstaff Department of

More information

Using Semantic Relations to Refine Coreference Decisions

Using Semantic Relations to Refine Coreference Decisions Using Semantic Relations to Refine Coreference Decisions Heng Ji David Westbrook Ralph Grishman Department of Computer Science New York University New York, NY, 10003, USA hengji@cs.nyu.edu westbroo@cs.nyu.edu

More information

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence. NLP Lab Session Week 8 October 15, 2014 Noun Phrase Chunking and WordNet in NLTK Getting Started In this lab session, we will work together through a series of small examples using the IDLE window and

More information

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words, A Language-Independent, Data-Oriented Architecture for Grapheme-to-Phoneme Conversion Walter Daelemans and Antal van den Bosch Proceedings ESCA-IEEE speech synthesis conference, New York, September 1994

More information

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics (L615) Markus Dickinson Department of Linguistics, Indiana University Spring 2013 The web provides new opportunities for gathering data Viable source of disposable corpora, built ad hoc for specific purposes

More information

Universiteit Leiden ICT in Business

Universiteit Leiden ICT in Business Universiteit Leiden ICT in Business Ranking of Multi-Word Terms Name: Ricardo R.M. Blikman Student-no: s1184164 Internal report number: 2012-11 Date: 07/03/2013 1st supervisor: Prof. Dr. J.N. Kok 2nd supervisor:

More information

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za

More information

Beyond the Pipeline: Discrete Optimization in NLP

Beyond the Pipeline: Discrete Optimization in NLP Beyond the Pipeline: Discrete Optimization in NLP Tomasz Marciniak and Michael Strube EML Research ggmbh Schloss-Wolfsbrunnenweg 33 69118 Heidelberg, Germany http://www.eml-research.de/nlp Abstract We

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF) SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF) Hans Christian 1 ; Mikhael Pramodana Agus 2 ; Derwin Suhartono 3 1,2,3 Computer Science Department,

More information

Distant Supervised Relation Extraction with Wikipedia and Freebase

Distant Supervised Relation Extraction with Wikipedia and Freebase Distant Supervised Relation Extraction with Wikipedia and Freebase Marcel Ackermann TU Darmstadt ackermann@tk.informatik.tu-darmstadt.de Abstract In this paper we discuss a new approach to extract relational

More information

Ensemble Technique Utilization for Indonesian Dependency Parser

Ensemble Technique Utilization for Indonesian Dependency Parser Ensemble Technique Utilization for Indonesian Dependency Parser Arief Rahman Institut Teknologi Bandung Indonesia 23516008@std.stei.itb.ac.id Ayu Purwarianti Institut Teknologi Bandung Indonesia ayu@stei.itb.ac.id

More information

Prediction of Maximal Projection for Semantic Role Labeling

Prediction of Maximal Projection for Semantic Role Labeling Prediction of Maximal Projection for Semantic Role Labeling Weiwei Sun, Zhifang Sui Institute of Computational Linguistics Peking University Beijing, 100871, China {ws, szf}@pku.edu.cn Haifeng Wang Toshiba

More information

Learning Computational Grammars

Learning Computational Grammars Learning Computational Grammars John Nerbonne, Anja Belz, Nicola Cancedda, Hervé Déjean, James Hammerton, Rob Koeling, Stasinos Konstantopoulos, Miles Osborne, Franck Thollard and Erik Tjong Kim Sang Abstract

More information

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases POS Tagging Problem Part-of-Speech Tagging L545 Spring 203 Given a sentence W Wn and a tagset of lexical categories, find the most likely tag T..Tn for each word in the sentence Example Secretariat/P is/vbz

More information

A Domain Ontology Development Environment Using a MRD and Text Corpus

A Domain Ontology Development Environment Using a MRD and Text Corpus A Domain Ontology Development Environment Using a MRD and Text Corpus Naomi Nakaya 1 and Masaki Kurematsu 2 and Takahira Yamaguchi 1 1 Faculty of Information, Shizuoka University 3-5-1 Johoku Hamamatsu

More information

Applications of memory-based natural language processing

Applications of memory-based natural language processing Applications of memory-based natural language processing Antal van den Bosch and Roser Morante ILK Research Group Tilburg University Prague, June 24, 2007 Current ILK members Principal investigator: Antal

More information

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities Yoav Goldberg Reut Tsarfaty Meni Adler Michael Elhadad Ben Gurion

More information

Corrective Feedback and Persistent Learning for Information Extraction

Corrective Feedback and Persistent Learning for Information Extraction Corrective Feedback and Persistent Learning for Information Extraction Aron Culotta a, Trausti Kristjansson b, Andrew McCallum a, Paul Viola c a Dept. of Computer Science, University of Massachusetts,

More information

Discriminative Learning of Beam-Search Heuristics for Planning

Discriminative Learning of Beam-Search Heuristics for Planning Discriminative Learning of Beam-Search Heuristics for Planning Yuehua Xu School of EECS Oregon State University Corvallis,OR 97331 xuyu@eecs.oregonstate.edu Alan Fern School of EECS Oregon State University

More information

Speech Recognition at ICSI: Broadcast News and beyond

Speech Recognition at ICSI: Broadcast News and beyond Speech Recognition at ICSI: Broadcast News and beyond Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 The DARPA Broadcast News task Aspects of ICSI

More information

AQUA: An Ontology-Driven Question Answering System

AQUA: An Ontology-Driven Question Answering System AQUA: An Ontology-Driven Question Answering System Maria Vargas-Vera, Enrico Motta and John Domingue Knowledge Media Institute (KMI) The Open University, Walton Hall, Milton Keynes, MK7 6AA, United Kingdom.

More information

Short Text Understanding Through Lexical-Semantic Analysis

Short Text Understanding Through Lexical-Semantic Analysis Short Text Understanding Through Lexical-Semantic Analysis Wen Hua #1, Zhongyuan Wang 2, Haixun Wang 3, Kai Zheng #4, Xiaofang Zhou #5 School of Information, Renmin University of China, Beijing, China

More information

Online Updating of Word Representations for Part-of-Speech Tagging

Online Updating of Word Representations for Part-of-Speech Tagging Online Updating of Word Representations for Part-of-Speech Tagging Wenpeng Yin LMU Munich wenpeng@cis.lmu.de Tobias Schnabel Cornell University tbs49@cornell.edu Hinrich Schütze LMU Munich inquiries@cislmu.org

More information

Training and evaluation of POS taggers on the French MULTITAG corpus

Training and evaluation of POS taggers on the French MULTITAG corpus Training and evaluation of POS taggers on the French MULTITAG corpus A. Allauzen, H. Bonneau-Maynard LIMSI/CNRS; Univ Paris-Sud, Orsay, F-91405 {allauzen,maynard}@limsi.fr Abstract The explicit introduction

More information

BYLINE [Heng Ji, Computer Science Department, New York University,

BYLINE [Heng Ji, Computer Science Department, New York University, INFORMATION EXTRACTION BYLINE [Heng Ji, Computer Science Department, New York University, hengji@cs.nyu.edu] SYNONYMS NONE DEFINITION Information Extraction (IE) is a task of extracting pre-specified types

More information

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments Cristina Vertan, Walther v. Hahn University of Hamburg, Natural Language Systems Division Hamburg,

More information

arxiv: v1 [cs.cl] 2 Apr 2017

arxiv: v1 [cs.cl] 2 Apr 2017 Word-Alignment-Based Segment-Level Machine Translation Evaluation using Word Embeddings Junki Matsuo and Mamoru Komachi Graduate School of System Design, Tokyo Metropolitan University, Japan matsuo-junki@ed.tmu.ac.jp,

More information

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy Informatics 2A: Language Complexity and the Chomsky Hierarchy September 28, 2010 Starter 1 Is there a finite state machine that recognises all those strings s from the alphabet {a, b} where the difference

More information

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm syntax: from the Greek syntaxis, meaning setting out together

More information

Linking Task: Identifying authors and book titles in verbose queries

Linking Task: Identifying authors and book titles in verbose queries Linking Task: Identifying authors and book titles in verbose queries Anaïs Ollagnier, Sébastien Fournier, and Patrice Bellot Aix-Marseille University, CNRS, ENSAM, University of Toulon, LSIS UMR 7296,

More information

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation Vivek Kumar Rangarajan Sridhar, John Chen, Srinivas Bangalore, Alistair Conkie AT&T abs - Research 180 Park Avenue, Florham Park,

More information

Cross Language Information Retrieval

Cross Language Information Retrieval Cross Language Information Retrieval RAFFAELLA BERNARDI UNIVERSITÀ DEGLI STUDI DI TRENTO P.ZZA VENEZIA, ROOM: 2.05, E-MAIL: BERNARDI@DISI.UNITN.IT Contents 1 Acknowledgment.............................................

More information

A Graph Based Authorship Identification Approach

A Graph Based Authorship Identification Approach A Graph Based Authorship Identification Approach Notebook for PAN at CLEF 2015 Helena Gómez-Adorno 1, Grigori Sidorov 1, David Pinto 2, and Ilia Markov 1 1 Center for Computing Research, Instituto Politécnico

More information

CS 598 Natural Language Processing

CS 598 Natural Language Processing CS 598 Natural Language Processing Natural language is everywhere Natural language is everywhere Natural language is everywhere Natural language is everywhere!"#$%&'&()*+,-./012 34*5665756638/9:;< =>?@ABCDEFGHIJ5KL@

More information

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers Chad Langley, Alon Lavie, Lori Levin, Dorcas Wallace, Donna Gates, and Kay Peterson Language Technologies Institute Carnegie

More information

Learning Methods in Multilingual Speech Recognition

Learning Methods in Multilingual Speech Recognition Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex

More information

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation tatistical Parsing (Following slides are modified from Prof. Raymond Mooney s slides.) tatistical Parsing tatistical parsing uses a probabilistic model of syntax in order to assign probabilities to each

More information

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks POS tagging of Chinese Buddhist texts using Recurrent Neural Networks Longlu Qin Department of East Asian Languages and Cultures longlu@stanford.edu Abstract Chinese POS tagging, as one of the most important

More information

On the Combined Behavior of Autonomous Resource Management Agents

On the Combined Behavior of Autonomous Resource Management Agents On the Combined Behavior of Autonomous Resource Management Agents Siri Fagernes 1 and Alva L. Couch 2 1 Faculty of Engineering Oslo University College Oslo, Norway siri.fagernes@iu.hio.no 2 Computer Science

More information

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus Language Acquisition Fall 2010/Winter 2011 Lexical Categories Afra Alishahi, Heiner Drenhaus Computational Linguistics and Phonetics Saarland University Children s Sensitivity to Lexical Categories Look,

More information

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models Netpix: A Method of Feature Selection Leading to Accurate Sentiment-Based Classification Models 1 Netpix: A Method of Feature Selection Leading to Accurate Sentiment-Based Classification Models James B.

More information

Software Maintenance

Software Maintenance 1 What is Software Maintenance? Software Maintenance is a very broad activity that includes error corrections, enhancements of capabilities, deletion of obsolete capabilities, and optimization. 2 Categories

More information

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Good Judgment Project: A large scale test of different methods of combining expert predictions The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania

More information

Age Effects on Syntactic Control in. Second Language Learning

Age Effects on Syntactic Control in. Second Language Learning Age Effects on Syntactic Control in Second Language Learning Miriam Tullgren Loyola University Chicago Abstract 1 This paper explores the effects of age on second language acquisition in adolescents, ages

More information

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2 Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2 Ted Pedersen Department of Computer Science University of Minnesota Duluth, MN, 55812 USA tpederse@d.umn.edu

More information

Developing Grammar in Context

Developing Grammar in Context Developing Grammar in Context intermediate with answers Mark Nettle and Diana Hopkins PUBLISHED BY THE PRESS SYNDICATE OF THE UNIVERSITY OF CAMBRIDGE The Pitt Building, Trumpington Street, Cambridge, United

More information

A Case Study: News Classification Based on Term Frequency

A Case Study: News Classification Based on Term Frequency A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center

More information

Methods for the Qualitative Evaluation of Lexical Association Measures

Methods for the Qualitative Evaluation of Lexical Association Measures Methods for the Qualitative Evaluation of Lexical Association Measures Stefan Evert IMS, University of Stuttgart Azenbergstr. 12 D-70174 Stuttgart, Germany evert@ims.uni-stuttgart.de Brigitte Krenn Austrian

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011

The Karlsruhe Institute of Technology Translation Systems for the WMT 2011 The Karlsruhe Institute of Technology Translation Systems for the WMT 2011 Teresa Herrmann, Mohammed Mediani, Jan Niehues and Alex Waibel Karlsruhe Institute of Technology Karlsruhe, Germany firstname.lastname@kit.edu

More information

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data Ebba Gustavii Department of Linguistics and Philology, Uppsala University, Sweden ebbag@stp.ling.uu.se

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

Detecting English-French Cognates Using Orthographic Edit Distance

Detecting English-French Cognates Using Orthographic Edit Distance Detecting English-French Cognates Using Orthographic Edit Distance Qiongkai Xu 1,2, Albert Chen 1, Chang i 1 1 The Australian National University, College of Engineering and Computer Science 2 National

More information

A DISTRIBUTIONAL STRUCTURED SEMANTIC SPACE FOR QUERYING RDF GRAPH DATA

A DISTRIBUTIONAL STRUCTURED SEMANTIC SPACE FOR QUERYING RDF GRAPH DATA International Journal of Semantic Computing Vol. 5, No. 4 (2011) 433 462 c World Scientific Publishing Company DOI: 10.1142/S1793351X1100133X A DISTRIBUTIONAL STRUCTURED SEMANTIC SPACE FOR QUERYING RDF

More information

THE VERB ARGUMENT BROWSER

THE VERB ARGUMENT BROWSER THE VERB ARGUMENT BROWSER Bálint Sass sass.balint@itk.ppke.hu Péter Pázmány Catholic University, Budapest, Hungary 11 th International Conference on Text, Speech and Dialog 8-12 September 2008, Brno PREVIEW

More information

Indian Institute of Technology, Kanpur

Indian Institute of Technology, Kanpur Indian Institute of Technology, Kanpur Course Project - CS671A POS Tagging of Code Mixed Text Ayushman Sisodiya (12188) {ayushmn@iitk.ac.in} Donthu Vamsi Krishna (15111016) {vamsi@iitk.ac.in} Sandeep Kumar

More information

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures Ulrike Baldewein (ulrike@coli.uni-sb.de) Computational Psycholinguistics, Saarland University D-66041 Saarbrücken,

More information

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS Julia Tmshkina Centre for Text Techitology, North-West University, 253 Potchefstroom, South Africa 2025770@puk.ac.za

More information

Lessons from a Massive Open Online Course (MOOC) on Natural Language Processing for Digital Humanities

Lessons from a Massive Open Online Course (MOOC) on Natural Language Processing for Digital Humanities Lessons from a Massive Open Online Course (MOOC) on Natural Language Processing for Digital Humanities Simon Clematide, Isabel Meraner, Noah Bubenhofer, Martin Volk Institute of Computational Linguistics

More information

Named Entity Recognition: A Survey for the Indian Languages

Named Entity Recognition: A Survey for the Indian Languages Named Entity Recognition: A Survey for the Indian Languages Padmaja Sharma Dept. of CSE Tezpur University Assam, India 784028 psharma@tezu.ernet.in Utpal Sharma Dept.of CSE Tezpur University Assam, India

More information

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments Vijayshri Ramkrishna Ingale PG Student, Department of Computer Engineering JSPM s Imperial College of Engineering &

More information

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models Richard Johansson and Alessandro Moschitti DISI, University of Trento Via Sommarive 14, 38123 Trento (TN),

More information

Guidelines for Writing an Internship Report

Guidelines for Writing an Internship Report Guidelines for Writing an Internship Report Master of Commerce (MCOM) Program Bahauddin Zakariya University, Multan Table of Contents Table of Contents... 2 1. Introduction.... 3 2. The Required Components

More information

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

Project in the framework of the AIM-WEST project Annotation of MWEs for translation Project in the framework of the AIM-WEST project Annotation of MWEs for translation 1 Agnès Tutin LIDILEM/LIG Université Grenoble Alpes 30 october 2014 Outline 2 Why annotate MWEs in corpora? A first experiment

More information

Developing a TT-MCTAG for German with an RCG-based Parser

Developing a TT-MCTAG for German with an RCG-based Parser Developing a TT-MCTAG for German with an RCG-based Parser Laura Kallmeyer, Timm Lichte, Wolfgang Maier, Yannick Parmentier, Johannes Dellert University of Tübingen, Germany CNRS-LORIA, France LREC 2008,

More information

On document relevance and lexical cohesion between query terms

On document relevance and lexical cohesion between query terms Information Processing and Management 42 (2006) 1230 1247 www.elsevier.com/locate/infoproman On document relevance and lexical cohesion between query terms Olga Vechtomova a, *, Murat Karamuftuoglu b,

More information

Specifying a shallow grammatical for parsing purposes

Specifying a shallow grammatical for parsing purposes Specifying a shallow grammatical for parsing purposes representation Atro Voutilainen and Timo J~irvinen Research Unit for Multilingual Language Technology P.O. Box 4 FIN-0004 University of Helsinki Finland

More information

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations 4 Interior point algorithms for network ow problems Mauricio G.C. Resende AT&T Bell Laboratories, Murray Hill, NJ 07974-2070 USA Panos M. Pardalos The University of Florida, Gainesville, FL 32611-6595

More information

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING SISOM & ACOUSTICS 2015, Bucharest 21-22 May THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING MarilenaăLAZ R 1, Diana MILITARU 2 1 Military Equipment and Technologies Research Agency, Bucharest,

More information

Search right and thou shalt find... Using Web Queries for Learner Error Detection

Search right and thou shalt find... Using Web Queries for Learner Error Detection Search right and thou shalt find... Using Web Queries for Learner Error Detection Michael Gamon Claudia Leacock Microsoft Research Butler Hill Group One Microsoft Way P.O. Box 935 Redmond, WA 981052, USA

More information

SEMAFOR: Frame Argument Resolution with Log-Linear Models

SEMAFOR: Frame Argument Resolution with Log-Linear Models SEMAFOR: Frame Argument Resolution with Log-Linear Models Desai Chen or, The Case of the Missing Arguments Nathan Schneider SemEval July 16, 2010 Dipanjan Das School of Computer Science Carnegie Mellon

More information

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing Grzegorz Chrupa la A dissertation submitted in fulfilment of the requirements for the award of Doctor of Philosophy (Ph.D.)

More information

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford

More information

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many Schmidt 1 Eric Schmidt Prof. Suzanne Flynn Linguistic Study of Bilingualism December 13, 2013 A Minimalist Approach to Code-Switching In the field of linguistics, the topic of bilingualism is a broad one.

More information

Reinforcement Learning by Comparing Immediate Reward

Reinforcement Learning by Comparing Immediate Reward Reinforcement Learning by Comparing Immediate Reward Punit Pandey DeepshikhaPandey Dr. Shishir Kumar Abstract This paper introduces an approach to Reinforcement Learning Algorithm by comparing their immediate

More information

Word Segmentation of Off-line Handwritten Documents

Word Segmentation of Off-line Handwritten Documents Word Segmentation of Off-line Handwritten Documents Chen Huang and Sargur N. Srihari {chuang5, srihari}@cedar.buffalo.edu Center of Excellence for Document Analysis and Recognition (CEDAR), Department

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

An Interactive Intelligent Language Tutor Over The Internet

An Interactive Intelligent Language Tutor Over The Internet An Interactive Intelligent Language Tutor Over The Internet Trude Heift Linguistics Department and Language Learning Centre Simon Fraser University, B.C. Canada V5A1S6 E-mail: heift@sfu.ca Abstract: This

More information

arxiv:cmp-lg/ v1 7 Jun 1997 Abstract

arxiv:cmp-lg/ v1 7 Jun 1997 Abstract Comparing a Linguistic and a Stochastic Tagger Christer Samuelsson Lucent Technologies Bell Laboratories 600 Mountain Ave, Room 2D-339 Murray Hill, NJ 07974, USA christer@research.bell-labs.com Atro Voutilainen

More information

Multi-Lingual Text Leveling

Multi-Lingual Text Leveling Multi-Lingual Text Leveling Salim Roukos, Jerome Quin, and Todd Ward IBM T. J. Watson Research Center, Yorktown Heights, NY 10598 {roukos,jlquinn,tward}@us.ibm.com Abstract. Determining the language proficiency

More information

The Strong Minimalist Thesis and Bounded Optimality

The Strong Minimalist Thesis and Bounded Optimality The Strong Minimalist Thesis and Bounded Optimality DRAFT-IN-PROGRESS; SEND COMMENTS TO RICKL@UMICH.EDU Richard L. Lewis Department of Psychology University of Michigan 27 March 2010 1 Purpose of this

More information

The taming of the data:

The taming of the data: The taming of the data: Using text mining in building a corpus for diachronic analysis Stefania Degaetano-Ortlieb, Hannah Kermes, Ashraf Khamis, Jörg Knappen, Noam Ordan and Elke Teich Background Big data

More information

Calibration of Confidence Measures in Speech Recognition

Calibration of Confidence Measures in Speech Recognition Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE

More information

Comparison of network inference packages and methods for multiple networks inference

Comparison of network inference packages and methods for multiple networks inference Comparison of network inference packages and methods for multiple networks inference Nathalie Villa-Vialaneix http://www.nathalievilla.org nathalie.villa@univ-paris1.fr 1ères Rencontres R - BoRdeaux, 3

More information

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen UNIVERSITY OF OSLO Department of Informatics Dialog Act Recognition using Dependency Features Master s thesis Sindre Wetjen November 15, 2013 Acknowledgments First I want to thank my supervisors Lilja

More information

An Empirical and Computational Test of Linguistic Relativity

An Empirical and Computational Test of Linguistic Relativity An Empirical and Computational Test of Linguistic Relativity Kathleen M. Eberhard* (eberhard.1@nd.edu) Matthias Scheutz** (mscheutz@cse.nd.edu) Michael Heilman** (mheilman@nd.edu) *Department of Psychology,

More information

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly Inflected Languages Classical Approaches to Tagging The slides are posted on the web. The url is http://chss.montclair.edu/~feldmana/esslli10/.

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

Exploiting Wikipedia as External Knowledge for Named Entity Recognition

Exploiting Wikipedia as External Knowledge for Named Entity Recognition Exploiting Wikipedia as External Knowledge for Named Entity Recognition Jun ichi Kazama and Kentaro Torisawa Japan Advanced Institute of Science and Technology (JAIST) Asahidai 1-1, Nomi, Ishikawa, 923-1292

More information

Lecture 10: Reinforcement Learning

Lecture 10: Reinforcement Learning Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation

More information

A Comparison of Two Text Representations for Sentiment Analysis

A Comparison of Two Text Representations for Sentiment Analysis 010 International Conference on Computer Application and System Modeling (ICCASM 010) A Comparison of Two Text Representations for Sentiment Analysis Jianxiong Wang School of Computer Science & Educational

More information

Natural Language Processing. George Konidaris

Natural Language Processing. George Konidaris Natural Language Processing George Konidaris gdk@cs.brown.edu Fall 2017 Natural Language Processing Understanding spoken/written sentences in a natural language. Major area of research in AI. Why? Humans

More information