Disambiguating Levin Verbs Using Untagged Data

Size: px
Start display at page:

Download "Disambiguating Levin Verbs Using Untagged Data"

Transcription

1 Disambiguating Levin Verbs Using Untagged Data Jianguo Li Department of Linguistics The Ohio State University Columbus, Ohio, USA Chris Brew Department of Linguistics The Ohio State University Columbus, Ohio, USA Abstract Lapata and Brew [8] (hereafter LB04) obtain from untagged texts a statistical prior model that is able to generate class preferences for ambiguous Levin [9] verbs. They also show that their informative priors, incorporated into a Naive Bayesian classifier deduced from handtagged data, can aid in verb class disambiguation. We re-examine the parameter estimation of LB04 s prior model and identify the only parameter in LB04 s prior model that determines the predominant class for a particular verb in a particular frame. In addition, we propose a method for training our classifier without using hand-tagged data. Our experiments suggest that although our verb class disambiguator does not match the performance of the ones that make use of hand-tagged data, it consistently outperforms the random baseline model. Our experiments also demonstrate that the informative priors derived from untagged texts help improve the performance of the classifier trained on untagged data. Keywords Keywords: lexical semantics, verb class disambiguation, Levin verb class, informative priors, untagged corpus, Naive Bayesian classifier 1 Introduction Much research in lexical acquisition has concentrated on verb classification [18, 11, 14, 7]. Many scholars hypothesize that the meaning of a verb determines to a large extent its syntactic behavior, particularly the realization and interpretation of its arguments, and therefore base their verb classification on the relation between verbs and their arguments [5, 4, 9, 16]. Such classifications can capture generalizations over a range of linguistic properties, and therefore can be used as a means of reducing redundancy in the lexicon and for filling gaps in lexical knowledge. Much of the work on verb classification in NLP has adopted the classification proposed by Levin [9]. Levin [9] argues that verbs that exhibit the same diathesis alternation can be assumed to share certain semantic components and to form a semantically coherent class. Applying this observation inductively, one can use surface syntactic cues to infer a verb s semantic class. In this paper, we focus on the task of verb classification for ambiguous Levin verbs (verbs belonging to two or more Levin classes). To be exact, given a verb in a particular frame, we want to assign it to one of its possible Levin classes. As noted by Lapata and Brew [8], this is a wide-spread and important disambiguation problem. Levin s verb inventory covers 3,024 verbs. Although only 784 verbs are polysemous verbs, the total frequency of polysemous verbs in the British National Corpus (BNC) is comparable to the total frequency of monosemous verbs (48.4%:51.6%). Consider the verb call in the following two sentences: 1. He called me a fool. 2. He called me a taxi. The verb call is ambiguous between the class of DUB and GET when occurring in the double object frame. We want to automatically identify call as a DUB verb in sentence (1) and a GET verb in sentence (2). The verb class of a particular verb token provides a significant amount of information about the verb. At semantic level, for example, knowing a token s verb class helps determine the thematic role of its arguments [14, 19] and at the syntactic level, it indicates what subcategorization frames and alternations are allowed [9, 6]. Word sense disambiguation (WSD) is usually cast as a problem in supervised learning, where a word class disambiguator is induced from hand-tagged data. The context within which the ambiguous word occurs is typically represented by a set of more or less linguistically-motivated features from which a learning algorithm induces a representative model that performs the disambiguation task. One classifier that has been extensively used is the Naive Bayesian classifier. A Naive Bayesian classifier usually consists of two parts: the prior and the posterior. Lapata and Brew [8] estimate an informative prior over Levin verb classes for a given verb in a given frame, training on untagged texts. Their prior model is able to generate a class preference for an ambiguous verb. Consider the verb call again. It is ambiguous between the class of DUB and GET when occurring in doubleobject frame. Their prior model predicts DUB to be the predominant class. The model s outcome is considered correct since hand-tagged corpus tokens also reveal a preference for the class DUB. To compute the posterior probability, LB04 uses contextual features (e.g. word collocation) extracted from a small handtagged corpus. Their experiments demonstrate that the informative priors obtained from untagged texts helps achieve improved disambiguation performance.

2 The major contribution of LB04 is that it highlights the importance for WSD of a suitable prior derived from untagged text: A prior model derived from untagged texts can help find with reasonable accuracy the predominant sense of a given target ambiguous word. Knowing the predominant sense of a target ambiguous word is valuable as the first sense heuristics which usually serve as a baseline for supervised WSD systems outperform many of these systems which take the surrounding contexts into account. McCarthy et al. [12] have recently also demonstrated the usefulness of a prior in WSD. They use parsed data to find distributionally similar words to the target ambiguous word and then use the associated similarity scores to discover the predominant sense for that target word. One benefit of both LB04 and McCarthy s method is that the predominant senses can be derived without relying on hand-tagged data, which may not be available for every domain and text type. This is important because the frequency of the senses of words depends on the genre and domain of the text under consideration. A prior model derived from untagged texts can also help improve the performance of a classifier over a uniform prior. This is exactly what is shown in LB04. However, although the informative priors in LB04 are derived from untagged texts, the posteriors are deduced from hand-tagged data. Using hand-tagged data to derive the posterior probability assumes the existence of such a corpus. However, if there is a hand-tagged corpus, then an empirical prior can be derived from such a corpus. We would expect a prior obtained from a hand-tagged corpus to be more accurate, therefore when combined with some contextual features should yield better performance. In this paper, we want to evaluate the usefulness of priors derived from a large unlabelled corpus or a small handlabelled corpus. Two experiments are conducted in this paper. First, we examined the estimation of LB04 s prior model because we suspected that some of its parameters are irrelevant to the ultimate outcome of the decision process. This examination confirmed our suspicion. We identified the only parameter that determines the predominant class and reformulated LB04 s prior model accordingly. Our reformulation shows that LB04 s prior model ignores the identity of individual verbs in determining the predominant class for a particular verb. We implemented LB04 s prior model using data parsed by two different full parsers [2, 1]. Second, we proposed a new way to train the verb disambiguator without relying on a hand-tagged corpus. More precisely, we used examples containing unambiguous verbs in a particular verb class as the training data for the ambiguous ones in that class. In doing so, both our informative priors and posteriors are obtained without using hand-tagged data. This method is available even if we are dealing with an unusual text type. We also tested the usefulness of our informative priors in aiding verb class disambiguation. 2 Experiment 1: The Prior Model 2.1 LB04 s Prior Model LB04 s prior model views the choice of a class c for a polysemous verb v in a given frame f as a maximization of the joint probability P(c, f, v), where v is a verb subcategorizing for the frame f with Levin class c: P(c, f, v) = P(v)P(f v)p(c v, f) (1) The estimation of P(c v, f) relies on the frequency of F(c, v, f), which could be obtained if a parsed corpus annotated with semantic class information were available. Without such a corpus, LB04 assumes that the semantic class determines the subcategorization patterns of its members independently of their identity: P(c v, f) P(c f) (2) By applying Bayes rule, P(c f, v) is rewritten as P(c f) = P(f c)p(c) P(f) Substituting (3) into (1), LB04 expresses P(c, f, v) as P(c, v, f) = P(v)P(f v)p(f c)p(c) P(f) (3) (4) 2.2 Examination of LB04 s Parameter Estimation To estimate P(c, f, v), LB04 has to estimate five parameters: P(v), P(f v), P(f), P(f c) and P(c), as shown in (4). However, for a given verb v in a given frame f, the value of P(v), P(f v) and P(f) do not vary with the choice of the class c. If we are only interested in knowing which class c is the predominant class for a given verb in a given frame, we could simply ignore them. Therefore, it is the value of P(f c) and P(c) that determines the predominant class for the verb. According to LB04, P(f c) and P(c) are estimated as P(f c) = P(c) = F(f, c) F(c) (5) F(c) Pi F(ci) (6) With (5) and (6), the value that determines the predominant class for a given verb in a given frame is calculated as F(f, c) F(c) P F(c) F(f, c) = F(ci) Pi F(ci) (7) i The value of the denominator i F(c i) is only a normalizing constant to ensure that we have a probability function. Again, if we are simply interested in which class is the predominant class for a given verb in a given frame, we can ignore it. It turns out that

3 Rank Class F(Class,V-NP) 1 CONT. LOCATION 70,471 2 ADMIRE 66,352 3 HURT 12,730 4 WIPE MANNER 10,294 5 ASSESS 9,872 6 PUSH-PULL 9,828 Table 1: Frequency of six classes with V-NP 1. Equal Distribution: dividing the overall frequency of a verb by the number of classes it belongs to: P(c v) = 1 classes(v) (11) 2. Unequal Distribution: distributing a verb s frequency unequally according to class size: P(c amb class) = c Pc amb class c (12) F(f, c) is the only value that determines the predominant class. According to LB04, F(f, c) is obtained by summing over all occurrences of verbs that are members of class c and attested in the corpus with frame f: F(f, c) = X i F(c,f, v i) (8) For monosemous verbs, F(c, f, v) reduces to the number of times these verbs have been attested in the corpus with a given frame. For polysemous verbs, F(c, f, v) is obtained by dividing the frequency of a verb with the given frame by the number of classes that the verb belongs to when occurring in the given frame. Note that our reformulation of LB04 does not result in a different model. All we did is getting rid of the parameters of LB04 s model that are irrelevant to the decision regarding the predominant class of a verb v in a frame f. Note two facts about the model from LB04: First, due to the independence assumption, the only parameter that matters for the prior model is F(c, f). The identity of a given verb is totally irrelevant. In other words, for a verb v that is ambiguous between class c 1 and c 2 in a given frame f, the predominant class for the verb v is c 1 if F(c 1,f) is greater than F(c 2,f) and c 2 otherwise. Table 1 ranks six verb classes according to their frequency of occurring with the transitive frame. For a verb that is ambiguous between any two of the classes listed in Table 1 when occurring in the transitive frame, the preferred class is determined by the rank of the class in the table. For example, both miss and support are ambiguous between the class AD- MIRE and CONT. LOCATION when occurring in the transitive frame. Since F(CONT. LOCATION, V-NP) is greater than F(ADMIRE, V-NP), the model selects CONT. LOCATION as the predominant class for both verbs. However, the prevalence in the manually annotated corpus data (BNC) suggests that CONT. LOCA- TION is the preferred class for miss while ADMIRE is the preferred class for support. The independence assumption makes it impossible for the model to select the right preferred class for both miss and support. The second fact about LB04 s prior model is that without our reformulation, LB04 has to estimate F(c): F(c) = X i F(c,v i) (9) F(v, c) = F(c)P(c v) (10) LB04 proposes two ways to estimate the value of P(c v): LB04 shows that in selecting the predominant class for a verb in a given frame, for the 34 ambiguous verbs with the transitive frame, its prior model is about 6% better using the equal distribution for the estimation of F(c) than using the unequal distribution. According to our reformulation of its prior model, the value of F(c) is totally irrelevant in choosing the predominant class for a verb. There should be no difference in the performance of the prior model between using equal and unequal distribution to estimate F(c). 2.3 Experiments on the Prior Model Methodology LB04 used a parsed version of the whole BNC made with GSearch [3], a tool that facilitates the search of arbitrary part-of-speech-tagged corpora for shallow syntactic patterns. It used a chunk grammar for recognizing the verbal complex, NPs and PPs, and applied GSearch to extract tokens matching frames specified in Levin. A set of linguistic heuristics were applied to the parser s output in order to filter out unreliable cues. Our implementation used two sets of frames acquired from the whole BNC using two different statistical parsers. (1) We parsed the whole BNC with Charniak s parser. (2) In addition, we obtained the frame set from Schulte im Walde [18]. This frame set was acquired from the whole BNC using a head-entity parser described in Carroll and Rooth (1998) (hereafter CR). We implemented LB04 s prior model (based on our reformulation) using these two separate sets of frames. We obtained test data from LB04. This test data, summarized in Table 2, consists of 5,078 ambiguous verb tokens involving 64 verb types and 3 frame types 1. It includes verbs with double object frame (V-NP- NP) (3.27 average class ambiguity), verbs with dative frame (V-NP-PP(to)) (average 2.94 class ambiguity) and verbs with transitive frame (V-NP) (2.77 average class ambiguity) Results of the Prior Model s Performance We report the results of our implementation of LB04 using accuracy by verb type. This accuracy is the percentage of verb types for which the prior model correctly selects the predominant class. The outcome 1 The test data we used here is not identical to that used in LB04. It has undergone both additional corrections and systematic adjustments before being released to us.

4 Frame Number of Verb Types V-NP-NP 12 V-NP-PP(to) 16 V-NP 34 Table 2: Test data of the prior model is considered correct if the class selected by the prior model agrees with the most frequent class found in the hand-tagged corpus. Table 3 provides a summary of the results for our implementation of LB04 s prior model. We also computed a baseline by randomly selecting a class out of all the possible classes for a given verb in a particular frame 2. Parser Charniak CR LB % 56.4% Baseline 39.7%±0.01 Table 3: Type accuracy for the prior model Table 3 shows that our reformulation of LB04 s prior model achieves a better performance (using either set of frame frequency) than the baseline. However, our results are lower than that reported in LB04. LB04 s prior model achieves an accuracy of 74.6%. This may be due to the different test data we used and different parsers we used to obtain frame frequency. 3 Experiment 2: Verb Class Disambiguation Using Untagged Texts 3.1 Motivation Recall that LB04 derives the informative priors from untagged texts, but the posteriors from a hand-tagged corpus. In this experiment, we attempt to address this weakness of LB04 s method: LB04 does not compare the performance of the Naive Bayesian classifier between using the informative priors derived from untagged texts (IPrior) and the empirical priors derived from a hand-tagged corpus (EPrior). It would be helpful to know if an IPrior outperforms an EPrior estimated from a very small hand-tagged corpus. LB04 derives IPrior from untagged texts. If the posteriors can also be deduced without using hand-tagged data, it will free us from our dependence on handtagged data for disambiguating Levin verbs. As noted above, Levin [9] has classified verbs according to their syntactic behavior. Verbs that show similar diathesis alternation are assumed to share certain semantic components and to form a coherent semantic class. Neighboring words are not taken into consideration in her verb classification. On the other hand, many scholars have shown that words with similar contextual features, typically neighboring words, are also semantically similar [17, 10]. Faced with these two dif- 2 We replicated this random selection 100 times and the result reported in Table 3 was obtained by averaging the results on the 100 selections. class ambiguous verbs unambiguous verbs anoint, baptize, brand, christen DUB call, make consecrate, crown, decree, dub vote name, nickname, pronounce, rule stamp, style, term book, buy, cash, catch charter, choose, earn, fetch GET call, find gain, gather, hire, keep leave, vote order, phone, pick, pluck procure, pull, reach, rent reserve, save, secure, shoot slaughter, steal, win Table 4: DUB and GET class ferent approaches to identifying semantically similar words, we may ask the following two questions: Are the semantic components shared by verbs in a Levin class correlated with their contexts words? Can we use the context words of the unambiguous verbs in a particular Levin class to disambiguate the ambiguous verbs in that class? To perform verb class disambiguation without relying on a hand-tagged corpus, we decided to train our verb class disambiguator using only data containing unambiguous verbs. Consider the verb call again, it is ambiguous between the class of DUB and GET when occurring in the double-object frame. However, most verbs in these two classes are not ambiguous, as shown in Table 4. For an unambiguous verb, we know for sure the class it belongs to without even examining the sentences in which it occurs. To disambiguate call in a double-object frame, we therefore used all sentences that are identified as double object frame and contain an unambiguous verb in the class DUB as the training data for the class DUB and did the same for the class GET. 3.2 Constructing Training Data We picked all the example sentences containing the relevant unambiguous verbs from Charniak-parsed BNC that are identified as double-object frame, transitive frame or dative frame. We understand that the training data constructed this way is noisy in that some false instances of the target frames are included in the training data. For example, a sentence like I fed the boy myself is incorrectly recognized as a double-object frame. Thus the training data we used may potentially have a negative effect on the verb class disambiguator. 3.3 Classifier and Feature Space A Naive Bayesian Classifier We employed a Naive Bayesian classifier for our disambiguation task. Although the Naive Bayesian classifier is simple, it is quite efficient and has shown good performance on WSD. Another reason for using a Naive Bayesian classifier is that it is easy to incorporate the prior information. Within a Naive Bayesian approach, the choice of the predominant class for an ambiguous verb v when occurring in a frame f given its context can be expressed as

5 C(v, f) = argmax ci [P(c i, f, v) n P(a 1,..., a n c i, f, v)] (13) i=1 Where C(v,f) represents the predominant class for an ambiguous verb v when occurring in a frame f. P(c i, f, v) is the prior probability of the ambiguous verb v belonging to class c i when occurring in frame f and n i=1 P(a 1,..., a n c i, f, v) is the posterior probability Feature Space As common in WSD, we used as features the neighboring words of a target ambiguous verb. We considered 8 different window sizes: L1R1, L1R2, L1R3, L1R4, L2R1, L2R2, L2R3 and L2R4. A window size such as L1R2 represents one word to the left and two words to the right of an ambiguous verb. Neighboring words are lemmatized using the English lemmatizer described in [15]. 3.4 Results and Discussion We used the same test data from the first experiment. We compare the performance of six different models. They differ from each other in whether the priors are derived from hand-tagged data and whether the classifier is trained on hand-tagged data. Prior IPrior: The informative priors derived from untagged texts as described in experiment 1. EPrior: The empirical priors derived from handtagged data. In our experiment, the empirical priors are derived from the test examples. UPrior: The uniform priors. Classifier NHTD: The classifier is trained without using hand-tagged data. In our experiment, the training data consists of all the examples containing only unambiguous verbs. The classifier is tested on all test examples. HTD: The classifier is trained on hand-tagged data. In our experiment, the classifier is trained and tested using 10-fold cross-validation on the test examples. The six models we experimented with are as follows: UPrior+NHTD, IPrior+NHTD, EPrior+NHTD, UPrior+HTD, IPrior+HTD and EPrior+HTD 3. For the purpose of comparison, we also report the performance of three different baseline models: Random Baseline (RB): We randomly selected a class from all those that are compatible with the given verb and frame. Selection was based on a unifrom distribution. 3 We also estimated a prior from the unambiguous examples only, but its performance is about the same as the IPrior. model average accuracy highest accuracy UPrior+NHTD 58.1% 64.8%(L1R3) IPrior+NHTD 62.3% 68.8%(L1R4) EPrior+NHTD 64.1% 71.0%(L2R4) UPrior+HTD 64.2% 72.3%(L1R4) IPrior+HTD 64.9% 73.9%(L2R4) EPrior+HTD 68.9% 77.4%(L2R4) RB 37.9% IPB 57.9% EPB 74.2% Table 5: Results for verb class disambiguation IPrior Baseline (IPB): We selected the class whose IPrior was the largest of the available possibilities. EPrior Baseline (EPB): We selected the class whose EPrior was the largest of the available possibilities. The results are summarized in Table 5. The average accuracy was obtained by averaging the accuracy over all 8 window sizes. We also report the highest accuracy and the window size where the highest accuracy were achieved. For example, using the window size L1R3 (see Table 5) the model UPrior+NHTD achieves its best performance of 64.8%. Several things are worth noting in the result Table 5: When the classifier is trained on hand-tagged data (HTD), using IPrior(IPrior+HTD) outperforms using UPrior(UPrior+HTD). This agrees with what is shown in LB04. However, using IPrior (IPrior+HTD) does not match the performance of using EPrior (EPrior+HTD). Both the average accuracy and the highest accuracy for model EPrior+HTD are higher than IPrior+HTD. A pair-wise t-test indicates that the difference is statistically significant (p-value = 0.021). This suggests that it is not always best to incorporate a prior derived from untagged texts into a classifier trained on hand-tagged data. It is better to derive a prior from a hand-tagged corpus if such a corpus is available. When the classifier is trained without using handtagged data (NHTD), neither using UPrior (UPrior+NHTD) nor using IPrior (IPrior+NHTD) performs better than any of the supervised models (HTD). However, they both (UPrior+NHTD and IPrior+NHTD) consistently outperform the random baseline, suggesting that verbs in the same Levin class do tend to share their context words. Our verb class disambiguator using untagged data can be used in the absence of a hand-tagged corpus. In addition, using the EPrior (EPrior+NHTD) achieves a performance comparable to that achieved by the supervised model with a UPrior (UPrior+HTD), suggesting a tagged corpus, if available, helps derive more accurate priors. When the classifier is trained without using handtagged data (NHTD), using IPrior (IPrior+NHTD) outperforms using UPrior (UPrior+NHTD). A pairwise t-test indicates that the improvement achieved by using IPrior is statistically significant (p-value = 0.026), suggesting that the IPrior derived from untagged data, thought not as accurate as the EPrior, can still can still help improve the performance of the classifier. All five models we experimented with outperform the IPB, but fail to achieve the performance of the EPB

6 with the exception of the model EPrior+HTD, the highest accuracy of which is about 3% better than the EPB. Again, annotation, if available, helps. 4 Conclusions and Future Work The main conclusions of this paper are the following: Our experiments confirm the importance of syntactic frame information in verb class disambiguation. In addition, we have also re-confirmed the importance of a good prior derived from untagged texts in WSD. However, instead of deriving the classifier from handtagged data like what LB04 did, we trained our classifier using examples containing unambiguous verbs. This offers us a way to disambiguate Levin verbs without relying on hand-tagged data. 4.1 About the Prior Model A contribution of our paper is a clearer reformulation of LB04 s prior model. This reveals that LB04 s prior model cannot distinguish between different verbs of the same class. This is a direct result of the independence assumption built into the model. To improve the performance of the prior model, we believe it is worthwhile finding new ways to bring the identity of each individual verb into the prior model [13]. 4.2 Disambiguation without a Handtagged Corpus We proposed a method for disambiguating Levin verbs that completely avoids the need for a hand-tagged corpus and analysed how it compares to various alternatives. Our experiments show that our verb class disambiguator is not as accurate as the supervised ones that make use of a hand-tagged corpus. One reason is that we relied on a statistical parser for identifying the target frames (double-object, transitive and dative) in constructing the training data. The training data obtained this way is noisy in that some false instances of the target frames are included. On the other hand, the training data (in this case it is the 5,078 test examples) used to train the supervised models has been examined by human annotators and is free of any false instances of the target frames. However, our method of disambiguating Levin verbs without using hand-tagged data consistently outperforms the random baseline, suggesting that it is feasible to use examples containing unambiguous verbs to disambiguate ambiguous ones. Levin s verb classification covers about 79 frames and many of them involve some ambiguity. In this paper, we only tested our verb class disambiguator on three of Levin s frames. It remains to be shown that it works equally well for other frames. We also plan to test our disambiguation method, namely using unambiguous words to disambiguate ambiguous ones, on different WSD data sets. 5 Acknowledgments This study was supported by NSF grant We are grateful to Mirella Lapata for providing the test data. Our thanks also go to Sabine Schulte im Walde for making available to us the frame set she acquired from BNC. References [1] G. Carroll and M. Rooth. Valence induction with a headlexicalized PCFG. In Proceedings of 3rd Conference of Empirical Methods of Natural Language Processing, pages 58 63, [2] E. Charniak. A maximum-entropy-inspired parser. In Proceedings of the 2000 Conference of the North American Chapter of the Association for Computation Linguistics, pages , [3] S. Corley, M. Corley, F. Keller, M. Cocker, and S. Trewin. Finding syntactic structure in unparsed corpora. Computers and the Humanities, 35(2):81 94, [4] A. Goldberg. Constructions. University of Chicago Press, Chicago, 1st edition, [5] R. Jackendoff. Semantics and Cognition. MIT Press, Cambridge, MA, [6] A. Korhonen. Subcategorization Acquisition. PhD thesis, Cambridge University, [7] A. Korhonen, Y. Krymolowski, and Z. Marx. Clustering polysemic subcategorization frame distributions semantically. In Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics, pages 48 55, [8] M. Lapata and C. Brew. Verb class disambiguation using informative priors. Computational Linguistics, 30(1):45 73, [9] B. Levin. English Verb Classes and Alternations: A Preliminary Investigation. University of Chicago Press, Chicago, 1st edition, [10] D. Lin. Automatic retrieval and clustering of similar words. In COLING-ACL 98, [11] D. McCarthy. Using semantic preference for identifying verbal participation in role switching alternations. In Proceedings of the 1st Conference of the North American Chapter of the Association for Computational Linguistics, pages 58 63, [12] D. McCarthy, R. Koeling, J. Weeds, and J. Carroll. Finding predominant senses in untagged text. In Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics, pages , [13] P. Merlo, E. Joanis, and J. Henderson. Unsupervised verb class disambiguation based on diathesis alternations. manuscripts, [14] P. Merlo and S. Stevenson. Automatic verb classification based on statistical distribution of argument structure. Computational Linguistics, 27(3): , [15] G. Minnen, J. Carroll, and D. Pearce. Applied morphological processing of English. Natural Language Engineering, 7(3): , [16] S. Pinker. Learnability and Cognition: The Acquisition of Argument Structure. MIT Press, Cambridge, MA, [17] D. Rohde, L. Gonnerman, and D. Plaut. An improved method for deriving word meaning from lexical co-occurrence. Cognitive Science, submitted. [18] S. Schulte im Walde. Clustering verbs semantically according to alternation behavior. In Proceedings of the 18th International Conference on Computational Linguistics, pages , [19] R. Swier and S. Stevenson. Unsupervised semantic role labelling. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, pages , 2004.

A Case Study: News Classification Based on Term Frequency

A Case Study: News Classification Based on Term Frequency A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center

More information

The Choice of Features for Classification of Verbs in Biomedical Texts

The Choice of Features for Classification of Verbs in Biomedical Texts The Choice of Features for Classification of Verbs in Biomedical Texts Anna Korhonen University of Cambridge Computer Laboratory 15 JJ Thomson Avenue Cambridge CB3 0FD, UK alk23@cl.cam.ac.uk Yuval Krymolowski

More information

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures Ulrike Baldewein (ulrike@coli.uni-sb.de) Computational Psycholinguistics, Saarland University D-66041 Saarbrücken,

More information

Can Human Verb Associations help identify Salient Features for Semantic Verb Classification?

Can Human Verb Associations help identify Salient Features for Semantic Verb Classification? Can Human Verb Associations help identify Salient Features for Semantic Verb Classification? Sabine Schulte im Walde Institut für Maschinelle Sprachverarbeitung Universität Stuttgart Seminar für Sprachwissenschaft,

More information

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data Ebba Gustavii Department of Linguistics and Philology, Uppsala University, Sweden ebbag@stp.ling.uu.se

More information

Proceedings of the 19th COLING, , 2002.

Proceedings of the 19th COLING, , 2002. Crosslinguistic Transfer in Automatic Verb Classication Vivian Tsang Computer Science University of Toronto vyctsang@cs.toronto.edu Suzanne Stevenson Computer Science University of Toronto suzanne@cs.toronto.edu

More information

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING SISOM & ACOUSTICS 2015, Bucharest 21-22 May THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING MarilenaăLAZ R 1, Diana MILITARU 2 1 Military Equipment and Technologies Research Agency, Bucharest,

More information

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus Language Acquisition Fall 2010/Winter 2011 Lexical Categories Afra Alishahi, Heiner Drenhaus Computational Linguistics and Phonetics Saarland University Children s Sensitivity to Lexical Categories Look,

More information

Linking Task: Identifying authors and book titles in verbose queries

Linking Task: Identifying authors and book titles in verbose queries Linking Task: Identifying authors and book titles in verbose queries Anaïs Ollagnier, Sébastien Fournier, and Patrice Bellot Aix-Marseille University, CNRS, ENSAM, University of Toulon, LSIS UMR 7296,

More information

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17. Semi-supervised methods of text processing, and an application to medical concept extraction Yacine Jernite Text-as-Data series September 17. 2015 What do we want from text? 1. Extract information 2. Link

More information

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

The Internet as a Normative Corpus: Grammar Checking with a Search Engine The Internet as a Normative Corpus: Grammar Checking with a Search Engine Jonas Sjöbergh KTH Nada SE-100 44 Stockholm, Sweden jsh@nada.kth.se Abstract In this paper some methods using the Internet as a

More information

Prediction of Maximal Projection for Semantic Role Labeling

Prediction of Maximal Projection for Semantic Role Labeling Prediction of Maximal Projection for Semantic Role Labeling Weiwei Sun, Zhifang Sui Institute of Computational Linguistics Peking University Beijing, 100871, China {ws, szf}@pku.edu.cn Haifeng Wang Toshiba

More information

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization Annemarie Friedrich, Marina Valeeva and Alexis Palmer COMPUTATIONAL LINGUISTICS & PHONETICS SAARLAND UNIVERSITY, GERMANY

More information

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics (L615) Markus Dickinson Department of Linguistics, Indiana University Spring 2013 The web provides new opportunities for gathering data Viable source of disposable corpora, built ad hoc for specific purposes

More information

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar Chung-Chi Huang Mei-Hua Chen Shih-Ting Huang Jason S. Chang Institute of Information Systems and Applications, National Tsing Hua University,

More information

Handling Sparsity for Verb Noun MWE Token Classification

Handling Sparsity for Verb Noun MWE Token Classification Handling Sparsity for Verb Noun MWE Token Classification Mona T. Diab Center for Computational Learning Systems Columbia University mdiab@ccls.columbia.edu Madhav Krishna Computer Science Department Columbia

More information

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities Yoav Goldberg Reut Tsarfaty Meni Adler Michael Elhadad Ben Gurion

More information

AQUA: An Ontology-Driven Question Answering System

AQUA: An Ontology-Driven Question Answering System AQUA: An Ontology-Driven Question Answering System Maria Vargas-Vera, Enrico Motta and John Domingue Knowledge Media Institute (KMI) The Open University, Walton Hall, Milton Keynes, MK7 6AA, United Kingdom.

More information

Using dialogue context to improve parsing performance in dialogue systems

Using dialogue context to improve parsing performance in dialogue systems Using dialogue context to improve parsing performance in dialogue systems Ivan Meza-Ruiz and Oliver Lemon School of Informatics, Edinburgh University 2 Buccleuch Place, Edinburgh I.V.Meza-Ruiz@sms.ed.ac.uk,

More information

Applications of memory-based natural language processing

Applications of memory-based natural language processing Applications of memory-based natural language processing Antal van den Bosch and Roser Morante ILK Research Group Tilburg University Prague, June 24, 2007 Current ILK members Principal investigator: Antal

More information

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases POS Tagging Problem Part-of-Speech Tagging L545 Spring 203 Given a sentence W Wn and a tagset of lexical categories, find the most likely tag T..Tn for each word in the sentence Example Secretariat/P is/vbz

More information

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence. NLP Lab Session Week 8 October 15, 2014 Noun Phrase Chunking and WordNet in NLTK Getting Started In this lab session, we will work together through a series of small examples using the IDLE window and

More information

The MEANING Multilingual Central Repository

The MEANING Multilingual Central Repository The MEANING Multilingual Central Repository J. Atserias, L. Villarejo, G. Rigau, E. Agirre, J. Carroll, B. Magnini, P. Vossen January 27, 2004 http://www.lsi.upc.es/ nlp/meaning Jordi Atserias TALP Index

More information

The stages of event extraction

The stages of event extraction The stages of event extraction David Ahn Intelligent Systems Lab Amsterdam University of Amsterdam ahn@science.uva.nl Abstract Event detection and recognition is a complex task consisting of multiple sub-tasks

More information

The Smart/Empire TIPSTER IR System

The Smart/Empire TIPSTER IR System The Smart/Empire TIPSTER IR System Chris Buckley, Janet Walz Sabir Research, Gaithersburg, MD chrisb,walz@sabir.com Claire Cardie, Scott Mardis, Mandar Mitra, David Pierce, Kiri Wagstaff Department of

More information

Learning Computational Grammars

Learning Computational Grammars Learning Computational Grammars John Nerbonne, Anja Belz, Nicola Cancedda, Hervé Déjean, James Hammerton, Rob Koeling, Stasinos Konstantopoulos, Miles Osborne, Franck Thollard and Erik Tjong Kim Sang Abstract

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Chinese Language Parsing with Maximum-Entropy-Inspired Parser Chinese Language Parsing with Maximum-Entropy-Inspired Parser Heng Lian Brown University Abstract The Chinese language has many special characteristics that make parsing difficult. The performance of state-of-the-art

More information

Unsupervised and Constrained Dirichlet Process Mixture Models for Verb Clustering

Unsupervised and Constrained Dirichlet Process Mixture Models for Verb Clustering Unsupervised and Constrained Dirichlet Process Mixture Models for Verb Clustering Andreas Vlachos Computer Laboratory University of Cambridge Cambridge CB3 0FD, UK av308l@cl.cam.ac.uk Anna Korhonen Computer

More information

A Bayesian Learning Approach to Concept-Based Document Classification

A Bayesian Learning Approach to Concept-Based Document Classification Databases and Information Systems Group (AG5) Max-Planck-Institute for Computer Science Saarbrücken, Germany A Bayesian Learning Approach to Concept-Based Document Classification by Georgiana Ifrim Supervisors

More information

Search right and thou shalt find... Using Web Queries for Learner Error Detection

Search right and thou shalt find... Using Web Queries for Learner Error Detection Search right and thou shalt find... Using Web Queries for Learner Error Detection Michael Gamon Claudia Leacock Microsoft Research Butler Hill Group One Microsoft Way P.O. Box 935 Redmond, WA 981052, USA

More information

Multilingual Sentiment and Subjectivity Analysis

Multilingual Sentiment and Subjectivity Analysis Multilingual Sentiment and Subjectivity Analysis Carmen Banea and Rada Mihalcea Department of Computer Science University of North Texas rada@cs.unt.edu, carmen.banea@gmail.com Janyce Wiebe Department

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Developing a TT-MCTAG for German with an RCG-based Parser

Developing a TT-MCTAG for German with an RCG-based Parser Developing a TT-MCTAG for German with an RCG-based Parser Laura Kallmeyer, Timm Lichte, Wolfgang Maier, Yannick Parmentier, Johannes Dellert University of Tübingen, Germany CNRS-LORIA, France LREC 2008,

More information

Which verb classes and why? Research questions: Semantic Basis Hypothesis (SBH) What verb classes? Why the truth of the SBH matters

Which verb classes and why? Research questions: Semantic Basis Hypothesis (SBH) What verb classes? Why the truth of the SBH matters Which verb classes and why? ean-pierre Koenig, Gail Mauner, Anthony Davis, and reton ienvenue University at uffalo and Streamsage, Inc. Research questions: Participant roles play a role in the syntactic

More information

Memory-based grammatical error correction

Memory-based grammatical error correction Memory-based grammatical error correction Antal van den Bosch Peter Berck Radboud University Nijmegen Tilburg University P.O. Box 9103 P.O. Box 90153 NL-6500 HD Nijmegen, The Netherlands NL-5000 LE Tilburg,

More information

An Interactive Intelligent Language Tutor Over The Internet

An Interactive Intelligent Language Tutor Over The Internet An Interactive Intelligent Language Tutor Over The Internet Trude Heift Linguistics Department and Language Learning Centre Simon Fraser University, B.C. Canada V5A1S6 E-mail: heift@sfu.ca Abstract: This

More information

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages Nuanwan Soonthornphisaj 1 and Boonserm Kijsirikul 2 Machine Intelligence and Knowledge Discovery Laboratory Department of Computer

More information

Measuring the relative compositionality of verb-noun (V-N) collocations by integrating features

Measuring the relative compositionality of verb-noun (V-N) collocations by integrating features Measuring the relative compositionality of verb-noun (V-N) collocations by integrating features Sriram Venkatapathy Language Technologies Research Centre, International Institute of Information Technology

More information

Word Segmentation of Off-line Handwritten Documents

Word Segmentation of Off-line Handwritten Documents Word Segmentation of Off-line Handwritten Documents Chen Huang and Sargur N. Srihari {chuang5, srihari}@cedar.buffalo.edu Center of Excellence for Document Analysis and Recognition (CEDAR), Department

More information

Procedia - Social and Behavioral Sciences 141 ( 2014 ) WCLTA Using Corpus Linguistics in the Development of Writing

Procedia - Social and Behavioral Sciences 141 ( 2014 ) WCLTA Using Corpus Linguistics in the Development of Writing Available online at www.sciencedirect.com ScienceDirect Procedia - Social and Behavioral Sciences 141 ( 2014 ) 124 128 WCLTA 2013 Using Corpus Linguistics in the Development of Writing Blanka Frydrychova

More information

! # %& ( ) ( + ) ( &, % &. / 0!!1 2/.&, 3 ( & 2/ &,

! # %& ( ) ( + ) ( &, % &. / 0!!1 2/.&, 3 ( & 2/ &, ! # %& ( ) ( + ) ( &, % &. / 0!!1 2/.&, 3 ( & 2/ &, 4 The Interaction of Knowledge Sources in Word Sense Disambiguation Mark Stevenson Yorick Wilks University of Shef eld University of Shef eld Word sense

More information

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models Netpix: A Method of Feature Selection Leading to Accurate Sentiment-Based Classification Models 1 Netpix: A Method of Feature Selection Leading to Accurate Sentiment-Based Classification Models James B.

More information

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Twitter Sentiment Classification on Sanders Data using Hybrid Approach IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 4, Ver. I (July Aug. 2015), PP 118-123 www.iosrjournals.org Twitter Sentiment Classification on Sanders

More information

Reducing Features to Improve Bug Prediction

Reducing Features to Improve Bug Prediction Reducing Features to Improve Bug Prediction Shivkumar Shivaji, E. James Whitehead, Jr., Ram Akella University of California Santa Cruz {shiv,ejw,ram}@soe.ucsc.edu Sunghun Kim Hong Kong University of Science

More information

Short Text Understanding Through Lexical-Semantic Analysis

Short Text Understanding Through Lexical-Semantic Analysis Short Text Understanding Through Lexical-Semantic Analysis Wen Hua #1, Zhongyuan Wang 2, Haixun Wang 3, Kai Zheng #4, Xiaofang Zhou #5 School of Information, Renmin University of China, Beijing, China

More information

Switchboard Language Model Improvement with Conversational Data from Gigaword

Switchboard Language Model Improvement with Conversational Data from Gigaword Katholieke Universiteit Leuven Faculty of Engineering Master in Artificial Intelligence (MAI) Speech and Language Technology (SLT) Switchboard Language Model Improvement with Conversational Data from Gigaword

More information

Review in ICAME Journal, Volume 38, 2014, DOI: /icame

Review in ICAME Journal, Volume 38, 2014, DOI: /icame Review in ICAME Journal, Volume 38, 2014, DOI: 10.2478/icame-2014-0012 Gaëtanelle Gilquin and Sylvie De Cock (eds.). Errors and disfluencies in spoken corpora. Amsterdam: John Benjamins. 2013. 172 pp.

More information

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS Julia Tmshkina Centre for Text Techitology, North-West University, 253 Potchefstroom, South Africa 2025770@puk.ac.za

More information

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY Chen, Hsin-Hsi Department of Computer Science and Information Engineering National Taiwan University Taipei, Taiwan E-mail: hh_chen@csie.ntu.edu.tw Abstract

More information

LEARNING A SEMANTIC PARSER FROM SPOKEN UTTERANCES. Judith Gaspers and Philipp Cimiano

LEARNING A SEMANTIC PARSER FROM SPOKEN UTTERANCES. Judith Gaspers and Philipp Cimiano LEARNING A SEMANTIC PARSER FROM SPOKEN UTTERANCES Judith Gaspers and Philipp Cimiano Semantic Computing Group, CITEC, Bielefeld University {jgaspers cimiano}@cit-ec.uni-bielefeld.de ABSTRACT Semantic parsers

More information

Ensemble Technique Utilization for Indonesian Dependency Parser

Ensemble Technique Utilization for Indonesian Dependency Parser Ensemble Technique Utilization for Indonesian Dependency Parser Arief Rahman Institut Teknologi Bandung Indonesia 23516008@std.stei.itb.ac.id Ayu Purwarianti Institut Teknologi Bandung Indonesia ayu@stei.itb.ac.id

More information

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy Informatics 2A: Language Complexity and the Chomsky Hierarchy September 28, 2010 Starter 1 Is there a finite state machine that recognises all those strings s from the alphabet {a, b} where the difference

More information

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

OCR for Arabic using SIFT Descriptors With Online Failure Prediction OCR for Arabic using SIFT Descriptors With Online Failure Prediction Andrey Stolyarenko, Nachum Dershowitz The Blavatnik School of Computer Science Tel Aviv University Tel Aviv, Israel Email: stloyare@tau.ac.il,

More information

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

Project in the framework of the AIM-WEST project Annotation of MWEs for translation Project in the framework of the AIM-WEST project Annotation of MWEs for translation 1 Agnès Tutin LIDILEM/LIG Université Grenoble Alpes 30 october 2014 Outline 2 Why annotate MWEs in corpora? A first experiment

More information

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments Cristina Vertan, Walther v. Hahn University of Hamburg, Natural Language Systems Division Hamburg,

More information

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Exploration. CS : Deep Reinforcement Learning Sergey Levine Exploration CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 4 due on Wednesday 2. Project proposal feedback sent Today s Lecture 1. What is exploration? Why is it a problem?

More information

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Innov High Educ (2009) 34:93 103 DOI 10.1007/s10755-009-9095-2 Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Phyllis Blumberg Published online: 3 February

More information

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2 Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2 Ted Pedersen Department of Computer Science University of Minnesota Duluth, MN, 55812 USA tpederse@d.umn.edu

More information

Graph Alignment for Semi-Supervised Semantic Role Labeling

Graph Alignment for Semi-Supervised Semantic Role Labeling Graph Alignment for Semi-Supervised Semantic Role Labeling Hagen Fürstenau Dept. of Computational Linguistics Saarland University Saarbrücken, Germany hagenf@coli.uni-saarland.de Mirella Lapata School

More information

Accurate Unlexicalized Parsing for Modern Hebrew

Accurate Unlexicalized Parsing for Modern Hebrew Accurate Unlexicalized Parsing for Modern Hebrew Reut Tsarfaty and Khalil Sima an Institute for Logic, Language and Computation, University of Amsterdam Plantage Muidergracht 24, 1018TV Amsterdam, The

More information

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation tatistical Parsing (Following slides are modified from Prof. Raymond Mooney s slides.) tatistical Parsing tatistical parsing uses a probabilistic model of syntax in order to assign probabilities to each

More information

A heuristic framework for pivot-based bilingual dictionary induction

A heuristic framework for pivot-based bilingual dictionary induction 2013 International Conference on Culture and Computing A heuristic framework for pivot-based bilingual dictionary induction Mairidan Wushouer, Toru Ishida, Donghui Lin Department of Social Informatics,

More information

Combining a Chinese Thesaurus with a Chinese Dictionary

Combining a Chinese Thesaurus with a Chinese Dictionary Combining a Chinese Thesaurus with a Chinese Dictionary Ji Donghong Kent Ridge Digital Labs 21 Heng Mui Keng Terrace Singapore, 119613 dhji @krdl.org.sg Gong Junping Department of Computer Science Ohio

More information

Disambiguation of Thai Personal Name from Online News Articles

Disambiguation of Thai Personal Name from Online News Articles Disambiguation of Thai Personal Name from Online News Articles Phaisarn Sutheebanjard Graduate School of Information Technology Siam University Bangkok, Thailand mr.phaisarn@gmail.com Abstract Since online

More information

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Good Judgment Project: A large scale test of different methods of combining expert predictions The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania

More information

Bootstrapping and Evaluating Named Entity Recognition in the Biomedical Domain

Bootstrapping and Evaluating Named Entity Recognition in the Biomedical Domain Bootstrapping and Evaluating Named Entity Recognition in the Biomedical Domain Andreas Vlachos Computer Laboratory University of Cambridge Cambridge, CB3 0FD, UK av308@cl.cam.ac.uk Caroline Gasperin Computer

More information

The Ups and Downs of Preposition Error Detection in ESL Writing

The Ups and Downs of Preposition Error Detection in ESL Writing The Ups and Downs of Preposition Error Detection in ESL Writing Joel R. Tetreault Educational Testing Service 660 Rosedale Road Princeton, NJ, USA JTetreault@ets.org Martin Chodorow Hunter College of CUNY

More information

A Statistical Approach to the Semantics of Verb-Particles

A Statistical Approach to the Semantics of Verb-Particles A Statistical Approach to the Semantics of Verb-Particles Colin Bannard School of Informatics University of Edinburgh 2 Buccleuch Place Edinburgh EH8 9LW, UK c.j.bannard@ed.ac.uk Timothy Baldwin CSLI Stanford

More information

Unsupervised Learning of Narrative Schemas and their Participants

Unsupervised Learning of Narrative Schemas and their Participants Unsupervised Learning of Narrative Schemas and their Participants Nathanael Chambers and Dan Jurafsky Stanford University, Stanford, CA 94305 {natec,jurafsky}@stanford.edu Abstract We describe an unsupervised

More information

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF) SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF) Hans Christian 1 ; Mikhael Pramodana Agus 2 ; Derwin Suhartono 3 1,2,3 Computer Science Department,

More information

Parsing of part-of-speech tagged Assamese Texts

Parsing of part-of-speech tagged Assamese Texts IJCSI International Journal of Computer Science Issues, Vol. 6, No. 1, 2009 ISSN (Online): 1694-0784 ISSN (Print): 1694-0814 28 Parsing of part-of-speech tagged Assamese Texts Mirzanur Rahman 1, Sufal

More information

BYLINE [Heng Ji, Computer Science Department, New York University,

BYLINE [Heng Ji, Computer Science Department, New York University, INFORMATION EXTRACTION BYLINE [Heng Ji, Computer Science Department, New York University, hengji@cs.nyu.edu] SYNONYMS NONE DEFINITION Information Extraction (IE) is a task of extracting pre-specified types

More information

TextGraphs: Graph-based algorithms for Natural Language Processing

TextGraphs: Graph-based algorithms for Natural Language Processing HLT-NAACL 06 TextGraphs: Graph-based algorithms for Natural Language Processing Proceedings of the Workshop Production and Manufacturing by Omnipress Inc. 2600 Anderson Street Madison, WI 53704 c 2006

More information

LEXICAL COHESION ANALYSIS OF THE ARTICLE WHAT IS A GOOD RESEARCH PROJECT? BY BRIAN PALTRIDGE A JOURNAL ARTICLE

LEXICAL COHESION ANALYSIS OF THE ARTICLE WHAT IS A GOOD RESEARCH PROJECT? BY BRIAN PALTRIDGE A JOURNAL ARTICLE LEXICAL COHESION ANALYSIS OF THE ARTICLE WHAT IS A GOOD RESEARCH PROJECT? BY BRIAN PALTRIDGE A JOURNAL ARTICLE Submitted in partial fulfillment of the requirements for the degree of Sarjana Sastra (S.S.)

More information

arxiv: v1 [cs.cl] 2 Apr 2017

arxiv: v1 [cs.cl] 2 Apr 2017 Word-Alignment-Based Segment-Level Machine Translation Evaluation using Word Embeddings Junki Matsuo and Mamoru Komachi Graduate School of System Design, Tokyo Metropolitan University, Japan matsuo-junki@ed.tmu.ac.jp,

More information

Lexical category induction using lexically-specific templates

Lexical category induction using lexically-specific templates Lexical category induction using lexically-specific templates Richard E. Leibbrandt and David M. W. Powers Flinders University of South Australia 1. The induction of lexical categories from distributional

More information

Using Semantic Relations to Refine Coreference Decisions

Using Semantic Relations to Refine Coreference Decisions Using Semantic Relations to Refine Coreference Decisions Heng Ji David Westbrook Ralph Grishman Department of Computer Science New York University New York, NY, 10003, USA hengji@cs.nyu.edu westbroo@cs.nyu.edu

More information

Using computational modeling in language acquisition research

Using computational modeling in language acquisition research Chapter 8 Using computational modeling in language acquisition research Lisa Pearl 1. Introduction Language acquisition research is often concerned with questions of what, when, and how what children know,

More information

Distant Supervised Relation Extraction with Wikipedia and Freebase

Distant Supervised Relation Extraction with Wikipedia and Freebase Distant Supervised Relation Extraction with Wikipedia and Freebase Marcel Ackermann TU Darmstadt ackermann@tk.informatik.tu-darmstadt.de Abstract In this paper we discuss a new approach to extract relational

More information

Defragmenting Textual Data by Leveraging the Syntactic Structure of the English Language

Defragmenting Textual Data by Leveraging the Syntactic Structure of the English Language Defragmenting Textual Data by Leveraging the Syntactic Structure of the English Language Nathaniel Hayes Department of Computer Science Simpson College 701 N. C. St. Indianola, IA, 50125 nate.hayes@my.simpson.edu

More information

Word Sense Disambiguation

Word Sense Disambiguation Word Sense Disambiguation D. De Cao R. Basili Corso di Web Mining e Retrieval a.a. 2008-9 May 21, 2009 Excerpt of the R. Mihalcea and T. Pedersen AAAI 2005 Tutorial, at: http://www.d.umn.edu/ tpederse/tutorials/advances-in-wsd-aaai-2005.ppt

More information

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels Jörg Tiedemann Uppsala University Department of Linguistics and Philology firstname.lastname@lingfil.uu.se Abstract

More information

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks POS tagging of Chinese Buddhist texts using Recurrent Neural Networks Longlu Qin Department of East Asian Languages and Cultures longlu@stanford.edu Abstract Chinese POS tagging, as one of the most important

More information

Some Principles of Automated Natural Language Information Extraction

Some Principles of Automated Natural Language Information Extraction Some Principles of Automated Natural Language Information Extraction Gregers Koch Department of Computer Science, Copenhagen University DIKU, Universitetsparken 1, DK-2100 Copenhagen, Denmark Abstract

More information

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions. to as a linguistic theory to to a member of the family of linguistic frameworks that are called generative grammars a grammar which is formalized to a high degree and thus makes exact predictions about

More information

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers Chad Langley, Alon Lavie, Lori Levin, Dorcas Wallace, Donna Gates, and Kay Peterson Language Technologies Institute Carnegie

More information

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT PRACTICAL APPLICATIONS OF RANDOM SAMPLING IN ediscovery By Matthew Verga, J.D. INTRODUCTION Anyone who spends ample time working

More information

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS ELIZABETH ANNE SOMERS Spring 2011 A thesis submitted in partial

More information

The Role of the Head in the Interpretation of English Deverbal Compounds

The Role of the Head in the Interpretation of English Deverbal Compounds The Role of the Head in the Interpretation of English Deverbal Compounds Gianina Iordăchioaia i, Lonneke van der Plas ii, Glorianna Jagfeld i (Universität Stuttgart i, University of Malta ii ) Wen wurmt

More information

A Bootstrapping Model of Frequency and Context Effects in Word Learning

A Bootstrapping Model of Frequency and Context Effects in Word Learning Cognitive Science 41 (2017) 590 622 Copyright 2016 Cognitive Science Society, Inc. All rights reserved. ISSN: 0364-0213 print / 1551-6709 online DOI: 10.1111/cogs.12353 A Bootstrapping Model of Frequency

More information

Matching Similarity for Keyword-Based Clustering

Matching Similarity for Keyword-Based Clustering Matching Similarity for Keyword-Based Clustering Mohammad Rezaei and Pasi Fränti University of Eastern Finland {rezaei,franti}@cs.uef.fi Abstract. Semantic clustering of objects such as documents, web

More information

A Comparison of Two Text Representations for Sentiment Analysis

A Comparison of Two Text Representations for Sentiment Analysis 010 International Conference on Computer Application and System Modeling (ICCASM 010) A Comparison of Two Text Representations for Sentiment Analysis Jianxiong Wang School of Computer Science & Educational

More information

THE VERB ARGUMENT BROWSER

THE VERB ARGUMENT BROWSER THE VERB ARGUMENT BROWSER Bálint Sass sass.balint@itk.ppke.hu Péter Pázmány Catholic University, Budapest, Hungary 11 th International Conference on Text, Speech and Dialog 8-12 September 2008, Brno PREVIEW

More information

arxiv:cmp-lg/ v1 22 Aug 1994

arxiv:cmp-lg/ v1 22 Aug 1994 arxiv:cmp-lg/94080v 22 Aug 994 DISTRIBUTIONAL CLUSTERING OF ENGLISH WORDS Fernando Pereira AT&T Bell Laboratories 600 Mountain Ave. Murray Hill, NJ 07974 pereira@research.att.com Abstract We describe and

More information

Stefan Engelberg (IDS Mannheim), Workshop Corpora in Lexical Research, Bucharest, Nov [Folie 1] 6.1 Type-token ratio

Stefan Engelberg (IDS Mannheim), Workshop Corpora in Lexical Research, Bucharest, Nov [Folie 1] 6.1 Type-token ratio Content 1. Empirical linguistics 2. Text corpora and corpus linguistics 3. Concordances 4. Application I: The German progressive 5. Part-of-speech tagging 6. Fequency analysis 7. Application II: Compounds

More information

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly Inflected Languages Classical Approaches to Tagging The slides are posted on the web. The url is http://chss.montclair.edu/~feldmana/esslli10/.

More information