# blame v admire v fault n dispute n talk v debate v blame n appreciate v dispute n converse v admiration n disapprove v Questioning

Save this PDF as:

Size: px
Start display at page:

Download "blame v admire v fault n dispute n talk v debate v blame n appreciate v dispute n converse v admiration n disapprove v Questioning"

## Transcription

1 Automatic Labeling of Semantic Roles Daniel Gildea University of California, Berkeley, and International Computer Science Institute Daniel Jurafsky Department of Linguistics University of Colorado, Boulder Abstract We present a system for identifying the semantic relationships, or semantic roles, lled by constituents of a sentence within a semantic frame. Various lexical and syntactic features are derived from parse trees and used to derive statistical classiers from hand-annotated training data. 1 Introduction Identifying the semantic roles lled by constituents of a sentence can provide a level of shallow semantic analysis useful in solving a number of natural language processing tasks. Semantic roles represent the participants in an action or relationship captured by a semantic frame. For example, the frame for one sense of the verb \crash" includes the roles Agent, Vehicle and To-Location. This shallow semantic level of interpretation can be used for many purposes. Current information extraction systems often use domain-specic frame-and-slot templates to extract facts about, for example, nancial news or interesting political events. A shallow semantic level of representation is a more domain-independent, robust level of representation. Identifying these roles, for example, could allow a system to determine that in the sentence \The rst one crashed" the subject is the vehicle, but in the sentence \The rst one crashed it" the subject is the agent, which would help in information extraction in this domain. Another application is in wordsense disambiguation, where the roles associated with a word can be cues to its sense. For example, Lapata and Brew (1999) and others have shown that the dierent syntactic subcatgorization frames of a verb like \serve" can be used to help disambiguate a particular instance of the word \serve". Adding semantic role subcategorization information to this syntactic information could extend this idea to use richer semantic knowledge. Semantic roles could also act as an important intermediate representation in statistical machine translation or automatic text summarization and in the emerging eld of Text Data Mining (TDM) (Hearst, 1999). Finally, incorporating semantic roles into probabilistic models of language should yield more accurate parsers and better language models for speech recognition. This paper proposes an algorithm for automatic semantic analysis, assigning a semantic role to constituents in a sentence. Our approach to semantic analysis is to treat the problem of semantic role labeling like the similar problems of parsing, part of speech tagging, and word sense disambiguation. We apply statistical techniques that have been successful for these tasks, including probabilistic parsing and statistical classication. Our statistical algorithms are trained on a hand-labeled dataset: the FrameNet database (Baker et al., 1998). The FrameNet database denes a tagset of semantic roles called frame elements, and includes roughly 50,000 sentences from the British National Corpus which have been hand-labeled with these frame elements. The next section describes the set of frame elements/semantic roles used by our system. In the rest of this

2 paper we report on our current system, as well as a number of preliminary experiments on extensions to the system. 2 Semantic Roles Historically, two types of semantic roles have been studied: abstract roles such as Agent and Patient, and roles specic to individual verbs such as Eater and Eaten for \eat". The FrameNet project proposes roles at an intermediate level, that of the semantic frame. Frames are dened as schematic representations of situations involving various participants, props, and other conceptual roles (Fillmore, 1976). For example, the frame \conversation", shown in Figure 1, is invoked by the semantically related verbs \argue", \banter", \debate", \converse", and \gossip" as well as the nouns \argument", \dispute", \discussion" and \ti". The roles dened for this frame, and shared by all its lexical entries, include Protagonist1 and Protagonist2 or simply Protagonists for the participants in the conversation, as well as Medium, and Topic. Example sentences are shown in Table 1. Dening semantic roles at the frame level avoids some of the diculties of attempting to nd a small set of universal, abstract thematic roles, or case roles such as Agent, Patient, etc (as in, among many others, (Fillmore, 1968) (Jackendo, 1972)). Abstract thematic roles can be thought of as being frame elements dened in abstract frames such as\action" and \motion" which are at the top of in inheritance hierarchy of semantic frames (Fillmore and Baker, 2000). The preliminary version of the FrameNet corpus used for our experiments contained 67 frames from 12 general semantic domains chosen for annotation. Examples of domains (see Figure 1) include \motion", \cognition" and \communication". Within these frames, examples of a total of 1462 distinct lexical predicates, or target words, were annotated: 927 verbs, 339 nouns, and 175 adjectives. There are a total of 49,013 annotated sentences, and 99,232 annotated frame elements (which do not include the target words themselves). 3 Related Work Assignment of semantic roles is an important part of language understanding, and has been attacked by many computational systems. Traditional parsing and understanding systems, including implementations of unication-based grammars such as HPSG (Pollard and Sag, 1994), rely on handdeveloped grammars which must anticipate each way in which semantic roles may be realized syntactically. Writing such grammars is time-consuming, and typically such systems have limited coverage. Data-driven techniques have recently been applied to template-based semantic interpretation in limited domains by \shallow" systems that avoid complex feature structures, and often perform only shallow syntactic analysis. For example, in the context of the Air Traveler Information System (ATIS) for spoken dialogue, Miller et al. (1996) computed the probability that a constituent such as \Atlanta" lled a semantic slot such as Destination in a semantic frame for air travel. In a data-driven approach to information extraction, Rilo (1993) builds a dictionary of patterns for lling slots in a specic domain such as terrorist attacks, and Rilo and Schmelzenbach (1998) extend this technique to automatically derive entire case frames for words in the domain. These last systems make use of a limited amount of hand labor to accept or reject automatically generated hypotheses. They show promise for a more sophisticated approach to generalize beyond the relatively small number of frames considered in the tasks. More recently, a domain independent system has been trained on general function tags such as Manner and Temporal by Blaheta and Charniak (2000). 4 Methodology We divide the task of labeling frame elements into two subtasks: that of identifying the boundaries of the frame elements in the sentences, and that of labeling each frame element, given its boundaries, with the correct role. We rst give results for a system which

4 S VP SBAR S VP PP PRP VBD IN N VBD PRP IN NN He heard the sound of liquid slurping in a metal container as Farrell approached him from behind Theme Target Goal Source Figure 2: A sample sentence with parser output (above) and FrameNet annotation (below). Parse constituents corresponding to frame elements are highlighted. for example as a subject or object of a verb. As with phrase type, this feature was read from parse trees returned by the parser. After experimentation with various versions of this feature, we restricted it to apply only to s, as it was found to have little eect on other phrase types. Each 's nearest S or VP ancestor was found in the parse tree; s with an S ancestor were given the grammatical function subject and those with a VP ancestor were labeled object. In general, agenthood is closely correlated with subjecthood. For example, in the sentence \He drove the car over the cli", the rst is more likely to ll the Agent role than the second or third. Position: This feature simply indicates whether the constituent to be labeled occurs before or after the predicate dening the semantic frame. We expected this feature to be highly correlated with grammatical function, since subjects will generally appear before a verb, and objects after. Moreover, this feature may overcome the shortcomings of reading grammatical function from a constituent's ancestors in the parse tree, as well as errors in the parser output. Voice: The distinction between active and passive verbs plays an important role in the connection between semantic role and grammatical function, since direct objects of active verbs correspond to subjects of passive verbs. From the parser output, verbs were classied as active or passive by building a set of 10 passiveidentifying patterns. Each of the patterns requires both a passive auxiliary (some form of \to be" or \to get") and a past participle. Head Word: As previously noted, we expected lexical dependencies to be extremely important in labeling semantic roles, as indicated by their importance in related tasks such as parsing. Since the parser used assigns each constituent

6 Distribution Coverage Accuracy Performance P (rjt) 100% 40.9% 40.9% P (rjpt; t) P (rjpt; gf; t) P (rjpt; position; voice) P (rjpt; position; voice; t) P (rjh) P (rjh; t) P (rjh; pt; t) Table 2: Distributions Calculated for Semantic Role Identication: r indicates semantic role, pt phrase type, gf grammatical function, h head word, and t target word, or predicate. P (rjpt; gf; t) Count in training data P (r =Agtjpt =;gf =Subj;t =abduct) = :46 6 P (r =Thmjpt =;gf =Subj;t =abduct) = :54 7 P (r =Thmjpt =;gf =Obj;t =abduct) = 1 9 P (r =Agtjpt =PP;t =abduct) = :33 1 P (r =Thmjpt =PP;t =abduct) = :33 1 P (r =CoThmjpt =PP;t =abduct) = :33 1 P (r =Manrjpt =ADVP;t =abduct) = 1 1 Table 3: Sample probabilities for P (rjpt; gf; t) calculated from training data for the verb abduct. The variable gf is only dened for noun phrases. The roles dened for the removing frame in the motion domain are: Agent, Theme, CoTheme (\... had been abducted with him") and Manner. Other schemes for choosing values of, including giving more weight to distributions for which more training data was available, were found to have relatively little eect. We attribute this to the fact that the evaluation depends only the the ranking of the probabilities rather than their exact values. P(r h) P(r h, pt, t) P(r h, t) P(r pt, gf, t) P(r t) P(r pt, t) P(r pt, position, voice, t) P(r pt, position, voice) Figure 3: Lattice organization of the distributions from Table 2, with more specic distributions towards the top. In the \backo" combination method, a lattice was constructed over the distributions in Table 2 from more specic conditioning events to less specic, as shown in Figure 3. The less specic distributions were used only when no data was present for any more specic distribution. As before, probabilities were combined with both linear interpolation and a geometric mean. Combining Method Correct Linear Interpolation 79.5% Geometric Mean 79.6 Backo, linear interpolation 80.4 Backo, geometric mean 79.6 Baseline: Most common role 40.9 Table 4: Results on Development Set, 8148 observations The nal system performed at 80.4% accuracy, which can be compared to the 40.9% achieved by always choosing the most probable role for each target word, essentially chance performance on this task. Results for this system on test data, held out during development of the system, are shown in Table

8 and the frame as inputs, whereas a full language understanding system would also identify which frames come into play in a sentence essentially the task of word sense disambiguation. The main feature used was the path from the target word through the parse tree to the constituent in question, represented as a string of parse tree nonterminals linked by symbols indicating upward or downward movement through the tree, as shown in Figure 4. S VP The other features used were the identity of the target word and the identity of the constituent's head word. The probability distributions calculated from the training data were P (fejpath), P (fejpath; t), and P (fejh; t), where fe indicates an event where the parse constituent in question is a frame element, path the path through the parse tree from the target word to the parse constituent, t the identity of the target word, and h the head word of the parse constituent. By varying the probability threshold at which a decision is made, one can plot a precision/recall curve as shown in Figure 5. P (fejpath; t) performs relatively poorly due to fragmentation of the training data (recall only about 30 sentences are available for each target word). While the lexical statistic P (fejh; t) alone is not useful as a classier, using it in linear interpolation with the path statistics improves results. Note that this method can only identify frame elements that have a corresponding constituent in the automatically generated parse tree. For this reason, it is interesting to calculate how many true frame elements overlap with the results of the system, relaxing the criterion that the boundaries must match exactly. Results for partial matching are shown in Table 6. When the automatically identied constituents were fed through the role labeling system described above, 79.6% of the constituents which had been correctly identied in the rst stage were assigned the correct role in the second, roughly equivalent to the performance when assigning roles to constituents identied by hand. Pro He frame element V ate target word Det some N pancakes P(fe path) P(fe path, t).75*p(fe path)+.25*p(fe h, t) Figure 4: In this example, the path from the frame element \He" to the target word \ate" can be represented as " S # VP # V, with " indicating upward movement in the parse tree and # downward movement. recall precision Figure 5: Precison/Recall plot for various methods of identifying frame elements. Recall is calculated over only frame elements with matching parse constituents. 6 Conclusion Our preliminary system is able to automatically label semantic roles with fairly high accuracy, indicating promise for applications in various natural language tasks. Lexical statistics computed on constituent headwords were found to be the most important of the features used. While lexical statistics are quite accurate on the data covered by observations in the training set, the sparsity of the data when conditioned on lexical items meant that combining features was the key to high overall performance. While the combined system was far more accurate than any feature

9 Type of Overlap Identied Constituents Number Exactly Matching Boundaries 66% 5421 Identied constituent entirely within true frame element True frame element entirely within identied constituent Partial overlap 0 26 No match to true frame element Table 6: Results on Identifying Frame Elements (FEs), including partial matches. Results obtained using P (fejpath) with threshold at.5. A total of 7681 constituents were identied as FEs, 8167 FEs were present in hand annotations, of which matching parse constituents were present for 7053 (86%). taken alone, the specic method of combination used was less important. We plan to continue this work by integrating semantic role identication with parsing, by bootstrapping the system on larger, and more representative, amounts of data, and by attempting to generalize from the set of predicates chosen by FrameNet for annotation to general text. References Collin F. Baker, Charles J. Fillmore, and John B. Lowe The berkeley framenet project. In Proceedings of the COLING-ACL, Montreal, Canada. Dan Blaheta and Eugene Charniak Assigning function tags to parsed text. In Proceedings of the 1st Annual Meeting of the North American Chapter of the ACL (NAACL), Seattle, Washington. Glenn Carroll and Mats Rooth Valence induction with a head-lexicalized pcfg. In Proceedings of the 3rd Conference on Empirical Methods in Natural Language Processing (EMNLP 3), Granada, Spain. Michael Collins Three generative, lexicalised models for statistical parsing. In Proceedings of the 35th Annual Meeting of the ACL. Charles J. Fillmore and Collin F. Baker Framenet: Frame semantics meets the corpus. In Linguistic Society of America, January. Charles Fillmore The case for case. In Bach and Harms, editors, Universals in Linguistic Theory, pages 1{88. Holt, Rinehart, and Winston, New York. Charles J. Fillmore Frame semantics and the nature of language. In Annals of the New York Academy of Sciences: Conference on the Origin and Development of Language and Speech, volume 280, pages 20{32. Marti Hearst Untangling text data mining. In Proceedings of the 37rd Annual Meeting of the ACL. Thomas Hofmann and Jan Puzicha Statistical models for co-occurrence data. Memo, Massachussetts Institute of Technology Articial Intelligence Laboratory, February. Ray Jackendo Semantic Interpretation in Generative Grammar. MIT Press, Cambridge, Massachusetts. Maria Lapata and Chris Brew Using subcategorization to resolve verb class ambiguity. In Joint SIGDAT Conference on Empirical Methods in NLP and Very Large Corpora, Maryland. Dekang Lin Automatic retrieval and clustering of similar words. In Proceedings of the COLING-ACL, Montreal, Canada. Scott Miller, David Stallard, Robert Bobrow, and Richard Schwartz A fully statistical approach to natural language interfaces. In Proceedings of the 34th Annual Meeting of the ACL. Carl Pollard and Ivan A. Sag Head- Driven Phrase Structure Grammar. University of Chicago Press, Chicago. Ellen Rilo and Mark Schmelzenbach An empirical approach to conceptual case frame acquisition. In Proceedings of the Sixth Workshop on Very Large Corpora. Ellen Rilo Automatically constructing a dictionary for information extraction tasks. In Proceedings of the Eleventh National Conference on Articial Intelligence (AAAI).

### Explorations in Disambiguation Using XML Text Representation. Kenneth C. Litkowski CL Research 9208 Gue Road Damascus, MD

Explorations in Disambiguation Using XML Text Representation Kenneth C. Litkowski CL Research 9208 Gue Road Damascus, MD 20872 ken@clres.com Abstract In SENSEVAL-3, CL Research participated in four tasks:

### Learning Parse Decisions From Examples With Rich Context. as submitted to ACL'96 on January 8, Ulf Hermjakob and Raymond J.

Learning Parse Decisions From Examples With Rich Context as submitted to ACL'96 on January 8, 1996 Ulf Hermjakob and Raymond J. Mooney Dept. of Computer Sciences University of Texas at Austin Austin, TX

### Tree Kernel Engineering for Proposition Re-ranking

Tree Kernel Engineering for Proposition Re-ranking Alessandro Moschitti, Daniele Pighin, and Roberto Basili Department of Computer Science University of Rome Tor Vergata, Italy {moschitti,basili}@info.uniroma2.it

### Alternative Syntactic Theories

Alternative Syntactic Theories L614 Spring 2015 Syntactic analysis Generative grammar: collection of words and rules with which we generate strings of those words, i.e., sentences Syntax attempts to capture

### CS474 Natural Language Processing. Word sense disambiguation. Machine learning approaches. Dictionary-based approaches

CS474 Natural Language Processing! Today Lexical semantic resources: WordNet» Dictionary-based approaches» Supervised machine learning methods» Issues for WSD evaluation Word sense disambiguation! Given

### Introduction to Advanced Natural Language Processing (NLP)

Advanced Natural Language Processing () L645 / B659 Dept. of Linguistics, Indiana University Fall 2015 1 / 24 Definition of CL 1 Computational linguistics is the study of computer systems for understanding

### Evaluating the Effectiveness of Ensembles of Decision Trees in Disambiguating Senseval Lexical Samples

Evaluating the Effectiveness of Ensembles of Decision Trees in Disambiguating Senseval Lexical Samples Ted Pedersen Department of Computer Science University of Minnesota Duluth, MN, 55812 USA tpederse@d.umn.edu

### 6.891: Lecture 4 (September 20, 2005) Parsing and Syntax II

6.891: Lecture 4 (September 20, 2005) Parsing and Syntax II Overview Weaknesses of PCFGs Heads in context-free rules Dependency representations of parse trees Two models making use of dependencies Weaknesses

### Using WordNet to Extend FrameNet Coverage

Using WordNet to Extend FrameNet Coverage Johansson, Richard; Nugues, Pierre Published in: LU-CS-TR: 2007-240 Published: 2007-01-01 Link to publication Citation for published version (APA): Johansson,

### COMS W4705x: Natural Language Processing FINAL EXAM December 18th, 2008

COMS W4705x: Natural Language Processing FINAL EXAM December 18th, 2008 DIRECTIONS This exam is closed book and closed notes. It consists of four parts. Each part is labeled with the amount of time you

### CS474 Introduction to Natural Language Processing Final Exam December 15, 2005

Name: CS474 Introduction to Natural Language Processing Final Exam December 15, 2005 Netid: Instructions: You have 2 hours and 30 minutes to complete this exam. The exam is a closed-book exam. # description

### Speech and Language Processing. Today

Speech and Language Processing Formal Grammars Chapter 12 Formal Grammars Today Context-free grammar Grammars for English Treebanks Dependency grammars 9/26/2013 Speech and Language Processing - Jurafsky

### Chapter 1. Introduction

Chapter 1 Introduction This thesis is concerned with experiments on the automatic induction of German semantic verb classes. In other words, (a) the focus of the thesis is verbs, (b) I am interested in

### Representing Caused Motion in Embodied Construction Grammar

Representing Caused Motion in Embodied Construction Grammar Ellen K. Dodge International Computer Science Institute Berkeley, CA, USA 94704 edodge@icsi.berkeley.edu Miriam R. L. Petruck International Computer

### CS497:Learning and NLP Lec 3: Natural Language and Statistics

CS497:Learning and NLP Lec 3: Natural Language and Statistics Spring 2009 January 28, 2009 Lecture Corpora and its analysis Motivation for statistical approaches Statistical properties of language (e.g.,

### Statistical NLP: linguistic essentials. Updated 10/15

Statistical NLP: linguistic essentials Updated 10/15 Parts of Speech and Morphology syntactic or grammatical categories or parts of Speech (POS) are classes of word with similar syntactic behavior Examples

### A TAG-based noisy channel model of speech repairs

A TAG-based noisy channel model of speech repairs Mark Johnson and Eugene Charniak Brown University ACL, 2004 Supported by NSF grants LIS 9720368 and IIS0095940 1 Talk outline Goal: Apply parsing technology

### Assignment 4. CMSC 473/673 Introduction to Natural Language Processing. Due Monday December 11, 2017, 11:59 AM

Assignment 4 CMSC 473/673 Introduction to Natural Language Processing Due Monday December 11, 2017, 11:59 AM Item Summary Assigned Tuesday November 21st, 2017 Due Monday December 11th, 2017 Topic Syntax

### Mention Detection: Heuristics for the OntoNotes annotations

Mention Detection: Heuristics for the OntoNotes annotations Jonathan K. Kummerfeld, Mohit Bansal, David Burkett and Dan Klein Computer Science Division University of California at Berkeley {jkk,mbansal,dburkett,klein}@cs.berkeley.edu

### Wrap-Up: a Trainable Discourse. Module for Information Extraction. Abstract

Journal of Articial Intelligence Research 2 (1994) 131-158 Submitted 4/94; published 12/94 Wrap-Up: a Trainable Discourse Module for Information Extraction Stephen Soderland Wendy Lehnert Department of

### Context Free Grammars

Context Free Grammars Synchronic Model of Language Syntactic Lexical Morphological Semantic Pragmatic Discourse Syntactic Analysis Syntax expresses the way in which words are arranged together. The kind

### Part-of-Speech Tagging & Sequence Labeling. Hongning Wang

Part-of-Speech Tagging & Sequence Labeling Hongning Wang CS@UVa What is POS tagging Tag Set NNP: proper noun CD: numeral JJ: adjective POS Tagger Raw Text Pierre Vinken, 61 years old, will join the board

### Natural Language Processing: An approach to Parsing and Semantic Analysis

Natural Language Processing: An approach to Parsing and Semantic Analysis Shabina Dhuria Department of Computer Science, DAV College, Sector-10, Chandigarh Abstract: Natural language processing is the

### Chunking. Ewan Klein ICL 14 November 2005

in NLTK-Lite in Cass as Tagging Ewan Klein ewan@inf.ed.ac.uk ICL 14 November 2005 in NLTK-Lite in Cass as Tagging in NLTK-Lite in Cass as Tagging in NLTK-Lite in Cass as Tagging Problems with Full Parsing,

### Topics in Computational Linguistics Grammar Engineering

Topics in Computational Linguistics Grammar Engineering Dan Flickinger CSLI Stanford & Saarland University danf@csli..edu Stephan Oepen Universitetet i Oslo & CSLI Stanford oe@csli..edu http://lingo..edu/courses/05/ge/

### Multi-Class Sentiment Analysis with Clustering and Score Representation

Multi-Class Sentiment Analysis with Clustering and Score Representation Mohsen Farhadloo Erik Rolland mfarhadloo@ucmerced.edu 1 CONTENT Introduction Applications Related works Our approach Experimental

### Chapter 2 EXISTING TEXT MINING SYSTEMS Information Retrieval (IR)

Chapter 2 EXISTING TEXT MINING SYSTEMS In this chapter, I give an overview of the methods used in text mining and information extraction in the biomedical field nowadays and also what the problems with

### Natural Language Processing SoSe Parsing. (based on the slides of Dr. Saeedeh Momtazi)

Natural Language Processing SoSe 2015 Parsing Dr. Mariana Neves May 18th, 2014 (based on the slides of Dr. Saeedeh Momtazi) Parsing Finding structural relationships between words in a sentence (http://nlp.stanford.edu:8080/parser)

### Statistical Approaches to Natural Language Processing CS 4390/5319 Spring Semester, 2003 Syllabus

Statistical Approaches to Natural Language Processing CS 4390/5319 Spring Semester, 2003 Syllabus http://www.cs.utep.edu/nigel/nlp.html Time and Location 15:00 16:25, Tuesdays and Thursdays Computer Science

### Dependency Grammar. Lilja Øvrelid INF5830 Fall With thanks to Markus Dickinson, Sandra Kübler and Joakim Nivre. Dependency Grammar 1(37)

Dependency Grammar Lilja Øvrelid INF5830 Fall 2015 With thanks to Markus Dickinson, Sandra Kübler and Joakim Nivre Dependency Grammar 1(37) Course overview Overview INF5830 so far general methodology statistical,

### Experiments on the LIMSI Broadcast News Data

Experiments on the LIMSI Broadcast News Data Interim Report for SNF Project 105211-112133: Rule-Based Language Model for Speech Recognition Tobias Kaufmann Institut für Technische Informatik und Kommunikationsnetze

### Part II. Statistical NLP

Advanced Artificial Intelligence Part II. Statistical NLP Applications of HMMs and PCFGs in NLP Wolfram Burgard, Luc De Raedt, Bernhard Nebel, Lars Schmidt-Thieme Most slides taken (or adapted) from Adam

### Lexical Acquisition in Statistical NLP

Lexical Acquisition in Statistical NLP Adapted from: Manning and Schütze, 1999 Chapter 8 (pp. 265-278; 308-312) Anjana Vakil University of Saarland Outline What is lexical information? Why is it important

### Scanning Phase. Tokenization. Partial Parsing. Rule Application. Target Information

The Use of Word Sense Disambiguation in an Information Extraction System Joyce Yue Chai 30 Saw Mill River Rd. IBM T. J. Watson Research Center Hawthorne, NY 10532 jchai@us.ibm.com Alan W. Biermann Computer

### Syntactic Reordering of Source Sentences for Statistical Machine Translation

Syntactic Reordering of Source Sentences for Statistical Machine Translation Mohammad Sadegh Rasooli Columbia University rasooli@cs.columbia.edu April 9, 2013 M. S. Rasooli (Columbia University) Syntactic

### Outline. Statistical Natural Language Processing. Symbolic NLP Insufficient. Statistical NLP. Statistical Language Models

Outline Statistical Natural Language Processing July 8, 26 CS 486/686 University of Waterloo Introduction to Statistical NLP Statistical Language Models Information Retrieval Evaluation Metrics Other Applications

### UIO-Lien: Entailment Recognition using Minimal Recursion Semantics

UIO-Lien: Entailment Recognition using Minimal Recursion Semantics Elisabeth Lien Department of Informatics University of Oslo, Norway elien@ifi.uio.no Milen Kouylekov Department of Informatics University

### PRELIMINARY EVALUATION OF THE VOYAGER SPOKEN LANGUAGE SYSTEM*

PRELIMINARY EVALUATION OF THE VOYAGER SPOKEN LANGUAGE SYSTEM* Victor Zue, James Glass, David Goodine, Hong Leung, Michael Phillips, Joseph Polifroni, and Stephanie Seneff Spoken Language Systems Group

### A Translation Aid System Using Flexible Text Retrieval Based on Syntax-Matching

A Translation Aid System Using Flexible Text Retrieval Based on Syntax-Matching Eiichiro SUMITA and Yutaka TSUTSUMI Tokyo Research Laboratory, IBM Japan, LTD. Abstract : ETOC (Easy TO Consult) is a translation

### Probability and Statistics in NLP. Niranjan Balasubramanian Jan 28 th, 2016

Probability and Statistics in NLP Niranjan Balasubramanian Jan 28 th, 2016 Natural Language Mechanism for communicating thoughts, ideas, emotions, and more. What is NLP? Building natural language interfaces

### University of Toronto, Department of Computer Science CSC 485/2501F Computational Linguistics, Fall Assignment 1

University of Toronto, Department of Computer Science CSC 485/2501F Computational Linguistics, Fall 2017 Assignment 1 Due date: 14:10, Friday 6 October 2017, in tutorial. Late assignments will not be accepted

### Part-of-Speech Tagging. Yan Shao Department of Linguistics and Philology, Uppsala University 19 April 2017

Part-of-Speech Tagging Yan Shao Department of Linguistics and Philology, Uppsala University 19 April 2017 Last time N-grams are used to create language models The probabilities are obtained via on corpora

### The Aims of Linguistic Theory

The Aims of Linguistic Theory Petr Horáček, Eva Zámečníková and Ivana Burgetová Department of Information Systems Faculty of Information Technology Brno University of Technology Božetěchova 2, 612 00 Brno,

### Statistically-Enhanced New Word Identification in a Rule-Based Chinese System

Statistically-Enhanced New Word Identification in a Rule-Based Chinese System Andi Wu Microsoft Research One Microsoft Way Redmond, WA 98052 Andiwu @microsoft.com Zixin Jiang Microsoft Research One Microsoft

### Building Applied Natural Language Generation Systems. Robert Dale and Ehud Reiter

Building Applied Natural Language Generation Systems Robert Dale and Ehud Reiter 1 Overview 1 An Introduction to NLG 2 Requirements Analysis for NLG 3 NLG Architecture and System Design 4 A Case Study

### Disambiguating Levin Verbs Using Untagged Data

Disambiguating Levin Verbs Using Untagged Data Jianguo Li Department of Linguistics The Ohio State University Columbus, Ohio, USA jianguo@ling.ohio-state.edu Chris Brew Department of Linguistics The Ohio

### RESOLVING PART-OF-SPEECH AMBIGUITY IN THE GREEK LANGUAGE USING LEARNING TECHNIQUES

RESOLVING PART-OF-SPEECH AMBIGUITY IN THE GREEK LANGUAGE USING LEARNING TECHNIQUES Georgios Petasis, Georgios Paliouras, Vangelis Karkaletsis, Constantine D. Spyropoulos and Ion Androutsopoulos Software

### NLANGP: Supervised Machine Learning System for Aspect Category Classification and Opinion Target Extraction

NLANGP: Supervised Machine Learning System for Aspect Category Classification and Opinion Target Extraction Zhiqiang Toh Institute for Infocomm Research 1 Fusionopolis Way Singapore 138632 ztoh@i2r.a-star.edu.sg

### Sentence Reduction for Automatic Text Summarization

Sentence Reduction for Automatic Text Summarization Hongyan Jing Department of Computer Science Columbia University New York, NY 10027, USA hj ing@cs.columbia.edu Abstract We present a novel sentence reduction

### Non-parametric Bayesian models for computational morphology

Non-parametric Bayesian models for computational morphology Dissertation defence Kairit Sirts Institute of Informatics Tallinn University of Technology 18.06.2015 1 Outline 1. NLP and computational morphology

### A Case Study: News Classification Based on Term Frequency

A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center

### Towards Automatic Verb Acquisition from VerbNet for Spoken Dialog Processing

Towards Automatic Verb Acquisition from VerbNet for Spoken Dialog Processing Mary Swift Department of Computer Science University of Rochester Rochester, NY 14607 USA swift@cs.rochester.edu Abstract This

### Chinese Syntactic Parsing Based on Extended GLR Parsing Algorithm with PCFG*

Chinese Syntactic Parsing Based on Extended GLR Parsing Algorithm with PCFG* Yan Zhang, Bo Xu and Chengqing Zong National Laboratory of Pattern Recognition, Institute of Automation Chinese Academy of sciences,

### Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar Induction

Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar Induction Hiroshi Noji Yusuke Miyao Mark Johnson Nara Institute of Science and Technology National Institute of Informatics

### Corpus-based terminology extraction applied to information access

Corpus-based terminology extraction applied to information access Anselmo Peñas, Felisa Verdejo and Julio Gonzalo {anselmo,felisa,julio}@lsi.uned.es Dpto. Lenguajes y Sistemas Informáticos, UNED, Spain

### Text Summarization Using Lexical Chains

Text Summarization Using Lexical Chains Meru Brunn Yllias Chali Christopher J. Pinchak Department of Mathematics and Computer Science University of Lethbridge 4401 University Drive Lethbridge, Alberta,

### Two hierarchical text categorization approaches for BioASQ semantic indexing challenge. BioASQ challenge 2013 Valencia, September 2013

Two hierarchical text categorization approaches for BioASQ semantic indexing challenge Francisco J. Ribadas Víctor M. Darriba Compilers and Languages Group Universidade de Vigo (Spain) http://www.grupocole.org/

### NaturalLanguageProcessing-Lecture12

NaturalLanguageProcessing-Lecture12 Instructor (Dan Jurafsky) :How about now? Oh, yeah, okay. Great, all right. So I m Dan Jurafsky. I m taking for Chris today as he emailed you because he had to leave

### The Aims of Linguistic Theory

The Aims of Linguistic Theory Petr Horáček, Eva Zámečníková and Ivana Burgetová Department of Information Systems Faculty of Information Technology Brno University of Technology Božetěchova 2, 612 00 Brno,

### n?1 Y CF P (N1; ::::; N n ) = log(p (N1jS P re ) P (N i js In ) P (N n js P ost )) (1) i=2 C(N1; S P re ) P (N1jS P re ) = Pn i=1 C(N i; S P re ) C(N

Segmenting Korean Compound Nouns using Statistical Information and a Preference Rule Bo-Hyun Yun, Min-Jeung Cho, Hae-Chang Rim Department of Computer Science, Korea University 1, 5-ka, Anam-dong, SEOUL,

### The use of speech recognition confidence scores in dialogue systems

The use of speech recognition confidence scores in dialogue systems GABRIEL SKANTZE gabriel@speech.kth.se Department of Speech, Music and Hearing, KTH This paper discusses the interpretation of speech

### Extracting and Using Trace-Free Functional Dependencies from the Penn Treebank to Reduce Parsing Complexity

Extracting and Using Trace-Free Functional Dependencies from the Penn Treebank to Reduce Parsing Complexity Gerold Schneider Institute of Computational Linguistics, University of Zurich Department of Linguistics,

### Dependency Parsing. Prashanth Mannem

Dependency Parsing Prashanth Mannem mannemp@eecs.oregonstate.edu Outline Introduction Dependency Parsing Formal definition Parsing Algorithms Introduction Dynamic programming Deterministic search 2 Syntax

### IITB System for CoNLL 2013 Shared Task: A Hybrid Approach to Grammatical Error Correction

IITB System for CoNLL 2013 Shared Task: A Hybrid Approach to Grammatical Error Correction Anoop Kunchukuttan Ritesh Shah Pushpak Bhattacharyya Department of Computer Science and Engineering, IIT Bombay

Artificial Intelligence Programming Statistical NLP Chris Brooks Outline n-grams Applications of n-grams review - Context-free grammars Probabilistic CFGs Information Extraction Advantages of IR approaches

### Word Sense Disambiguation with Semi-Supervised Learning

Word Sense Disambiguation with Semi-Supervised Learning Thanh Phong Pham 1 and Hwee Tou Ng 1,2 and Wee Sun Lee 1,2 1 Department of Computer Science 2 Singapore-MIT Alliance National University of Singapore

### Dialogue manager. Natural. language parser). Black Board. Domain model. Topsy

The Intellimedia WorkBench - an environment for building multimodal systems Tom Brndsted ftb@cpk.auc.dkg, Lars Bo Larsen flbl@cpk.auc.dkg, Michael Manthey fmanthey@cs.auc.dkg, Paul Mc Kevitt fpmck@cpk.auc.dkg,

### Naive Bayes Classifier Approach to Word Sense Disambiguation

Naive Bayes Classifier Approach to Word Sense Disambiguation Daniel Jurafsky and James H. Martin Chapter 20 Computational Lexical Semantics Sections 1 to 2 Seminar in Methodology and Statistics 3/June/2009

### A Combined Memory-Based Semantic Role Labeler of English

A Combined Memory-Based Semantic Role Labeler of English Roser Morante, Walter Daelemans, Vincent Van Asch CNTS - Language Technology Group University of Antwerp Prinsstraat 13, B-2000 Antwerpen, Belgium

### Natural Language Processing: Part II Overview of Natural Language Processing (L90): ACS

Natural Language Processing: Part II Overview of Natural Language Processing (L90): ACS Ann Copestake Computer Laboratory University of Cambridge October 2017 Outline of today s lecture Overview of the