SEMAFOR: Frame Argument Resolution with Log-Linear Models

Similar documents
The stages of event extraction

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

LTAG-spinal and the Treebank

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Prediction of Maximal Projection for Semantic Role Labeling

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

AQUA: An Ontology-Driven Question Answering System

Linking Task: Identifying authors and book titles in verbose queries

The Strong Minimalist Thesis and Bounded Optimality

Heuristic Sample Selection to Minimize Reference Standard Training Set for a Part-Of-Speech Tagger

Ensemble Technique Utilization for Indonesian Dependency Parser

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Rule-based Expert Systems

Developing a large semantically annotated corpus

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

University of Alberta. Large-Scale Semi-Supervised Learning for Natural Language Processing. Shane Bergsma

Constraining X-Bar: Theta Theory

Exploration. CS : Deep Reinforcement Learning Sergey Levine

MENTORING. Tips, Techniques, and Best Practices

Annotation Projection for Discourse Connectives

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization

Natural Language Processing: Interpretation, Reasoning and Machine Learning

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Context Free Grammars. Many slides from Michael Collins

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

The Discourse Anaphoric Properties of Connectives

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Graph Alignment for Semi-Supervised Semantic Role Labeling

Memory-based grammatical error correction

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation

Distant Supervised Relation Extraction with Wikipedia and Freebase

Which verb classes and why? Research questions: Semantic Basis Hypothesis (SBH) What verb classes? Why the truth of the SBH matters

The Role of Semantic and Discourse Information in Learning the Structure of Surgical Procedures

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

How to analyze visual narratives: A tutorial in Visual Narrative Grammar

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

The MEANING Multilingual Central Repository

Applications of memory-based natural language processing

Rule Learning With Negation: Issues Regarding Effectiveness

Proof Theory for Syntacticians

The Role of the Head in the Interpretation of English Deverbal Compounds

Grammars & Parsing, Part 1:

NCEO Technical Report 27

Leveraging Sentiment to Compute Word Similarity

Natural Language Processing. George Konidaris

Beyond the Pipeline: Discrete Optimization in NLP

Summary / Response. Karl Smith, Accelerations Educational Software. Page 1 of 8

Disambiguation of Thai Personal Name from Online News Articles

Unsupervised Learning of Narrative Schemas and their Participants

The Smart/Empire TIPSTER IR System

Unit 3. Design Activity. Overview. Purpose. Profile

BYLINE [Heng Ji, Computer Science Department, New York University,

Control and Boundedness

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

cmp-lg/ Jan 1998

arxiv: v1 [cs.cl] 2 Apr 2017

Text Type Purpose Structure Language Features Article

Building a Semantic Role Labelling System for Vietnamese

Can Human Verb Associations help identify Salient Features for Semantic Verb Classification?

Lecture 1: Basic Concepts of Machine Learning

Aspectual Classes of Verb Phrases

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Multilingual Sentiment and Subjectivity Analysis

A Case Study: News Classification Based on Term Frequency

Using dialogue context to improve parsing performance in dialogue systems

Author: Justyna Kowalczys Stowarzyszenie Angielski w Medycynie (PL) Feb 2015

Firms and Markets Saturdays Summer I 2014

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

Grammar Extraction from Treebanks for Hindi and Telugu

A Framework for Customizable Generation of Hypertext Presentations

Chapter 2 Rule Learning in a Nutshell

The Choice of Features for Classification of Verbs in Biomedical Texts

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

What is a Mental Model?

Summarizing A Nonfiction

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Intensive Writing Class

Update on Soar-based language processing

Concept Acquisition Without Representation William Dylan Sabo

Reading Grammar Section and Lesson Writing Chapter and Lesson Identify a purpose for reading W1-LO; W2- LO; W3- LO; W4- LO; W5-

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2

The College Board Redesigned SAT Grade 12

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Postprint.

Course Law Enforcement II. Unit I Careers in Law Enforcement

The Conversational User Interface

Practice Examination IREB

Radius STEM Readiness TM

Some Principles of Automated Natural Language Information Extraction

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Transcription:

SEMAFOR: Frame Argument Resolution with Log-Linear Models Desai Chen or, The Case of the Missing Arguments Nathan Schneider SemEval July 16, 2010 Dipanjan Das School of Computer Science Carnegie Mellon University Noah A. Smith SemEval Task 10: Linking Events and their Participants in Discourse We describe an approach to frame-semantic role labeling and evaluate it on data from this task.

SEMAFOR: Frame Argument Resolution with Log-Linear Models Desai Chen or, The Case of the Missing Arguments Nathan Schneider (guy in the front of the room) Dipanjan Das Noah A. Smith SemEval Task 10: Linking Events and their Participants in Discourse We describe an approach to frame-semantic role labeling and evaluate it on data from this task.

New wrinkle in this version of the task: classifying and resolving missing arguments. Frame SRL NNP VBP IN PRP NN IN IN PRP VBD VBN VBN WRB PRP VBD DT NN. Holmes sprang in his chair as if he had been stung when I read the headline. (SemEval 2010 trial data) 2 This is a full annotation of a sentence in terms of its frames/arguments. Note that this is a *partial* semantic representation: it shows a certain amount of relational meaning but doesn t encode, for instance, that as if he had been stung is a hypothetical used to provide imagery for the manner of motion (we infer that it must have been rapid and brought upon by a shocking stimulus). The SRL task: Given a sentence with POS tags, syntactic dependencies, predicates, and frame names, predict the arguments for each frame role.

New wrinkle in this version of the task: classifying and resolving missing arguments. Frame SRL NNP VBP IN PRP NN IN IN PRP VBD VBN VBN WRB PRP VBD DT NN. Holmes sprang in his chair as if he had been stung when I read the headline. EXPERIENCER_OBJ READING SELF_MOTION (SemEval 2010 trial data) 2 This is a full annotation of a sentence in terms of its frames/arguments. Note that this is a *partial* semantic representation: it shows a certain amount of relational meaning but doesn t encode, for instance, that as if he had been stung is a hypothetical used to provide imagery for the manner of motion (we infer that it must have been rapid and brought upon by a shocking stimulus). The SRL task: Given a sentence with POS tags, syntactic dependencies, predicates, and frame names, predict the arguments for each frame role.

New wrinkle in this version of the task: classifying and resolving missing arguments. Frame SRL NNP VBP IN PRP NN IN IN PRP VBD VBN VBN WRB PRP VBD DT NN. Holmes sprang in his chair as if he had been stung when I read the headline. Experiencer EXPERIENCER_OBJ Stimulus: INI READING Reader Text SELF_MOTION Self_mover Place Manner Time (SemEval 2010 trial data) 2 This is a full annotation of a sentence in terms of its frames/arguments. Note that this is a *partial* semantic representation: it shows a certain amount of relational meaning but doesn t encode, for instance, that as if he had been stung is a hypothetical used to provide imagery for the manner of motion (we infer that it must have been rapid and brought upon by a shocking stimulus). The SRL task: Given a sentence with POS tags, syntactic dependencies, predicates, and frame names, predict the arguments for each frame role.

New wrinkle in this version of the task: classifying and resolving missing arguments. Frame SRL NNP VBP IN PRP NN IN IN PRP VBD VBN VBN WRB PRP VBD DT NN. Holmes sprang in his chair as if he had been stung when I read the headline. EXPERIENCER_OBJ READING SELF_MOTION Experiencer Stimulus: INI Reader What the Experiencer felt is missing! Self_mover Place Manner Time Text (SemEval 2010 trial data) 2 This is a full annotation of a sentence in terms of its frames/arguments. Note that this is a *partial* semantic representation: it shows a certain amount of relational meaning but doesn t encode, for instance, that as if he had been stung is a hypothetical used to provide imagery for the manner of motion (we infer that it must have been rapid and brought upon by a shocking stimulus). The SRL task: Given a sentence with POS tags, syntactic dependencies, predicates, and frame names, predict the arguments for each frame role.

Contributions Evaluate frame SRL on new data Experiment with a classifier for null instantiations (NIs) implicit interactions in a discourse 3

Overview Background: frame SRL Overt argument identification Null instantiation resolution Conclusion 4

FrameNet FrameNet (Fillmore et al., 2003) defines semantic frames, roles, and associated predicates provides a linguistically rich representation for predicate-argument structures based on the theory of frame semantics (Fillmore, 1982) 5

FrameNet MAKE_NOISE Sound Place Time Noisy_event Sound_source cough.v, gobble.v, hiss.v, ring.v, yodel.v,... http://framenet.icsi.berkeley.edu 6 The FrameNet lexicon is a repository of expert information, storing the semantic frames and a number of (frame-specific) roles. Each frame represents a holistic event or scenario, generalizing over specific predicates. It also defines roles for the participants, props, and attributes of the scenario.

FrameNet frame name MAKE_NOISE Sound Place roles Time Noisy_event Sound_source cough.v, gobble.v, hiss.v, ring.v, yodel.v,... group of predicates ( lexical units ) http://framenet.icsi.berkeley.edu 7 For example, here we show the Make_noise frame that has several roles such as Sound, Noisy_event, Sound_Source, etc. FrameNet also lists some possible lexical units which could evoke these frames. Examples for this frame are cough, gobble, hiss, ring, and so on.

FrameNet EVENT Event Place Time event.n, happen.v, occur.v, take place.v,... TRANSITIVE_ACTION Event Place CAUSE_TO_MAKE_NOISE Purpose Place MAKE_NOISE Sound Place OBJECTIVE_INFLUENCE Place Time Agent Time Agent Time Noisy_event Time Influencing_entity Influencing_situation Cause Patient Cause Sound_maker blare.v, honk.v, play.v, ring.v, toot.v,... Sound_source cough.v, gobble.v, hiss.v, ring.v, yodel.v,... Dependent_entity affect.v, effect.n, impact.n, impact.v,... Inheritance relation Excludes relation Causative_of relation relationships between frames and between roles http://framenet.icsi.berkeley.edu 8 The FrameNet lexicon also provides relationships between frames and between roles

Annotated Data 9 [SE 07] has ANC travel guides, PropBank news, and (mostly) NTI reports on weapons stockpiles. Unlike other participants, we do not use the 139,000 lexicographic exemplar sentences (except indirectly through features) because the annotations are partial (only 1 frame) and the sample of sentences is biased (they were chosen manually to illustrate variation of arguments). [SE 10] also has coreference, though we do not make use of this information.

Annotated Data Full-text annotations: all frames + arguments [SE 07] SemEval 2007 task data: news, popular nonfiction, bureaucratic 2000 sentences, 50K words 9 [SE 07] has ANC travel guides, PropBank news, and (mostly) NTI reports on weapons stockpiles. Unlike other participants, we do not use the 139,000 lexicographic exemplar sentences (except indirectly through features) because the annotations are partial (only 1 frame) and the sample of sentences is biased (they were chosen manually to illustrate variation of arguments). [SE 10] also has coreference, though we do not make use of this information.

Annotated Data Full-text annotations: all frames + arguments [SE 07] SemEval 2007 task data: news, popular nonfiction, bureaucratic 2000 sentences, 50K words [SE 10] New SemEval 2010 data: fiction 1000 sentences, 17K words ½ train, ½ test 9 [SE 07] has ANC travel guides, PropBank news, and (mostly) NTI reports on weapons stockpiles. Unlike other participants, we do not use the 139,000 lexicographic exemplar sentences (except indirectly through features) because the annotations are partial (only 1 frame) and the sample of sentences is biased (they were chosen manually to illustrate variation of arguments). [SE 10] also has coreference, though we do not make use of this information.

Overview Background: frame SRL Overt argument identification Null instantiation resolution Conclusion 10

Frame SRL: Overt Arguments We train a classifier SELF_MOTION.Place to pick an argument for each role of each sprang (parse) frame. SRL IN PRP NN in his chair (Das et al., 2010) 11 See NAACL 2010 paper

Frame SRL: Overt Arguments We train a classifier SELF_MOTION.Place to pick an argument for each role of each sprang (parse) frame. SRL IN PRP NN in his chair a probabilistic model with features looking at the span, the frame, the role, and the observed sentence structure (Das et al., 2010) 11 See NAACL 2010 paper

Frame SRL: Overt Arguments sprang ~ SELF_MOTION 12 An example of the desired mapping. For the predicate sprang, each role of the evoked frame is considered separately, and filled with a phrase in the sentence or left empty.

Frame SRL: Overt Arguments sprang ~ SELF_MOTION Self_mover Place Path Goal Time Manner... 12 An example of the desired mapping. For the predicate sprang, each role of the evoked frame is considered separately, and filled with a phrase in the sentence or left empty.

Frame SRL: Overt Arguments sprang ~ SELF_MOTION Self_mover Place Path Goal Time Manner... NNP Holmes PRP I IN PRP NN in his chair IN IN PRP VBD VBN PRP NN his chair WRB PRP VBD DT NN when I read the headline DT VBN as if he had been stung PRP VBD VBN VBN he had been stung PRP he NN the headline... 12 An example of the desired mapping. For the predicate sprang, each role of the evoked frame is considered separately, and filled with a phrase in the sentence or left empty.

Frame SRL: Overt Arguments sprang ~ SELF_MOTION Self_mover Place Path Goal Time Manner... NNP Holmes PRP I IN PRP NN in his chair IN IN PRP VBD VBN PRP NN his chair WRB PRP VBD DT NN when I read the headline DT VBN as if he had been stung PRP VBD VBN VBN he had been stung PRP he NN the headline... 12 An example of the desired mapping. For the predicate sprang, each role of the evoked frame is considered separately, and filled with a phrase in the sentence or left empty.

Frame SRL: Overt Arguments sprang ~ SELF_MOTION Self_mover Place Path NNP Holmes IN IN PRP VBD VBN IN PRP NN in his chair PRP NN his chair VBN as if he had been stung Goal Time Manner... PRP he WRB PRP VBD DT when I read the headline PRP I PRP VBD VBN VBN he had been stung... DT NN NN the headline 12 An example of the desired mapping. For the predicate sprang, each role of the evoked frame is considered separately, and filled with a phrase in the sentence or left empty.

Frame SRL: Overt Arguments sprang ~ SELF_MOTION Self_mover Place Path Goal Time Manner... NNP Holmes PRP I IN PRP NN in his chair IN IN PRP VBD VBN PRP NN his chair WRB PRP VBD DT NN when I read the headline DT VBN as if he had been stung PRP VBD VBN VBN he had been stung PRP he NN the headline... 12 An example of the desired mapping. For the predicate sprang, each role of the evoked frame is considered separately, and filled with a phrase in the sentence or left empty.

Frame SRL: Overt Arguments stung ~ EXPERIENCER_OBJ Experiencer Stimulus Degree Time Manner... NNP Holmes PRP I IN PRP NN in his chair IN IN PRP VBD VBN PRP NN his chair WRB PRP VBD DT NN when I read the headline DT VBN as if he had been stung PRP VBD VBN VBN he had been stung PRP he NN the headline... 13...and likewise for stung, etc.

Frame SRL: Overt Arguments stung ~ EXPERIENCER_OBJ Experiencer Stimulus Degree Time NNP Holmes IN PRP NN in his chair IN IN PRP VBD VBN PRP NN his chair VBN as if he had been stung PRP VBD VBN VBN he had been stung PRP he Manner... WRB PRP VBD DT PRP I NN when I read the headline DT NN the headline... 13...and likewise for stung, etc.

Frame SRL: Experimental Setup SRL component of SEMAFOR 1.0 (Das et al., 2010; http://www.ark.cs.cmu.edu/semafor) Task scoring script for overt argument precision, recall, F1 on test set Strict matching criterion: argument spans must be exact 14

SRL on SE 10 Test Data P 0.65 0.67 0.67 (2 documents, ~500 sentences) 0.40 R Ch. 13 0.46 0.49 F1 F 0.50 0.54 0.57 Training Sets SE 07 SE 07 + ½ SE 10 SE 07 + SE 10 # sentences 2000 2250 2500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 15 SE 07: SEMAFOR trained only on old data (different domain from test set) SE 10: new training data included (same domain as test set) Adding a small amount of new data helps a lot: (~7% F1): domain issue + so little data to begin with. Suggests even more data might yield substantial improvements! Scores are microaveraged according to the number of frames in each of the 2 test documents.

Overview Background: frame SRL Overt argument identification Null instantiation resolution Conclusion 16

Null Instantiations New this year: classification and resolution of null instantiations (NIs), arguments that are nonlocal or implicit in the discourse a role is said to be null-instantiated if it has no (overt) argument in the sentence, but has an implicit contextual filler see also (Gerber & Chai, 2010), which considers implicit argument resolution for several (nominal) predicates (Fillmore, 1986; Ruppenhofer, 2005) 17

Null Instantiations indefinite null instantiation (INI): the referent is vague/deemphasized We ate Thing_eaten. He was stung Stimulus. definite null instantiation (DNI): a specific referent is obvious from the discourse They ll arrive soon Goal. (the goal is implicitly the speaker s location) (Fillmore, 1986; Ruppenhofer, 2005) 18

Null Instantiations indefinite null instantiation (INI): the referent is vague/deemphasized We ate Thing_eaten. He was stung Stimulus. definite null instantiation (DNI): a specific referent is obvious from the discourse They ll arrive soon Goal. (the goal is implicitly the speaker s location) (Fillmore, 1986; Ruppenhofer, 2005) 18

DNI Example: overt nonlocal referent I think I Experiencer shall be in a position to make the situation rather more clear to you before long. It has been an exceedingly difficult and most complicated business Experiencer. (SemEval 2010 test data) 19 The other frame-evoking words are bolded, but their arguments are not shown.

DNI Example: overt nonlocal referent I think I Experiencer shall be in a position to make the situation rather more clear to you before long. It has been an exceedingly difficult and most Degree complicated business Experiencer. Activity DIFFICULTY (SemEval 2010 test data) 19 The other frame-evoking words are bolded, but their arguments are not shown.

DNI Example: overt nonlocal referent I think I Experiencer shall be in a position to make the situation rather more clear to you before long. It has been an exceedingly difficult and most Degree complicated business Experiencer. Activity DIFFICULTY (SemEval 2010 test data) 19 The other frame-evoking words are bolded, but their arguments are not shown.

Prevalence of NIs 60 91 90 2,589 62 82% 277 INI 303 DNIs DNI, unresolved DNI, referent in same sentence DNI, referent within 3 previous sentences DNI, other referent Overt (SemEval 2010 new training data) 20 These numbers may be approximate. They show how few NIs there are compared to overt args, and why the DNI resolution task is so hard.

Modeling Approach for NIs SELF_MOTION.Place sprang (parse) SRL We try a straightforward approach for null instantiations: a IN PRP NN in his chair if a core role NI Resolution second classifier INI DNI DNI+referent NONE 21 The SRL module selects an argument span or none for each role. For core roles, we then build a second classifier for disambiguating types of null elements. This uses the same mathematical techniques to predict a different kind of outputs. Ideally, the NI module would be able to predict whether each core role was INI, DNI + its referent, if applicable, or not NI. Our system only considers DNIs with referents in the previous 3 sentences. Experiments show that a large search space, while leading to high *oracle* recall, confuses the model in practice.

Modeling Approach for NIs SELF_MOTION.Place sprang (parse) SRL We try a straightforward approach for null instantiations: a IN PRP NN in his chair if a core role NI Resolution second classifier INI DNI DNI+referent NONE 21 features encode roles null instantiation preferences The SRL module selects an argument span or none for each role. For core roles, we then build a second classifier for disambiguating types of null elements. This uses the same mathematical techniques to predict a different kind of outputs. Ideally, the NI module would be able to predict whether each core role was INI, DNI + its referent, if applicable, or not NI. Our system only considers DNIs with referents in the previous 3 sentences. Experiments show that a large search space, while leading to high *oracle* recall, confuses the model in practice.

Modeling Approach for NIs SELF_MOTION.Place sprang (parse) SRL We try a straightforward approach for null instantiations: a IN PRP NN in his chair if a core role NI Resolution second classifier INI DNI DNI+referent NONE nominals, NPs from previous 3 sentences as possible referents 21 The SRL module selects an argument span or none for each role. For core roles, we then build a second classifier for disambiguating types of null elements. This uses the same mathematical techniques to predict a different kind of outputs. Ideally, the NI module would be able to predict whether each core role was INI, DNI + its referent, if applicable, or not NI. Our system only considers DNIs with referents in the previous 3 sentences. Experiments show that a large search space, while leading to high *oracle* recall, confuses the model in practice.

NI-only results on SE 10 Test Data P Ch. 13 0.47 0.74 (2 documents, ~500 sentences) R 0.36 0.62 F1 F 0.49 0.53 Training Sets ½ SE 10 SE 10 # sentences 250 500 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 NIs only, oracle overt args 22 Also: NI subtask confusion matrix Evaluating NI performance only. We train only on the new SemEval 2010 data because the SemEval 2007 data used different annotation practices for null instantiations. The evaluation criterion actually doesn t distinguish between INIs and unresolved DNIs. We predicted only the former.

Overview Background: frame SRL Overt argument identification Null instantiation resolution Conclusion 23

Contributions & Claims 1. Evaluated frame SRL on new data Amount of training data makes a big difference Still lots of room for improvement 2. Experimented with a classifier for null instantiations, with mixed success Resolving nonlocal referents is much harder than classifying the instantiation type 3. Learned models achieve higher recall, and consequently F1, than custom heuristics used by other teams Our modeling framework is extensible: it should allow us to incorporate many of these in a soft way as features 24

Size of Data SE 07 SE 10 train SE 10 test PropBank 25 Sizes of frame-annotated data provided for SemEval 07 and 10 tasks, as compared to PropBank. The bottom graph is in terms of tokens. Whereas FrameNet provides a linguistically rich representation, PropBank has much higher coverage/annotated data.

Size of Data SE 3,000 sentences PropBank 0 5 10 15 20 25 30 35 40 45 50 thousands of sentences SE 07 SE 10 train SE 10 test PropBank 25 Sizes of frame-annotated data provided for SemEval 07 and 10 tasks, as compared to PropBank. The bottom graph is in terms of tokens. Whereas FrameNet provides a linguistically rich representation, PropBank has much higher coverage/annotated data.

Size of Data SE 3,000 sentences PropBank 50,000 sentences 0 5 10 15 20 25 30 35 40 45 50 thousands of sentences SE 07 SE 10 train SE 10 test PropBank 25 Sizes of frame-annotated data provided for SemEval 07 and 10 tasks, as compared to PropBank. The bottom graph is in terms of tokens. Whereas FrameNet provides a linguistically rich representation, PropBank has much higher coverage/annotated data.

Size of Data SE 3,000 sentences PropBank 50,000 sentences 0 5 10 15 20 25 30 35 40 45 50 thousands of sentences SE 07 SE 10 train SE 10 test PropBank SE frame annotations 15,000 frames PropBank predicates 113,000 predicates 0 16 33 49 66 82 99 115 thousands of instances 25 Sizes of frame-annotated data provided for SemEval 07 and 10 tasks, as compared to PropBank. The bottom graph is in terms of tokens. Whereas FrameNet provides a linguistically rich representation, PropBank has much higher coverage/annotated data.

Conclusion Next challenge: data sparseness in frame SRL obtaining quality frame annotations from experts is expensive opportunity for semi-supervised learning additional knowledge/constraints in modeling non-expert annotations? bridging across lexical-semantic resources (FrameNet, WordNet, PropBank, VerbNet, NomBank, ) 26

Task 10 (Frame SRL) Posters (101) CLR: Linking Events and Their Participants in Discourse Using a Comprehensive FrameNet Dictionary Ken Litkowski (102) VENSES++: Adapting a deep semantic processing system to the identification of null instantiations Sara Tonelli & Rodolfo Delmonte 27 if you re interested in this task

Thank you! http://www.ark.cs.cmu.edu/semafor Image from http://commons.wikimedia.org/wiki/file:sherlockholmes.jpg

Thank you! JUDGMENT_DIRECT_ADDRESS Communicator: DNI Reason: DNI Addressee http://www.ark.cs.cmu.edu/semafor Image from http://commons.wikimedia.org/wiki/file:sherlockholmes.jpg

Extra Slides NI subtask confusion matrix NI-only and full results table 29

NI-only Subtask: Confusion Matrix Gold Predicted overt DNI INI masked inc. total overt 2068 (1630) 5 362 327 0 2762 DNI 64 12 (3) 182 90 0 348 INI 41 2 214 96 0 353 masked 73 0 240 1394 0 1707 inc. 12 2 55 2 0 71 total 2258 21 1053 1909 0 3688 correct 30 from the paper

Results Table: NI-only and Full Chapter 13 Chapter 14 Training Data Prec. Rec. F 1 Prec. Rec. F 1 SemEval 2010 new: 100% 0.40 0.64 0.50 0.53 0.60 0.56 SemEval 2010 new: 75% 0.66 0.37 0.50 0.70 0.37 0.48 SemEval 2010 new: 50% 0.73 0.38 0.51 0.75 0.35 0.48 Full All 0.35 0.55 0.43 0.56 0.49 0.52 NI-only 31