Learning Parse Decisions From Examples With Rich Context. as submitted to ACL'96 on January 8, Ulf Hermjakob and Raymond J.

Similar documents
have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Parsing of part-of-speech tagged Assamese Texts

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Three New Probabilistic Models. Jason M. Eisner. CIS Department, University of Pennsylvania. 200 S. 33rd St., Philadelphia, PA , USA

Accuracy (%) # features

An Interactive Intelligent Language Tutor Over The Internet

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

CS 598 Natural Language Processing

AQUA: An Ontology-Driven Question Answering System

An Introduction to the Minimalist Program

Loughton School s curriculum evening. 28 th February 2017

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Linking Task: Identifying authors and book titles in verbose queries

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Cross Language Information Retrieval

Citation for published version (APA): Veenstra, M. J. A. (1998). Formalizing the minimalist program Groningen: s.n.

Proceedings of the 19th COLING, , 2002.

Using dialogue context to improve parsing performance in dialogue systems

Prediction of Maximal Projection for Semantic Role Labeling

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Ensemble Technique Utilization for Indonesian Dependency Parser

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Some Principles of Automated Natural Language Information Extraction

Learning Computational Grammars

A Case Study: News Classification Based on Term Frequency

Accurate Unlexicalized Parsing for Modern Hebrew

Natural Language Processing. George Konidaris

The College Board Redesigned SAT Grade 12

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

LTAG-spinal and the Treebank

Specifying a shallow grammatical for parsing purposes

Developing a TT-MCTAG for German with an RCG-based Parser

The Strong Minimalist Thesis and Bounded Optimality

Grammars & Parsing, Part 1:

MYCIN. The MYCIN Task

Language Acquisition Chart

A Grammar for Battle Management Language

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Constraining X-Bar: Theta Theory

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

SEMAFOR: Frame Argument Resolution with Log-Linear Models

Compositional Semantics

A Domain Ontology Development Environment Using a MRD and Text Corpus

Multiple case assignment and the English pseudo-passive *

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

LEARNING A SEMANTIC PARSER FROM SPOKEN UTTERANCES. Judith Gaspers and Philipp Cimiano

The stages of event extraction

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

Update on Soar-based language processing

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

The Discourse Anaphoric Properties of Connectives

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

In Udmurt (Uralic, Russia) possessors bear genitive case except in accusative DPs where they receive ablative case.

Distant Supervised Relation Extraction with Wikipedia and Freebase

A Comparison of Two Text Representations for Sentiment Analysis

Adapting Stochastic Output for Rule-Based Semantics

THE VERB ARGUMENT BROWSER

Infrastructure Issues Related to Theory of Computing Research. Faith Fich, University of Toronto

Context Free Grammars. Many slides from Michael Collins

Proof Theory for Syntacticians

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

National Literacy and Numeracy Framework for years 3/4

LING 329 : MORPHOLOGY

Improving Simple Bayes. Abstract. The simple Bayesian classier (SBC), sometimes called

cmp-lg/ Jul 1995

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Rule Learning With Negation: Issues Regarding Effectiveness

The Smart/Empire TIPSTER IR System

phone hidden time phone

A Computational Evaluation of Case-Assignment Algorithms

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Procedia - Social and Behavioral Sciences 154 ( 2014 )

Heuristic Sample Selection to Minimize Reference Standard Training Set for a Part-Of-Speech Tagger

Speech Recognition at ICSI: Broadcast News and beyond

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Software Maintenance

Transcription:

Learning Parse Decisions From Examples With Rich Context as submitted to ACL'96 on January 8, 1996 Ulf Hermjakob and Raymond J. Mooney Dept. of Computer Sciences University of Texas at Austin Austin, TX 78712, USA ulf@cs.utexas.edu mooney@cs.utexas.edu Abstract We present a knowledge and context-based system for parsing natural language and evaluate it on sentences from the Wall Street Journal. Applying machine learning techniques, the system uses parse action examples acquired under supervision to generate a deterministic shift-reduce parser in the form of a decision structure. It relies heavily on context, as encoded in features which describe the morpholgical, syntactical, semantical and other aspects of a given parse state. 1. INTRODUCTION The parsing of unrestricted text, with its enormous word and structural ambiguity, still poses a great challenge in natural language processing. While systems that attempted to capture parse grammars in absolute hand-coded rules just could not penetrate the complexity of ambiguity of unrestricted text, probabilistic approaches with often only very limited context sensitivity have hit a performance ceiling even when trained on very large corpora. To cope with the complexity of unrestricted text, parse rules of any kind of formalism will have to consider a complex context with many dierent features of e.g. morphological, syntactic or semantic type. This can present a signicant problem, because even linguistically trained natural language developers have great diculties writing and even more so extending explicit parse grammars covering a wide range of natural language. On the other hand it is much easier for humans to decide how specic sentences should be analyzed, although for some tricky sentences even this can still require substantial linguistic expertise. We therefore propose an approach to parsing based on example-based learning with a very strong emphasis on context, integrating morphological, syntactical, semantical and other aspects relevant to making good parse decisions, thereby also allowing the parsing to be deterministic. Applying machine learning techniques, the system uses parse action examples acquired under supervision to generate a deterministic shift-reduce type parser in the form of a decision structure. The generated parser transforms input sentences into a phrase-structure and case-frame tree, powerful enough to be fed into a transfer and a generation module to complete the full process of machine translation. Relieving the NL-developer from the hard if not impossible task of writing an explicit grammar, the focus on relevant features at the same time requires a relatively modest number of training examples when compared to more statistically oriented approaches. The approach we take provides a high degree of manageability of the `grammar' when increasing its coverage. 1

2. BASIC PARSING PARADIGM As the basic mechanism for parsing text into a shallow semantic representation, we choose a shift-reduce type parser (Marcus, 1980). It breaks parsing into an ordered sequence of small and manageable parse actions such as shift and reduce. This ordered `left-to-right' parsing is much closer to how humans parse a sentence than e.g. chart oriented parsers; it allows a very transparent control structure and makes the parsing process relative intuitive for humans. This is very important, because during the training phase, the system is guided by a human supervisor for whom the ow of control needs to be as transparent and intuitive as possible. The parsing does not have separate phases for part-of-speech selection and syntax and semantic processing, but rather integrates all of them into a single parsing phase. Since the system has all morphological, syntactical and semantical context information available at all times, the system can make well-based decisions very early, allowing a single path, i.e. deterministic parse, which eliminates wasting computation on `dead end' alternatives. Before the parsing itself starts, the input string is segmented into a list of words incl. punctuation marks, which then are sent through a morphological analyzer that, using a lexicon, produces primitive frames for the segmented words. There is one primitive frame for each word and part of speech that the word can be, so that several primitive frames for one word would reect ambiguity of the word with respect to part of speech. (Morphological ambiguity is captured within a frame.) parse stack input list "John" "called" synt: verb "Mary" * "again" synt: adv (R 2 TO S-VP AS PRED (OBJ PAT)) "reduce the 2 top elements of the parse stack to a frame with syntax vp and roles pred and obj and pat " "John" "called Mary" synt: vp sub: (pred) (obj pat) * "again" synt: adv "called" synt: verb "Mary" Figure 1: A typical parse action (simplied); boxes represent frames The central data structure for the parser contains a parse stack and an input list. The parse stack and the input list contain trees of frames of words or phrases. Core slots of frames are surface and lexical form, syntactic and semantic category, subframes with syntactic and semantic roles, and form restrictions such as number, person, and tense. Other slots can include information like the numerical 2

value of number words, ags whether or not a (German) verb has a separable or inseparable prex etc. Initially, the parse stack is empty and the input list contains the primitive frames produced by the morphological analyzer. After initialization, the deterministic parser applies a sequence of parse actions to the parse structure. The most frequent parse actions are shift, which shifts a frame from the input list onto the parse stack or backwards, and reduce, which combines one or several frames on the parse stack into one new frame. The frames to be combined are typically, but not necessarily, next to each other at the top of the stack. As shown in gure 1, the action (R 2 TO VP AS PRED (OBJ PAT)) for example reduces the two top frames of the stack into a new frame that is marked as a verb phrase and contains the next-to-the-top frame as its predicate (or head) and the top frame of the stack as its object and patient. Other parse actions include `add-into' which adds frames arbitrarily deep into an existing frame tree, `mark' that marks some slot of some frame with some value and operations to introduce empty categories (traces and `PRO', as in \She i wanted PRO i to win."). Parse actions can have numerous arguments, making the parse action language very powerful. The parse action examples needed for training the system are acquired interactively. For each training sentence, the system and the supervisor parse the sentence step by step with the supervisor entering the next parse action command, e.g. (R 2 TO VP AS PRED (OBJ PAT)), and the system executing it, repeating this sequence until the sentence is fully parsed. At least for the very rst sentence, the supervisor actually has to type in the full parse action commands. With a growing number of parse action examples available, the system, as described below in more detail, can be trained on those previous examples. In such a partially trained system, the parse actions are then proposed by the system using a parse decision structure which \classies" the current context. The proper classication is the specic action or sequence of actions that (the system believes) should be performed next. During further training, the supervisor then enters parse action commands by either conrming what the system proposes or overruling it by providing the proper action. As the corpus of parse example grows and the system gets trained on more and more data, the system becomes more rened, so that the supervisor has to overrule the system with decreasing frequency. The sequence of correct parse actions for a sentence is then recorded in a log le. 3. FEATURES To make good parse decisions, a wide range of features at various degrees of abstraction have to be considered. To express such a wide range of features, we dened a feature language. Given a particular parse state and a feature, the system can interpret the feature and compute its value for the given parse state, often using additional knowledge resources such as 1. a general knowledge base (KB), which currently consists of a directed acyclic graph of concepts, with currently 3608 is-a-relationship links, e.g. \CAR N OU N?CON CEP T is-a V EHICLE N OU N?CON CEP T", and 2. subcategorazation tables that describe the syntactic and semantic role structure(s) for verbs and nouns with currently close to a total of 200 entries. The following examples, for easier understanding rendered in English and not in feature language syntax, illustrate the expressiveness of the feature language: the general syntactic class of the top element of the stack (e.g. adjective, noun phrase), 3

the specic nite tense of the second stack element (e.g. present tense, past tense), whether or not some element could be a nominal degree adverb, whether or not some phrase already contains a subject, the semantic role of some noun phrase with respect to some verb phrase (e.g. agent, time; this involves pattern matching with corresponding entries in the verb subcategorization table), whether or not some noun and verb phrase agree. Features can in principal refer to any element on the parse stack or input list, and any of their subelements, at any depth. Since all of the currently 151 features are supposed to bear some linguistic relevance, none of them actually refer to anything too far removed from the current focus of a parse state. The set of features is globally used for all examples and can easily be extended when the need arises. 4. LEARNING DECISION STRUCTURES Traditional statistical techniques also use features, but often have to sharply limit their number (for some trigram approaches to three fairly simple features) to avoid overwhelming computational complexity. In parsing, only a very small number of features are crucial over a wide range of examples, while most features are critical in only a few examples, being used to `ne-tune' the decision structure for special cases. So in order to overcome the antagonism between the importance of having a large number of features and the need to control the number of examples required for learning, particularly when acquiring examples under supervision, we choose a decision-tree-type learning algorithm, which recursively selects the most discrimating feature of the corresponding subset of training examples, eventually ignoring all locally irrelevant features, thereby tailoring the size of the nal decision structure to the complexity of the training data. Given a set of features, the system can feed parse actions from the log le through the parse engine, and automatically augment each parse action with its corresponding feature-value vector. While parse actions might be complex for the action interpreter, they are atomic with respect to the decision structure learner; e.g. \(R 2 TO VP AS PRED (OBJ PAT))" would be such an atomic classication. A set of examples consisting of a feature-value vector and a parse action as a classication is then fed into an ID3-based learning routine that generates a decision structure, which can then `classify' any given parse state (which implicitely assigns a values to every feature) by proposing what parse action to perform next. As an extension to the standard ID3 model (Quinlan, 1986), our system uses a exible decision structure that currently consists of a decision list of hierarchical decision trees. 5. WALL STREET JOURNAL EXPERIMENTS We now present intermediate results on training and testing a prototype implementation of the system with sentences from the Wall Street Journal, a prominent corpus of `real' text, as collected on the ACL-CD. In order to limit the size of the required lexicon, we work on a reduced corpus that includes all those sentences that are fully covered by the 3000 most frequently occurring words (ignoring numbers etc.) in the entire corpus. The rst 144 sentences used in this experiment vary in length from 4 to 45 words, averaging at 17.0 words. One of these sentence is \Canadian manufacturers' new orders fell 4

to $20.80 billion (Canadian) in January, down 4% from December's $21.67 billion on a seasonally adjusted basis, Statistics Canada, a federal agency, said.". For the following test series, the corpus of 144 sentences that currently have parse action logs associated with them is divided into 9 blocks of 16 sentences. Each of these 9 blocks is then consecutively used for testing. For each of the 9 sub-tests, a varying number of sentences from the other blocks is used for training the parse decision structure, so that within a sub-test, none of the training sentences are ever used as a test sentence. The results of the 9 sub-tests of each series are then averaged. Number of training sentences 16 32 64 128 Precision 84.6% 84.6% 87.3% 89.9% Recall 80.4% 83.0% 86.3% 89.4% Labelled precision 81.0% 81.6% 85.1% 88.3% Labelled recall 77.4% 80.0% 83.8% 87.6% Tagging accuracy 97.2% 97.3% 97.5% 97.9% Crossings per sentence 2.6 2.5 2.0 1.8 Sent. with 0 crossings 27.8% 30.6% 39.6% 42.4% Sent. with up to 1 crossing 46.5% 48.6% 58.3% 61.8% Sent. with up to 2 crossings 59.7% 60.4% 69.4% 76.4% Sent. with up to 3 crossings 73.6% 75.0% 81.3% 84.7% Sent. with up to 4 crossings 79.2% 81.3% 88.2% 87.5% Correct operations 78.8% 82.4% 86.1% 89.3% Sent. with correct OpSequence 2 5 7 19 Sent. with correct Struct&Label 5 8 18 28 Sentences with endless loop 4 2 2 0 Table 1: Evaluation results using a total of 144 test sentences for each of the four test series number of correct constituents in system parse Precision: number of constituents in system parse number of correct constituents in system parse Recall: number of constituents in logged parse Crossing brackets: number of constituents which violate constituent boundaries with a constituent in the logged parse. Labelled precision/recall measures not only structural correctness, but also the correctness of the syntactic label. Correct operations measures the number of correct operations during a parse continuously corrected based on the logged sequence. A sentence has a correct operating sequence, OpSequence, if the system fully predicts the logged parse action sequence, and a correct struture and labeling, Struct&Label, if the structure and syntactic labeling of the nal system parse of a sentence is 100% correct, regardless of the operations leading to it. The current set of 151 features was sucient to always discrimate examples with dierent parse actions, thereby producing 100% test-accuracy on sentences that the system had previously been trained on. While that percentage is certainly less important than the accuracy gures for unseen sentences, it nevertheless represents an upper ceiling, which for many statistical systems lies signicantly below 100%. Many of the mistakes are due to encountering constructions that just have not been seen before at all, typically causing several erroneous parse decisions in a row. This observation also supports our hope that with more training sentences, the testing accuracy for unseen sentence will still rise signicantly. 5

6. RELATED WORK Our basic parsing and interactive training paradigm is based on (Simmons and Yu, 1992). We have extended their work by signicantly increasing the expressiveness of the parse action and feature languages, in particular by moving far beyond a few simple features, limited to syntax only, by adding several knowledge resources and by introducing a sophisticated machine learning component. (Magerman, 1995) uses a decision tree model similar to ours, training his system SPATTER with parse action sequences for 40,000 Wall Street Journal sentences derived from the Penn Treebank. Questioning the traditional n-grams, Magerman already advocates a heavier reliance on contextual information. Going beyond Magerman's still relatively rigid set of 36 features, we propose a yet richer, basically unlimited feature language and set. Our parse action sequences are too complex to be derived from a treebank like Penn's. While this necessitates the involvement of a supervisor for training, we are able to perform deterministic parsing and get already very good test results for only 128 training sentences. 7. CONCLUSION We try to bridge the gap between the typically hard-to-scale, context-rich rule approach and the typically large-scale, context-poor probabilistic approach for unrestricted text parsing. Using rich and unied context and a complex parse action language, we can demonstrate good intermediate results for single-pass deterministic parsing. While many limited-context probabilistic approaches have already reached a performance ceiling, we still expect to signicantly improve our results when increasing our training base beyond the currently 128 sentences. Even then the training size will compare favorably with the huge number of training sentences necessary for many probabilistic systems. REFERENCES E. Black, J. Laerty, and S. Roukos. 1992. Development and evaluation of a broad-coverage probabilistic grammar of English-language computer manuals. In 30th ACL Proceedings, pages 185{192. Graeme Hirst. 1987. Semantic interpretation and the resolution of ambiguity. Cambridge University Press. D. M. Magerman. 1995. Statistical Decision-Tree Models for Parsing In Proceedings of ACL. M. P. Marcus. 1980. A Theory of Syntactic Recognition for Natural Language. MIT Press. S. Nirenburg, J. Carbonell, M. Tomita, and K. Goodman. 1992. Machine Translation: A Knowledge- Based Approach. San Mateo, CA: Morgan Kaufmann. J.R. Quinlan. 1986. Induction of decision trees. In Machine Learning 1 (1). Robert F. Simmons and Yeong-Ho Yu. 1992. The Acquisition and Use of Context-Dependent Grammars for English. In Computational Linguistics, December 1992. 6