Specifying a shallow grammatical for parsing purposes

Similar documents
arxiv:cmp-lg/ v1 7 Jun 1997 Abstract

CS 598 Natural Language Processing

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Advanced Grammar in Use

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

ELD CELDT 5 EDGE Level C Curriculum Guide LANGUAGE DEVELOPMENT VOCABULARY COMMON WRITING PROJECT. ToolKit

Developing a TT-MCTAG for German with an RCG-based Parser

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Refining the Design of a Contracting Finite-State Dependency Parser

THE VERB ARGUMENT BROWSER

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

BULATS A2 WORDLIST 2

Definition Corpus for Finnish Voutilainen, Atro; Linden, Krister; Purtonen, Tanja Katariina Voutilainen, A, Linden, K & Purtonen, T K 2011, '

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Emmaus Lutheran School English Language Arts Curriculum

Loughton School s curriculum evening. 28 th February 2017

Today we examine the distribution of infinitival clauses, which can be

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Words come in categories

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

The Discourse Anaphoric Properties of Connectives

The College Board Redesigned SAT Grade 12

California Department of Education English Language Development Standards for Grade 8

An Interactive Intelligent Language Tutor Over The Internet

Adjectives tell you more about a noun (for example: the red dress ).

Writing a composition

Grammars & Parsing, Part 1:

Opportunities for Writing Title Key Stage 1 Key Stage 2 Narrative

Underlying and Surface Grammatical Relations in Greek consider

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Chapter 4: Valence & Agreement CSLI Publications

Context Free Grammars. Many slides from Michael Collins

a) analyse sentences, so you know what s going on and how to use that information to help you find the answer.

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Subject: Opening the American West. What are you teaching? Explorations of Lewis and Clark

Possessive have and (have) got in New Zealand English Heidi Quinn, University of Canterbury, New Zealand

Accurate Unlexicalized Parsing for Modern Hebrew

SAMPLE. Chapter 1: Background. A. Basic Introduction. B. Why It s Important to Teach/Learn Grammar in the First Place

Heuristic Sample Selection to Minimize Reference Standard Training Set for a Part-Of-Speech Tagger

FinnTreeBank: Creating a research resource and service for language researchers with Constraint Grammar

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

GERM 3040 GERMAN GRAMMAR AND COMPOSITION SPRING 2017

Ch VI- SENTENCE PATTERNS.

A First-Pass Approach for Evaluating Machine Translation Systems

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

What the National Curriculum requires in reading at Y5 and Y6

Author: Justyna Kowalczys Stowarzyszenie Angielski w Medycynie (PL) Feb 2015

Intension, Attitude, and Tense Annotation in a High-Fidelity Semantic Representation

Senior Stenographer / Senior Typist Series (including equivalent Secretary titles)

LING 329 : MORPHOLOGY

AQUA: An Ontology-Driven Question Answering System

Proof Theory for Syntacticians

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Parsing of part-of-speech tagged Assamese Texts

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

EAGLE: an Error-Annotated Corpus of Beginning Learner German

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Linking Task: Identifying authors and book titles in verbose queries

Sample Goals and Benchmarks

Chapter 9 Banked gap-filling

Procedia - Social and Behavioral Sciences 154 ( 2014 )

Participate in expanded conversations and respond appropriately to a variety of conversational prompts

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

On the Notion Determiner

Development of the First LRs for Macedonian: Current Projects

The Smart/Empire TIPSTER IR System

Some Principles of Automated Natural Language Information Extraction

The Role of the Head in the Interpretation of English Deverbal Compounds

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

Developing Grammar in Context

ELA/ELD Standards Correlation Matrix for ELD Materials Grade 1 Reading

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

Progressive Aspect in Nigerian English

Universal Grammar 2. Universal Grammar 1. Forms and functions 1. Universal Grammar 3. Conceptual and surface structure of complex clauses

University of Waterloo School of Accountancy. AFM 102: Introductory Management Accounting. Fall Term 2004: Section 4

National Literacy and Numeracy Framework for years 3/4

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Nancy Hennessy M.Ed. 1

A Framework for Customizable Generation of Hypertext Presentations

5 th Grade Language Arts Curriculum Map

Character Stream Parsing of Mixed-lingual Text

Comprehension Recognize plot features of fairy tales, folk tales, fables, and myths.

Impact of Controlled Language on Translation Quality and Post-editing in a Statistical Machine Translation Environment

Controlled vocabulary

Control and Boundedness

Adapting Stochastic Output for Rule-Based Semantics

5. UPPER INTERMEDIATE

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS

Constraining X-Bar: Theta Theory

Welcome to the Purdue OWL. Where do I begin? General Strategies. Personalizing Proofreading

Interactive Corpus Annotation of Anaphor Using NLP Algorithms

Copyright and moral rights for this thesis are retained by the author

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Appendix D IMPORTANT WRITING TIPS FOR GRADUATE STUDENTS

Transcription:

Specifying a shallow grammatical for parsing purposes representation Atro Voutilainen and Timo J~irvinen Research Unit for Multilingual Language Technology P.O. Box 4 FIN-0004 University of Helsinki Finland At ro.voutilainen,timo. J arvinen@helsinki.fi Abstract Is it possible to specify a grammatical representation (descriptors and their application guidelines) to such a degree that it can be consistently applied by different grammarians e.g. for producing a benchmark corpus for p.arser evaluation? Arguments for and against have been given, but very little empirical evidence. In this article we report on a double-blind experiment with a surfaceoriented morphosyntactic grammatical representation used in a large-scale English parser. We argue that a consistently applicable representation for morphology and also shallow syntax can be specified. A grammatical representation with a near- coverage of running text can be specified with a reasonable effort, especially if the representation is based on structural distinctions (i.e. it is structurally resolvable). Introduction The central task of a parser is to assign grammatical descriptions onto input sentences. Evaluating a parser's output (as well as designing a computational lexicon and grammar) presupposes a predefined, parser-independent specification of the grammatical representation. Perhaps surprisingly, the possibility of specifying a workable grammatical representation is a matter of controversy, even at lower levels of analysis, e.g. morphology (incl. parts of speech). Consider the following setting (the double-blind experiment). Two linguists trained to apply a tag set to running text according to application guidelines (a "style sheet") are to analyse a given data individually. The results are then automatically compared, and the differences are jointly examined by these linguists to see whether the differences are due to inattention, or whether they are intentional (i.e. there is a genuine difference in analysis). - How many percentage points of all words in running text are retain a different analysis after the differences due to inattention have been omitted? The higher this percentage, the more susceptible seems the possibility of specifying a workable grammatical representation. According to a pessimistic view (e.g. Church 992), the part of speech of several percentage points of words in running text is impossible to agree on by different judges, even after negotiations. A more optimistic view can be found in (Leech and Eyes 993, p. 39; Marcus et al. 993, p. 328); they argue that a near- interjudge agreement is possible, provided the part-of-speech annotation is done carefully by experts. Unfortunately, they give very little empirical evidence for their position e.g. in terms of double-blind experiments. Supposing defining these lower levels of grammatical representation is so problematic, the more distinctive levels should be even more difficult. If specifying the task of the parser - what the parser is supposed to do - turns out to be so problematic, one could even question the rationality of natural language parser design as a whole. In other words, the controversy regarding the specifiability of a grammatical representation is a fundamental issue. In this article we report on a double-blind experiment with a surface-oriented morphosyntactic grammatical representation used by a large-scale English parser. We show that defining a grammatical representation is possible, even relatively straightforward. We present results from part-ofspeech annotation and shallow syntactic analysis. Our three main findings are:. A practically interjudge agreement can be reached at the level of morphological (incl. part-of-speech) analysis provided that (i) the grammatical representation is based on structural distinctions and (ii) the individual descriptive practices of the most frequent 'problem cases' are properly documented. 2. A shallow dependency-oriented functional syntax can be defined, very much like a morphological representation. The only substan- 20

tial difference seems to be that somewhat more effort for documenting the individual solution is needed at the level of syntax. 3. A grammatical representation (morphosyntactic descriptors and their application guidelines) can be specified with a reasonable effort. In addition to general descriptive principles, only a few dozen construction-specific entries seem necessary for reaching a high coverage of running text. In short: In this paper we give empirical evidence for the possibility of specifying a grammatical representation in enough detail to make it (almost) consistently applicable. What we are less specific about here is the exact formal properties that make a representation easy to specify; this topic remains open for future investigation. 2 Grammatical representation in English Constraint Grammar In the experiment to be reported in Section 3, we employed the grammatical representation that defines the descriptive task of the English Constraint Grammar Parser ENGCG (Karlsson et al. (eds.) 995). 2. Morphology The morpholexical component in ENGCG employs 39 morphological tags for part of speech, inflection, derivation and certain syntactic properties (e.g. verb classification). Each morphological analysis usually consists of several tags, and many words get several analyses as alternatives. The following analysis of the sentence That round table might collapse is a rather extreme example: "<round>" "table" "table" <*> <**CLB> CS <*> DET CENTRAL DEM SG <*> ADV <*> PRON DEM SG <*> <**CLB> <Rel> PRON SG/PL <SVO> V SUBJUNCTIVE VFIN <SVO> V IMP VFIN <SVO> V INF <SVO> V PRES -SG3 VFIN PREP N NOM SG A ABS ADV N NOM SG <SVO> V SUBJUNCTIVE VFIN aa list of the ENGCG tags can be retrieved via e-mail by sending an empty mail message to engcginfo@ling.helsinki.fi. The returned document will also tell how to analyse own samples using the ENGCG server. "table" <SVO> V IMP VFIN "table" <SV0> V INF "table" <SVO> V PRES -SG3 VFIN "might" N NOM SG "might" V AUXMOD VFIN "collapse" N NOM SG "collapse" <SVO> V SUBJUNCTIVE VFIN "collapse" <SVO> V IMP VFIN "collapse" <SVO> V INF "collapse" <SVO> V PRES -SG3 VFIN,,<$.>,, The morphological analyser produces about 80 different tag combinations. To compare the ENGCG morphological description with another well-known tag set, the Brown Corpus tag set: ENGCG is more distinctive in that the part of speech distinction is spelled out in the description of determiner-pronoun, preposition-conjunction, and determiner-adverb-pronoun homographs, as well as uninflected verb forms, which are represented as ambiguous due to the subjunctive, imperative, infinitive and present tense readings. On the other hand, ENGCG does not spell out partof-speech ambiguity in the description of -ing and nonfinite -ed forms, noun-adjective homographs when the core meanings of the adjective and noun readings are similar, nor abbreviations vs. proper vs. common nouns. Generally, the ENGCG morphological tag set avoids the introduction of structurally unjustified distinctions. 2.2 Syntax ENGCG syntax employs 30 dependency-oriented functional tags that indicate the surface-syntactic roles of nominal heads (subject, object, preposition complement, apposition, etc.) and modifiers (premodifiers, postmodifiers). The shallow structure of verb chains is also given - the tag set distinguishes between auxiliaries and main verbs, finite and nonfinite. Also the structure of adverbials as well as prepositional and adjective phrases is given, though some of the attachments of adverbials is left underspecified. Finally, a disambiguated sample analysis of the above sample sentence: <*> DET CENTRAL DEM SG @DN> "<ro~d>' A ABS @AN> "table" N N0M SG @SUBJ "might" V AUXMOD VFIN ~+FAUXV "collapse" <SV0> V INF @-FMAINV "<$. >,, 2

Syntactic tags are flanked with the @-sign; 2 morphological tags and the base form are given to the left of the syntactic tags. 3 The experiment This section reports on an experiment on partof-speech and syntactic disambiguation by human experts (the authors of this article). Three 2,000- word texts were successively used: a software manual, a scientific magazine, and a newspaper. 3. Setting The experiment was conducted as follows.. The text was morphologically analysed using the ENGCG morphological analyser. For the analysis of unrecognised words, we used a rule-based heuristic component that assigns morphological analyses, one or more, to each word not represented in the lexicon of the system. 2. Two experts in the ENGCG grammaticalrepresentation independently marked the correct alternative analyses in the.ambiguous input, using mainly structural, but in some structurally unresolvable cases also higher-level, information. The corpora consisted of continuous text rather than isolated sentences; this made the use of textual knowledge possible in the selection of the correct alternative. In the rare cases where two analyses were regarded as equally legitimate, both could be marked. The judges were encouraged to consult the documentation of the grammatical representation. 3. These tagged versions were compared to each other using the Unix sdiff program. 4. The differences were jointly examined by the judges in order to see whether they were due to (i) inattention, (ii) incomplete specification of the grammatical representation or (iii) an undecidable analysis. 5. A 'consensus' version of the tagged corpus was prepared. Usually only a unique analysis was given. However, there were three situations where a multiple analysis was accepted: When the judges disagree about the correct analysis even after negotiations. In this case, comments were added to distinguish it from the other two types. Neutralisation: both analyses were regarded as equivalent. (This often indicates a redundancy in the lexicon.) 2,,@DN>, represents determiners; "@AN>" represents premodifying adjectives; "@SUB J" represents subjects; "@+FAUXV" represents finite auxiliaries; and "@-FMAINV" represents nonfinite m~in verbs. Global ambiguity: the sentence was agreed to be globally ambiguous. 6. Whenever an undefined construction was detected during the joint examination, the grammar definition manual was updated. 7. The preparation of the syntactic version was the next main step. For each contextually appropriate morphological reading, all syntactic tags were introduced with a mapping program. An example: 3 <*> DET CENTRAL DEM SG @DN> "<r ouzld>" A ABS @AN> "table" N NOM SG @NPHR ~SUBJ @OBJ @I-OBJ @PCOMPL-S @PCOMPL-0 @APP @NN> @<P @O-ADVL "might" V AUXMOD VFIN @+FAUXV "collapse" <SVO> V INF @-FMAINV @<P-FMAINV @<NOM-FMAINV -<$.>,, 8. Steps 2-6 were applied to these syntactic ambiguities. This procedure was successively applied to the three texts to see how much previous updates of the grammar definition manual decreased the need for further updates and how much the interjudge agreement might increase even after the first mechanical comparison (cf. Step 3). 3.2 Results The results are given in Figure (next page). Some comments are in order, first about morphology. The initial consistency rate was constantly above 99%. After negotiations, the judges agreed about the correct analysis or analyses in all cases. The vast majority of the initial differences were due to inattention, and the remaining few to incomplete specification of the morphological representation. Some representative examples about these jointly examined 3,,@NPHR, represents stray nominal heads; "@OBJ" represents objects; "@I-OBJ" represents indirect objects; "@PCOMPL-S" represents subject complements; "@PCOMPL-O" represents object complements; "@APP" represents appositions; "@NN>" represents premodifying nouns (and nonfinal noun parts in compounds); "@<P" represents nominal preposition complements; "@O-ADVL" represents nominal adverbials; "@<P-FMAINV" represents nonfinite m~in verbs as preposition complements; and "@<NOM-FMAINV" represents postmodifying nonfinite main verbs. 22

- that text news technical magazine ] total words [[ 999 999 2073 607 morph.tags/word in input morphologically ambiguous words agreement after mechanical comparison updates to morphology manual agreement after negotiations morph.tags/word in consensus corpus.78 4.% 99.3%.00(+2) syn.tags/word in input 3.50 syntactically ambiguous words 42.0% agreement after mechanical comparison 95.8% updates to syntax manual agreement after negotiations 5 syn.tags/word in consensus corpus.0 (+8).95 45.4% 99.3%.00 (+0) 3.53 4.9% 97.0%.0 (+) Figure : Results of a tagging test..72 36.8% 99.%.00 (+ ) 3.36 44.9% 97.4% '.00 (+3).82 4.0% 99.2% 3.00(+3) 3.46 42.9% 96.8% 7.0 (+32) differences are in order. (Words followed by an expression of the form (X/Y) were initially tagged differently by the judges. After joint examination, Y was agreed to be the correct alternative in all cases but (5), where X and Y were regarded as equally possible.) 4. As we go(v INF / Y PRES) to(inf- MARK / PREP) press(v INF / N), George Bush's decision not to sign the Biodiversity Convention, and Britain's apparent intention to follow suit, are grievous blows.. 2... they were circulating a letter expressing concern that(pron REL / CS) it would give the developing countries a blank cheque to demand money from donors to finance sustainable development. 3. That(PRON DEM / CS) there was no outburst of protest over the new policy suggests that public anxiety over genetic engineering has ebbed in recent years. 4. The value-added information is the kind(a /N) we want ourselves." 5... they had not seen before at one(num / PRON) of the busiest times of the school year. 6. I don't think people get(v INF / V PRES) a great deal from bald figures. 7. She had to ask because some of the six-year-olds from other schools who attend(v INF / V PRES) her classes know the names of as(prep/ad-a>) many hard drugs as she does. Only three ulsdates were needed to the morphological part of the manual. 4Before an "of" phrase, the pronoun/numeral distinction of "one" was regarded as neutralised. This observation was also added to the morphology manual. Though multiple analyses were considered acceptable in the case of (even semantically) undecidable situations, very few were actually needed: only 3 words out of 6,07 received two analyses (for example, it was agreed that more could be analysed both as an adverb and as a pronoun in.. free trade will mean you destroy more.). Next, some observations about syntax. At the level of syntax, most of the initial differences were identified as obvious mistakes, e.g.: - He was (@+PMAINV/@ FA UXV) addressing his hosts.. Sometimes, however, there was a need to discuss the descriptive policies. Consider the following sentence fragment: 5 managers'(@gn>) keeping?@-fmainv / @SUB J)in(@ADVL / @<NOM) touch with employees enhances communication.. In principle, managers'could be described as a subject in a nonfinite clause, and keeping accordingly as a nonfinite main verb. However, the ENGCG syntactic representation does not recognise the subject category in nonfinite clauses; therefore, in the name of consistency, keeping in the above example should be assigned a nominal rather than a verbal function - finite clause subject, in this case. Initially, the syntactic representation was less completely specified than the morphological representation. The grammar definition manual initially comprised twelve entries for syntactic functions; seven additions were made s"@gn>" represents genitival premoditlers, and "~<NOM" represents postmodifiers. 23

during the experiment. This had a positive effect: the initial disagreements decreased from 4.2% to 2.6% during the three rounds. The entries in the syntax manual can be classifted into three types:. Two or more alternatives are structurally plausible, but one is to be consistently preferred; e.g. A number that occurs after a proper noun and is surrounded by commas is a postmodifier (rather than an apposition) 2. Elimination of a distinction in certain contexts; e.g. Premodifying -ing forms are to be analysed as adjectives (rather than as nouns) 3. An unorthodox policy is adopted; e.g. In sentences with a formal subject, what is usually regarded as a notional subject is here analysedas a subject complement. Multiple analyses were given to 32 words (0.5% of all words). PP attachment, in particular the distinction between clause level (@ADVL) and postmodifying (@<NOM) functions, proved to be the most difficult syntactic phenomenon to define uniquely; often the analyses remained somewhat indeterminate. With the first sample, 5.7% of all prepositions were initially annotated differently; even with the last sample, 4.5% of the initial analyses differed. Unsurprisingly, a frequent agreement in the analysis of these cases was to accept both alternatives as legitimate. Some further examples are in order. These examples show some possible structural ambiguities in which text-level semantic information was needed to decide upon the preferred analysis Y over less plausible alternative X. Note that the adopted decision often determines the correct analysis of one or more subsequent items (the "domino effect").... his priority was (@ FA UXV / @+FMAINV) keeping his country's biotechnology industry free.. 2. Germany wants the heads of European governments and perhaps Japan (@OBJ / @<P) to issue a 'declaration of likeminded par~ies". 3. We were (@-FMAINV / @+FAUXV) pleased (@PCOMPL-S / @-FMAINV) with (@<NOM / @ADVL) the report. The last type was recurrent because the ENGCG morphology offers only a past participle reading to -ed-forms. We prefer the verbal reading and predicative -ed-forms are listed as exceptions in the coding manual. 4 Conclusion A satisfactory definition of the grammatical representation appears possible, not only at the level of morphology, but also at the level of shallow dependency-oriented functional syntax. In our experiments, a practically consensus was reached at both these levels during the joint examination. Our results agree, at least at the level of morphology, with (Leech and Eyes 993; Marcus et al. 993). In our experiment, the main differences between morphology and syntax were that (i) specifying the syntactic representation takes a few more pages in the definition manual, and (it) there seem to be more cases in syntax where multiple analyses have to be accepted - but relatively few even then. The grammatical representation should employ intuitively clear grammatical descriptors that (i) represent all constructions in the language and (it) reflect distributional distinctions. Proposing a too fine-grained classification of e.g. -ing forms, as may be the case in the tagged Brown Corpus, can make the principled selection of the appropriate analysis very difficult, even with detailed manuals. As a minor point we may add that errors due to inattention tend to occur in the preparation of e.g. benchmark corpora; however, almost all of them can be eliminated using the double-blind method. Acknowledgements We would like to thank Jussi Piitulainen, Past Tapanainen and two EACL referees for useful comments on a previous version of this paper. The usual disclaimers hold. References Kenneth W. Church 992. Current Practice in Part of Speech Tagging and Suggestions for the Future. In Simmons (ed.), Sbornik praci: In Honor of Henry KuSera, Michigan Slavic Studies. Michigan. 3-48. Elizabeth Eyes and Geoffrey Leech 993. Syntactic Annotation: Linguistic Aspects of Grammatical Tagging and Skeleton Parsing. In Ezra Black, Roger Garside and Geoffrey Leech (eds.) 993. Statistically-Driven Computer Grammars of English: The IBM/Lancaster Approach. Amsterdam and Atlanta: Rodopi. 36-6. Fred Karlsson, Atro Voutilainen, Juha Heikkil and Arto Anttila (eds.) 995. Constraint Grammar. A Language-Independent System for Parsing Unrestricted Text. Berlin and New York: Mouton de Gruyter. Mitchell Marcus, Beatrice Santorini and Mary Ann Marcinkiewicz 993. Building a Large Annotated Corpus of English: The Penn Treebank. Computational Linguistics 9:2. 33-330. 24