Dependency grammar and dependency parsing

Similar documents
11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Grammars & Parsing, Part 1:

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

Some Principles of Automated Natural Language Information Extraction

CS 598 Natural Language Processing

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

Chapter 4: Valence & Agreement CSLI Publications

Parsing of part-of-speech tagged Assamese Texts

Context Free Grammars. Many slides from Michael Collins

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Ensemble Technique Utilization for Indonesian Dependency Parser

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

Specifying a shallow grammatical for parsing purposes

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Parsing natural language

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Refining the Design of a Contracting Finite-State Dependency Parser

Prediction of Maximal Projection for Semantic Role Labeling

An Efficient Implementation of a New POP Model

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

Proof Theory for Syntacticians

The Effect of Multiple Grammatical Errors on Processing Non-Native Writing

What Can Neural Networks Teach us about Language? Graham Neubig a2-dlearn 11/18/2017

Natural Language Processing. George Konidaris

Experiments with a Higher-Order Projective Dependency Parser

MODELING DEPENDENCY GRAMMAR WITH RESTRICTED CONSTRAINTS. Ingo Schröder Wolfgang Menzel Kilian Foth Michael Schulz * Résumé - Abstract

Compositional Semantics

Developing a TT-MCTAG for German with an RCG-based Parser

Accurate Unlexicalized Parsing for Modern Hebrew

The Interface between Phrasal and Functional Constraints

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

Minimalism is the name of the predominant approach in generative linguistics today. It was first

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

Analysis of Probabilistic Parsing in NLP

An Interactive Intelligent Language Tutor Over The Internet

Type Theory and Universal Grammar

Argument structure and theta roles

Construction Grammar. University of Jena.

A Computational Evaluation of Case-Assignment Algorithms

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Words come in categories

Character Stream Parsing of Mixed-lingual Text

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Feature-Based Grammar

Ch VI- SENTENCE PATTERNS.

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

On the Notion Determiner

The Smart/Empire TIPSTER IR System

Underlying and Surface Grammatical Relations in Greek consider

Inleiding Taalkunde. Docent: Paola Monachesi. Blok 4, 2001/ Syntax 2. 2 Phrases and constituent structure 2. 3 A minigrammar of Italian 3

A deep architecture for non-projective dependency parsing

Adapting Stochastic Output for Rule-Based Semantics

Cross Language Information Retrieval

Achim Stein: Diachronic Corpora Aston Corpus Summer School 2011

Taught Throughout the Year Foundational Skills Reading Writing Language RF.1.2 Demonstrate understanding of spoken words,

Theoretical Syntax Winter Answers to practice problems

Pseudo-Passives as Adjectival Passives

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

BULATS A2 WORDLIST 2

LING 329 : MORPHOLOGY

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Learning Computational Grammars

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

First Grade Curriculum Highlights: In alignment with the Common Core Standards

Second Exam: Natural Language Parsing with Neural Networks

AQUA: An Ontology-Driven Question Answering System

Part III: Semantics. Notes on Natural Language Processing. Chia-Ping Chen

Constraining X-Bar: Theta Theory

Chapter 9 Banked gap-filling

The building blocks of HPSG grammars. Head-Driven Phrase Structure Grammar (HPSG) HPSG grammars from a linguistic perspective

Copyright and moral rights for this thesis are retained by the author

Using dialogue context to improve parsing performance in dialogue systems

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

RANKING AND UNRANKING LEFT SZILARD LANGUAGES. Erkki Mäkinen DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF TAMPERE REPORT A ER E P S I M S

Control and Boundedness

Sample Goals and Benchmarks

Three New Probabilistic Models. Jason M. Eisner. CIS Department, University of Pennsylvania. 200 S. 33rd St., Philadelphia, PA , USA

Can Human Verb Associations help identify Salient Features for Semantic Verb Classification?

Hyperedge Replacement and Nonprojective Dependency Structures

Houghton Mifflin Reading Correlation to the Common Core Standards for English Language Arts (Grade1)

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

LTAG-spinal and the Treebank

BASIC ENGLISH. Book GRAMMAR

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Survey on parsing three dependency representations for English

What the National Curriculum requires in reading at Y5 and Y6

The College Board Redesigned SAT Grade 12

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Multiple case assignment and the English pseudo-passive *

Advanced Grammar in Use

Towards a MWE-driven A* parsing with LTAGs [WG2,WG3]

CX 101/201/301 Latin Language and Literature 2015/16

Transcription:

Dependency grammar and dependency parsing Syntactic analysis (5LN455) 2012-12-11 Joakim Nivre Department of Linguistics and Philology Based on slides by Marco Kuhlmann

Dependency grammar

Dependency grammar The term dependency grammar does not refer to a specific grammar formalism. Rather, it refers to a specific way to describe the syntactic structure of a sentence.

Dependency grammar The notion of dependency The basic observation behind constituency is that groups of words may act as one unit. Example: noun phrase, prepositional phrase The basic observation behind dependency is that words have grammatical functions with respect to other words in the sentence. Example: subject, modifier

Dependency grammar Phrase structure trees S NP VP Pro Verb NP I booked Det Nom a Nom PP Noun from LA flight

Dependency grammar Dependency trees! dobj! subj det pmod I! booked a flight from LA! In an arc h d, the word h is called the head, and the word d is called the dependent. The arcs form a rooted tree.

Dependency grammar Heads in phrase structure grammar In phrase structure grammar, ideas from dependency grammar can be found in the notion of heads. Roughly speaking, the head of a phrase is the most important word of the phrase: the word that determines the phrase function. Examples: noun in a noun phrase, preposition in a prepositional phrase

Dependency grammar Heads in phrase structure grammar S NP VP Pro Verb NP I booked Det Nom a Nom PP Noun from LA flight

Dependency grammar The history of dependency grammar The notion of dependency can be found in some of the earliest formal grammars. Modern dependency grammar is attributed to Lucien Tesnière (1893 1954). Recent years have seen a revived interest in dependency-based description of natural language syntax.

Dependency grammar Head-dependency relations Verb + arguments Subject: Sandy writes poetry Object: Sandy writes poetry Noun + modifiers Determiner: the little black cat Adjectival modifier: the little black cat

Dependency grammar Some tricky cases Coordination Sandy and Kim write poetry Verb groups Sandy could have written poetry Prepositional phrases Sandy went to London

Dependency grammar Examples What dependency relations do you find in the following sentences? Her mother sent her a letter. Economic news had little effect on financial markets.

Dependency grammar Linguistic resources Descriptive dependency grammars exist for some natural languages. Dependency treebanks exist for a wide range of natural languages. These treebanks can be used to train accurate and efficient dependency parsers.

Overview Arc-factored dependency parsing Collins algorithm Eisner s algorithm Transition-based dependency parsing The arc-standard algorithm Evaluation of dependency parsers

Arc-factored dependency parsing

Ambiguity Just like phrase structure parsing, dependency parsing has to deal with ambiguity. dobj subj det pmod I booked a flight from LA

Ambiguity Just like phrase structure parsing, dependency parsing has to deal with ambiguity. dobj pmod subj det I booked a flight from LA

Disambiguation We need to disambiguate between alternative analyses. We develop mechanisms for scoring dependency trees, and disambiguate by choosing a dependency tree with the highest score.

Scoring models and parsing algorithms Distinguish two aspects: Scoring model: How do we want to score dependency trees? Parsing algorithm: How do we compute a highest-scoring dependency tree under the given scoring model?

The arc-factored model Split the dependency tree t into parts p1,..., pn, score each of the parts individually, and combine the score into a simple sum. score(t) = score(p1) + + score(pn) The simplest scoring model is the arc-factored model, where the scored parts are the arcs of the tree.

Arc-factored dependency parsing Features! dobj! subj det pmod I! booked a flight from LA! To score an arc, we define features that are likely to be relevant in the context of parsing. We represent an arc by its feature vector.

Arc-factored dependency parsing Examples of features The head is a verb. The dependent is a noun. The head is a verb and the dependent is a noun. The head is a verb and the predecessor of the head is a pronoun. The arc goes from left to right. The arc has length 2.

Arc-factored dependency parsing Feature vectors 1 booked flight Feature: The dependent is a noun. 0 flight from LA flight a booked I 0 Feature: The head is a verb. 1

Arc-factored dependency parsing Implementation of feature vectors We assign each feature a unique number. For each arc, we collect the numbers of those features that apply to that arc. The feature vector of the arc is the list of those numbers. Example: [1, 2, 42, 313, 1977, 2008, 2010]

Arc-factored dependency parsing Feature weights Arc-factored dependency parsers require a training phase. During training, our goal is to assign, to each feature fi, a feature weight wi. Intuitively, the weight wi quantifies the effect of the feature fi on the likelihood of the arc. How likely is is that we will see an arc with this feature in a useful dependency tree?

Arc-factored dependency parsing Feature weights We define the score of an arc h d as the weighted sum of all features of that arc: score(h d) = f1w1 + + fnwn

Arc-factored dependency parsing Training using structured prediction Take a sentence w and a gold-standard dependency tree g for w. Compute the highest-scoring dependency tree under the current weights; call it p. Increase the weights of all features that are in g but not in p. Decrease the weights of all features that are in p but not in g.

Arc-factored dependency parsing Training using structured prediction Training involves repeatedly parsing (treebank) sentences and refining the weights. Hence, training presupposes an efficient parsing algorithm. Next time we will look at parsing algorithms for the arc-factored model.

Arc-factored dependency parsing Higher-order models The arc-factored model is a first-order model, because scored subgraphs consist of a single arc. An nth-order model scores subgraphs consisting of (at most) n arcs. Second-order: siblings, grand-parents Third-order: tri-siblings, grand-siblings Higher-order models capture more linguistic structure and give higher parsing accuracy.

Arc-factored dependency parsing Summary The term arc-factored dependency parsing refers to dependency parsers that score a dependency tree by scoring its arcs. Arcs are scored by defining features and assigning weights to these features. The resulting parsers can be trained using structured prediction. More powerful scoring models exist.