Alternative Syntactic Theories

Similar documents
CS 598 Natural Language Processing

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Grammars & Parsing, Part 1:

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

"f TOPIC =T COMP COMP... OBJ

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Proof Theory for Syntacticians

Natural Language Processing. George Konidaris

Developing a TT-MCTAG for German with an RCG-based Parser

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Compositional Semantics

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Control and Boundedness

Parsing of part-of-speech tagged Assamese Texts

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

An Introduction to the Minimalist Program

An Interactive Intelligent Language Tutor Over The Internet

1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

Context Free Grammars. Many slides from Michael Collins

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Some Principles of Automated Natural Language Information Extraction

Dependency, licensing and the nature of grammatical relations *

Constraining X-Bar: Theta Theory

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

Towards a MWE-driven A* parsing with LTAGs [WG2,WG3]

Underlying and Surface Grammatical Relations in Greek consider

Chapter 4: Valence & Agreement CSLI Publications

LTAG-spinal and the Treebank

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Construction Grammar. University of Jena.

LING 329 : MORPHOLOGY

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Accurate Unlexicalized Parsing for Modern Hebrew

Hyperedge Replacement and Nonprojective Dependency Structures

Prediction of Maximal Projection for Semantic Role Labeling

Aspectual Classes of Verb Phrases

Inleiding Taalkunde. Docent: Paola Monachesi. Blok 4, 2001/ Syntax 2. 2 Phrases and constituent structure 2. 3 A minigrammar of Italian 3

Efficient Normal-Form Parsing for Combinatory Categorial Grammar

Authors note Chapter One Why Simpler Syntax? 1.1. Different notions of simplicity

CAS LX 522 Syntax I. Long-distance wh-movement. Long distance wh-movement. Islands. Islands. Locality. NP Sea. NP Sea

Heads and history NIGEL VINCENT & KERSTI BÖRJARS The University of Manchester

Type-driven semantic interpretation and feature dependencies in R-LFG

AQUA: An Ontology-Driven Question Answering System

Hans-Ulrich Block, Hans Haugeneder Siemens AG, MOnchen ZT ZTI INF W. Germany. (2) [S' [NP who][s does he try to find [NP e]]s IS' $=~

RANKING AND UNRANKING LEFT SZILARD LANGUAGES. Erkki Mäkinen DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF TAMPERE REPORT A ER E P S I M S

The Strong Minimalist Thesis and Bounded Optimality

Analysis of Probabilistic Parsing in NLP

Multiple case assignment and the English pseudo-passive *

Adapting Stochastic Output for Rule-Based Semantics

THE INTERNATIONAL JOURNAL OF HUMANITIES & SOCIAL STUDIES

Language properties and Grammar of Parallel and Series Parallel Languages

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Argument structure and theta roles

Update on Soar-based language processing

Refining the Design of a Contracting Finite-State Dependency Parser

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

The building blocks of HPSG grammars. Head-Driven Phrase Structure Grammar (HPSG) HPSG grammars from a linguistic perspective

A Version Space Approach to Learning Context-free Grammars

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Erkki Mäkinen State change languages as homomorphic images of Szilard languages

The Interface between Phrasal and Functional Constraints

Loughton School s curriculum evening. 28 th February 2017

Theoretical Syntax Winter Answers to practice problems

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

Interfacing Phonology with LFG

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Parsing with Treebank Grammars: Empirical Bounds, Theoretical Models, and the Structure of the Penn Treebank

Building an HPSG-based Indonesian Resource Grammar (INDRA)

Using dialogue context to improve parsing performance in dialogue systems

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Specifying Logic Programs in Controlled Natural Language

A Grammar for Battle Management Language

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

PHILOSOPHY & CULTURE Syllabus

On the Notion Determiner

Which verb classes and why? Research questions: Semantic Basis Hypothesis (SBH) What verb classes? Why the truth of the SBH matters

Teacher: Mlle PERCHE Maeva High School: Lycée Charles Poncet, Cluses (74) Level: Seconde i.e year old students

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

The Structure of Multiple Complements to V

The MEANING Multilingual Central Repository

(Sub)Gradient Descent

a) analyse sentences, so you know what s going on and how to use that information to help you find the answer.

SEMAFOR: Frame Argument Resolution with Log-Linear Models

Constructions with Lexical Integrity *

Som and Optimality Theory

Minimalism is the name of the predominant approach in generative linguistics today. It was first

Implementing a tool to Support KAOS-Beta Process Model Using EPF

Grade 11 Language Arts (2 Semester Course) CURRICULUM. Course Description ENGLISH 11 (2 Semester Course) Duration: 2 Semesters Prerequisite: None

College Pricing. Ben Johnson. April 30, Abstract. Colleges in the United States price discriminate based on student characteristics

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Procedia - Social and Behavioral Sciences 154 ( 2014 )

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Linking Task: Identifying authors and book titles in verbose queries

How to analyze visual narratives: A tutorial in Visual Narrative Grammar

Part III: Semantics. Notes on Natural Language Processing. Chia-Ping Chen

Transcription:

Alternative Syntactic Theories L614 Spring 2015

Syntactic analysis Generative grammar: collection of words and rules with which we generate strings of those words, i.e., sentences Syntax attempts to capture the nature of those rules 1 Colorless green ideas sleep furiously. 2 *Furiously sleep ideas green colorless. What generalizations are needed to capture the difference between grammatical sentences and ungrammatical sentences? Using a particular formalism, a theory encapsulates these generalizations i.e., a theory is a grammar

Formalism vs. Theory Will we actually look at theories?... Sort of. A theory describes a set of data and makes predictions for new data In this class, we will emphasize theories which are testable, i.e., can be verified or falsified A formalism provides a way of defining a theory with mathematical rigor It is essentially a set of beliefs and conditions that frame how generalizations can be made. The course name (Alternative Syntactic Theories) is a bit of a misnomer: we will actually be focusing on formalisms, and we will use theories to exemplify them.

Transformational syntax The Transformational tradition Roughly speaking, transformational syntax (GB, P&P,...) has focused on the following: Explanatory adequacy: does the theory fit with a deeper model (e.g., universal grammar)? Psychological modeling: does the grammar make sense in light of what we know of how the mind works? Universality: are the generalizations applicable to all languages? Transformations/Movement: are (surface) sentences derived from underlying sentences? (e.g. passives from actives) These kinds of theories have not generally been integrated with computational applications

Alternatives Making it computational How can grammatical theories be useful for computational lingusitics? Parsing: take an input sentence and return the syntactic analysis and/or state whether it is a valid sentence Generation: take a meaning representation and generate a valid sentence Both tasks are often subparts of practical applications (e.g., dialogue systems) Both can also provide feedback to the grammar writer

Alternatives Computational needs To use a grammar for parsing or generation, we need to have a grammar that meets several criteria: Accurate: gives a correct analysis Precise: tells a computer exactly what to do Efficient: able to parse a sentence and return one or only a small number of parses Useful: is relatively easy to map a syntactic structure of a sentence to its meaning Not necessarily why computational formalisms were developed, but the formalisms enable such uses

Computational Grammar Formalisms The formalisms we will look generally share several properties: Descriptively adequate Precisely encoded (implementable) Constrained in the mathematical formalism Monostratal (Usually) highly lexical

Descriptive Descriptively adequate One could explain the underlying mechanisms, but we are mostly concerned with being able to describe linguistic phenomena Provide a structural description for every well-formed sentence Define which sentences are well-formed in a language & which not Give an accurate encoding of a language Broad-coverage: describe all of a language Less of a distinction between core & periphery phenomena

Precise Precisely encoded Mathematical formalism: formal way to generate sets of strings Thus, we need to precisely define: elementary structures ways of combining those structures Such an emphasis on mathematical precision makes these grammar formalisms more easily implementable e.g., can answer the question of whether different parts of a grammar will conflict

Constrained Constrained in the mathematical formalism Formalism should (arguably) be constrained, i.e., cannot be allowed to specify all strings Linguistic motivation: Limit the scope of the theory of grammar Computational motivation: Allow one to define efficient processing models This is different than constraining a theory What is the minimum amount of mathematical overhead that we need to describe language?

Monostratal Monostratal Only have one (surface) syntactic level Make no recourse to movement or transformations Augment your basic (phrase structure) tree with information that can describe movement phenomena Need some way to relate different structures (e.g., active and passive) without invoking, e.g., traces Without having to refer to movement, easier to process sentences computationally

Lexical Lexical Some approaches: rules apply to broad classes & only some information in the lexicon (e.g., subcategorization) But more and more theories emphasize the role of individual lexical items in grammatical constructions Linguistic motivation: lexicon best way to specify some generalizations: He told/*divulged me the truth Computational motivation: lexical information can be derived from corpora Shift more of the information to the lexicon; each lexical item is thus a complex object

Brief mention of complexity We have touched on the complexity of different formalisms Type Automaton Grammar Memory Name Rule Name 0 Unbounded TM α β General rewrite 1 Bounded LBA β A γ β δ γ Context-sensitive 2 Stack PDA A β Context-free 3 None FSA A xb, A x Right linear TM: Turing Machine LBA: Linear-Bounded Automaton PDA: Push-Down Automaton FSA: Finite-State Automaton

Criteria Criteria under which to evaluate grammar formalisms Three kinds of criteria: linguistic naturalness mathematical power computational effectiveness and efficiency The weaker the type of grammar: the stronger the claim made about possible languages the greater the potential efficiency of the parsing procedure Reasons for choosing a stronger grammar class: to capture the empirical reality of actual languages to provide for elegant analyses capturing more generalizations (e.g., more compact grammars)

CFGs Context-Free Grammars (CFGs) Context-Free Grammars (CFGs): probably the most popular formalism for writing English grammars elementary structures: rules composed of nonterminal and terminal elements combine rules by rewriting them Example of a set of rules: S NP VP NP Det N VP V NP... Empirical downside: the rules are rather impoverished...

CFGs Are CFGs good enough? Data from Swiss German & other languages show that CFGs are not powerful enough to handle all natural language constructions CFGs are not easily lexicalized CFGs become complicated once we start taking into account agreement features, verb subcategorizations, unbounded dependency constructions, raising constructions, etc. We need more refined formalisms...

Beyond CFGs Beyond CFGs We want to move beyond CFGs to better capture language, but maintain that level of precision One can view this in different ways: Extend the basic model of CFGS with, e.g., complex categories, functional structure, feature structures,... Eliminate CFG model or derive it some other way The frameworks we will investigate explore different ways of looking at syntax...

Computational Grammar Frameworks What we will look at the rest of the semester: Dependency Grammar (DG) Tree-Adjoining Grammar (TAG) Lexical-Functional Grammar (LFG) Head-driven Phrase Structure Grammar (HPSG) Combinatory Categorial Grammar (CCG)

Dependency Grammar (DG) The way to analyze a sentence is by looking at the relations between words Generally speaking, no grouping (constituency) is used DG is not a unified framework; there are a host of different frameworks within this tradition DG bears similarity to functional structure, but have often been derived independent of CFG traditions Analyses tend to be closely related to the semantics of a sentence Some frameworks we ll investigate utilize insights from DG

Tree-Adjoining Grammar (TAG) Roughly: analysis looks like a CFG tree, but the way to obtain it is different Elementary structures are trees of arbitrary height Trees are rooted in lexical items, i.e. lexicalized In other words, the lexicon contains tree fragments as parts of lexical entries Put trees together by substituting and adjoining them, resulting in a final tree which looks like a CFG-derived tree

Lexical-Functional Grammar (LFG) Functional structure (subject, object, etc.) divided from constituent structure (tree structure) Akin to dependency structure + phrase structure The f-structures are potentially very complex Can express some generalizations in f-structure; some in c-structure; i.e., not restricted to saying everything in terms of trees

Head-driven Phrase Structure Grammar (HPSG) Sentences, phrases, & words all uniformly treated as linguistic signs, i.e., complex objects of features Analyses can rely on CFG backbone, but need not Similar to LFG in its use of a feature architecture Uses inheritance hierarchy to relate different objects e.g., nouns and determiners are both types of nominals

Combinatory Categorial Grammar (CCG) Categorial Grammar derives sentences in a proof-solving manner Maintains close link with semantic representation Lexical categories specify how to combine words into sentences Again, lexical entries contain tree-like information CCG has sophisticated mechnisms to deal with coordination, extraction, & other constructions