Constructions at work or at rest?

Similar documents
have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Accurate Unlexicalized Parsing for Modern Hebrew

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Some Principles of Automated Natural Language Information Extraction

Towards a MWE-driven A* parsing with LTAGs [WG2,WG3]

LFG Semantics via Constraints

CS 598 Natural Language Processing

Proof Theory for Syntacticians

An Efficient Implementation of a New POP Model

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

An Interactive Intelligent Language Tutor Over The Internet

Parsing of part-of-speech tagged Assamese Texts

"f TOPIC =T COMP COMP... OBJ

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

UNDERSTANDING DECISION-MAKING IN RUGBY By. Dave Hadfield Sport Psychologist & Coaching Consultant Wellington and Hurricanes Rugby.

Compositional Semantics

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

A Case Study: News Classification Based on Term Frequency

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

LEARNING A SEMANTIC PARSER FROM SPOKEN UTTERANCES. Judith Gaspers and Philipp Cimiano

Learning Computational Grammars

AQUA: An Ontology-Driven Question Answering System

Adapting Stochastic Output for Rule-Based Semantics

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

FOREWORD.. 5 THE PROPER RUSSIAN PRONUNCIATION. 8. УРОК (Unit) УРОК (Unit) УРОК (Unit) УРОК (Unit) 4 80.

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

Derivational and Inflectional Morphemes in Pak-Pak Language

LINGUISTICS. Learning Outcomes (Graduate) Learning Outcomes (Undergraduate) Graduate Programs in Linguistics. Bachelor of Arts in Linguistics

Constraining X-Bar: Theta Theory

AN INTRODUCTION (2 ND ED.) (LONDON, BLOOMSBURY ACADEMIC PP. VI, 282)

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

Applications of memory-based natural language processing

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Testing claims of a usage-based phonology with Liverpool English t-to-r 1

A relational approach to translation

Detecting English-French Cognates Using Orthographic Edit Distance

Possessive have and (have) got in New Zealand English Heidi Quinn, University of Canterbury, New Zealand

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

Using dialogue context to improve parsing performance in dialogue systems

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Procedia - Social and Behavioral Sciences 154 ( 2014 )

Authors note Chapter One Why Simpler Syntax? 1.1. Different notions of simplicity

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

The Interface between Phrasal and Functional Constraints

arxiv: v1 [cs.cl] 2 Apr 2017

Linking Task: Identifying authors and book titles in verbose queries

Ensemble Technique Utilization for Indonesian Dependency Parser

Parallel Evaluation in Stratal OT * Adam Baker University of Arizona

CEFR Overall Illustrative English Proficiency Scales

Problems of the Arabic OCR: New Attitudes

Construction Grammar. University of Jena.

cambridge occasional papers in linguistics Volume 8, Article 3: 41 55, 2015 ISSN

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

The Strong Minimalist Thesis and Bounded Optimality

Building an HPSG-based Indonesian Resource Grammar (INDRA)

LTAG-spinal and the Treebank

English Language and Applied Linguistics. Module Descriptions 2017/18

Control and Boundedness

LING 329 : MORPHOLOGY

Developing a TT-MCTAG for German with an RCG-based Parser

Discourse markers and grammaticalization

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Context Free Grammars. Many slides from Michael Collins

Lexical category induction using lexically-specific templates

Phonological encoding in speech production

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

A Version Space Approach to Learning Context-free Grammars

Rule Learning With Negation: Issues Regarding Effectiveness

Linguistics Program Outcomes Assessment 2012

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

Mandarin Lexical Tone Recognition: The Gating Paradigm

Interfacing Phonology with LFG

The College Board Redesigned SAT Grade 12

Iraide Ibarretxe Antuñano Universidad de Zaragoza

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

On the Notion Determiner

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Underlying and Surface Grammatical Relations in Greek consider

Prediction of Maximal Projection for Semantic Role Labeling

California Department of Education English Language Development Standards for Grade 8

Feature-Based Grammar

BYLINE [Heng Ji, Computer Science Department, New York University,

PH.D. IN COMPUTER SCIENCE PROGRAM (POST M.S.)

Using computational modeling in language acquisition research

1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

Multiple case assignment and the English pseudo-passive *

Jacqueline C. Kowtko, Patti J. Price Speech Research Program, SRI International, Menlo Park, CA 94025

A Framework for Customizable Generation of Hypertext Presentations

L1 and L2 acquisition. Holger Diessel

South Carolina English Language Arts

Today we examine the distribution of infinitival clauses, which can be

Transcription:

Constructions at work or at rest? RENS BOD* Abstract We question whether Adele Goldberg fulfills her self-declared goal in Constructions at Work, i.e. to develop a usage-based theory that can produce an open-ended number of novel utterances based on a finite amount of input. We point out converging trends in computational linguistics that suggest formalizations of Construction Grammar. In particular, we go into recent developments in Data-Oriented Parsing, such as U-DOP and LFG-DOP, that produce an unlimited number of new utterances based on a finite number of stored form-meaning pairs. Keywords: computational linguistics; data-oriented parsing; U-DOP, LFG-DOP; probabilistic linguistics; language acquisition; language processing. Constructions at Work is an impressive piece of work which has already become a landmark in linguistics. In my commentary I will not go into the wealth of linguistic insights presented by Adele Goldberg, nor will I question the basic assumptions of Construction Grammar (CxG). Instead, I will deal with only one question: to what extent are we o ered a model of language acquisition and/or language use in Goldberg s book? At various places in the book we read that one of Goldberg s major goals is to address How [do] learners acquire generalizations such that they can produce an open-ended number of novel utterances based on a * University of St Andrews, Scotland, and University of Amsterdam, Netherlands. Author s e-mail: 3rens.bod@gmail.com4. Many thanks to Willem Zuidema and an anonymous reviewer for comments on a previous version of this commentary. Cognitive Linguistics 20 1 (2009), 129 134 DOI 10.1515/COGL.2009.006 0936 5907/09/0020 0129 6 Walter de Gruyter

130 R. Bod finite amount of input (e.g., p. 11, p. 227). But at the end of the day we learn that Goldberg proposes no model or theory that can produce an open-ended number of novel utterances based on a finite amount of input. Sure enough, we get a lot of information about how constructions can be learned and generalized, and how a well-known machine learning algorithm, ADABoost, can be applied to combine multiple cues in generalizing constructions. But suprisingly enough, there is no exact model or theory in the book that describes how these constructions are used by learners to produce novel utterances. The question that continuously came up while reading Goldberg s book, was: where is the model? I could find no description of an inputoutput procedure nor of a process-model that tells us how new utterances are produced on the basis of previous utterances. There is no precise definition of (i) the notion of a productive unit in CxG, (ii) the way productive units are acquired step by step from incoming input utterances, and (iii) the combination operations that combine constructions into (an open-ended number of ) new utterances. At the end of the book, it even remained unclear to me what kind of an object a construction is according to Goldberg. Of course, I understand (and agree) that any linguistic pattern is recognized as a construction [...] (p. 5). Yet, the concept of any linguistic pattern is not well-defined. Linguistic patterns can be ( partial) strings, (partial) phrase-structures, (partial) dependency structures, (partial) attribute-value matrices, ( partial) typed-feature structures, etc. I doubt whether these are all taken as constructions. However, there is also a more precise definition of construction in the book: a formmeaning pairing. But then, is a construction a pairing of strings, a function from a string to a predicate-argument structure, or yet something else? And is the combination operation between constructions a concatenation operation, a substitution operation, a unification operation, some integration of these three, or something di erent? We learn on page 10 that Constructions are combined freely to form actual expressions as long as they are not in conflict. As a first informal exposition this may do, but it is not a definition of how two or more constructions can be combined. I guess that at least a notion of substitution is involved if a construction with open slots or variables is combined with another construction. And most importantly perhaps, how are the productive units acquired from previous utterances in a step-by-step, incremental way? My comments should not be seen as a pedantry: as long as we have no precise definition of the combination operation(s) between CxG s productive units, there is no way that CxG can be unequivocally tested as a linguistic model. I am not arguing for mathematical formalizations, only for

Constructions at work or rest? 131 precise definitions. Even when definitive definitions are out of reach at the moment, tentative definitions allow at least for testing CxG, after which these definitions can be adapted and improved. One could claim that it should not be a theorist s goal to come up with precise definitions for combination operations let alone to construct a computational model. While I agree that the construction of a full-fledged computational model of language acquisition may be outside the realm of language theory, precise definitions of its productive units and how they are combined form a necessary condition for any linguistic theory. Is this where Construction Grammar is moving to: a set of insightful but untestable ideas? Yet, it is not just Construction Grammar but almost any current linguistic theory that has given up on the construction of a precise, testable model of language use and language acquisition. Where are the good old days when every self-respecting linguist designed a precise model of language use and acquisition? But perhaps this is wishful thinking, and these days have never existed. Although I believe that Construction Grammar and Usage-Based Linguistics are by far the most promising approaches to language learning, they seem to underestimate the importance of exact definitions. As far as I am aware, the only discipline that takes this enterprise seriously, is computational linguistics. However, during the last decade or so this discipline has become so impenetrable for non-computational linguists that it has lost almost all connections with the other fields of linguistics. The fact that computational linguists mostly publish in conference proceedings has not helped either, although all proceedings from 1965 onwards are freely available at the ACL-archives ( http://acl.ldc. upenn.edu/). It is regrettable although perhaps understandable that Goldberg does not go into any of the work on statistical modeling in computational linguistics, which is so relevant for usage-based and constructionist models of language (given the importance of frequency in these models). Other linguists, such as Bybee (2006) and Hay and Bresnan (2006), have long recognized and emphasized the converging trends in theoretical and computational linguistics, suggesting a common development towards an integrated usage-based theory of language use and acquisition. In computational linguistics, there is nowadays a whole tradition that aims to combine construction/usage-based and computational approaches, such as Klein and Manning (2002), Steels (2004), Bod (2006), Zuidema (2006) and several others. To give a concrete example let me briefly point out some connections between CxG and a computational model I have worked on myself: data-oriented parsing (DOP, hereafter). I will only go into the more recent versions of DOP such as LFG-DOP (Bod and Kaplan 1998) and

132 R. Bod HPSG-DOP (Arnold and Linardaki 2007). In DOP, new utterances are produced and understood by statistically generalizing over a corpus of previous utterances. Informally, the productive units in DOP are, like in CxG, taken as any linguistic pattern be they morphological, syntactic, semantic or pragmatic in nature. But in concrete instantiations of DOP the productive units (or constructions) are usually taken as syntacticsemantic patterns. For example, in LFG-DOP the productive units are defined as any connected part of constituent structure and semantic structure (see Bod 2006). These parts are learned from exemplars of previous utterance-analyses and are formally equivalent to a functional mapping from subtrees (representing syntactic surface form) to attribute-value matrices (representing functional and semantic structure). The combination operation between the productive units consists of simple node substitution between subtrees combined with unification between corresponding attribute-value matrices. A statistical model is used to rank alternative utterances for a given meaning or alternative meanings for a given utterance. All these formal details of DOP may turn out to be inadequate in the long run, but they are precise enough to be turned into a testable computational model, which has indeed been carried out (e.g., Bod in press; Zuidema 2006). The acquisition of the productive units is dealt with by an unsupervised generalization of DOP known as U-DOP (Bod 2007). U-DOP initially allows any, possibly discontinuous fragment of previous utterances to form a productive unit, and lets the statistics decide which form-meaning pairs are actually learned and stored in the corpus (to which next DOP can be applied). This integrated U-DOP/DOP model has been tested against a wide variety of child language phenomena from the CHILDES database, ranging from learning separable particle verbs to learning auxiliary fronting (Bod in press). It o ers to linguists and cognitive scientists a precise model for testing the usage-based approach against child and adult data. It thus seems that the DOP approach is not only consonant with CxG but that it actually stands as a testable realization of CxG: it specifies how constructions are acquired, how they are combined and how they can undergo changes in the light of new input. Goldberg s notion of exemplar-based abstraction (p. 48) is furthermore congenial to U-DOP s acquisition procedure that abstracts over di erent utterances by statistical induction (Bod in press). Yet U-DOP/DOP is not the only possible computational realization of CxG. There are other computational approaches to language learning, based on somewhat di erent notions of productive unit and di erent combination operations, e.g., van Zaanen (2000), Clark (2001), Klein and Manning (2002), Dennis (2005), Zuidema (2006), Seginer (2007) to name a few.

Constructions at work or rest? 133 Despite the various, sometimes subtle di erences between these computational learning approaches, they share one very important goal, which is, as far as I understand, the same goal as in CxG: to develop a usagebased theory of language learning that acquires generalizations which can produce an unlimited number of new utterances based on a finite number of previous utterances. I am convinced that this goal can be achieved, if only the two linguistic and computational strands take each other s work into account. Received 23 October 2007 Revision received 22 March 2008 University of St Andrews and University of Amsterdam References Arnold, Doug and Evita Linardaki 2007 HPSG-DOP: Towards exemplar-based HPSG. Proceedings of the Workshop on Exemplar Based Models of Language Acquisition and Use. Dublin, 42 51. Bod, Rens 2006 Exemplar-based syntax: How to get productivity from examples. The Linguistic Review 23, 291 320. Bod, Rens 2007 Is the end of supervised parsing in sight? Proceedings ACL 2007. Prague, 400 407. Bod, Rens in press From exemplar to grammar: A probabilistic analogy-based model of language learning. Cognitive Science. Bod, Rens and Ronald Kaplan 1998 A probabilistic corpus-driven model for lexical-functional analysis. Proceedings COLING-AC 1998, 145 151. Bybee, Joan 2006 From usage to grammar: The mind s response to repetition. Language 82(4), 711 733. Clark, Alexander 2001 Unsupervised induction of stochastic context-free grammars using distributional clustering. Proceedings CoNLL 2001, 105 112. Dennis, Simon 2005 An exemplar-based approach to unsupervised parsing. Proceedings CogSci 2005. Stresa, Italy. Hay, Jennifer and Joan Bresnan 2006 Spoken syntax: The phonetics of giving a hand in new zealand english. The Linguistic Review 23, 321 349. Klein, Dan and Chris Manning 2002 A general constituent-context model for improved grammar induction. Proceedings ACL 2002. Philadelphia, 128 135. Seginer, Yoav 2007 Fast unsupervised incremental parsing. Proceedings ACL 2007, 384 391.

134 R. Bod Steels, Luc 2004 Constructivist development of grounded Construction Grammar. Proceedings ACL 2004, 9 16. van Zaanen, Menno 2000 ABL: Alignment-Based Learning. Proceedings COLING 2000, 961 967. Zuidema, Willem 2006 What are the productive units of natural language grammar? A DOP approach to the automatic identification of constructions. Proceedings CoNLL 2006, 29 36.