A Usage-Based Approach to Recursion in Sentence Processing

Similar documents
A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

The Perception of Nasalized Vowels in American English: An Investigation of On-line Use of Vowel Nasalization in Lexical Access

An Empirical and Computational Test of Linguistic Relativity

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Morphosyntactic and Referential Cues to the Identification of Generic Statements

Control and Boundedness

CS 598 Natural Language Processing

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Mandarin Lexical Tone Recognition: The Gating Paradigm

Loughton School s curriculum evening. 28 th February 2017

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

Aging and the Use of Context in Ambiguity Resolution: Complex Changes From Simple Slowing

Phonological and Phonetic Representations: The Case of Neutralization

Phenomena of gender attraction in Polish *

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

Minimalism is the name of the predominant approach in generative linguistics today. It was first

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

An Evaluation of the Interactive-Activation Model Using Masked Partial-Word Priming. Jason R. Perry. University of Western Ontario. Stephen J.

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Linking object names and object categories: Words (but not tones) facilitate object categorization in 6- and 12-month-olds

Advanced Grammar in Use

Degeneracy results in canalisation of language structure: A computational model of word learning

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

ELD CELDT 5 EDGE Level C Curriculum Guide LANGUAGE DEVELOPMENT VOCABULARY COMMON WRITING PROJECT. ToolKit

The Effect of Extensive Reading on Developing the Grammatical. Accuracy of the EFL Freshmen at Al Al-Bayt University

The Strong Minimalist Thesis and Bounded Optimality

Proof Theory for Syntacticians

Using dialogue context to improve parsing performance in dialogue systems

Constraining X-Bar: Theta Theory

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

FOREWORD.. 5 THE PROPER RUSSIAN PRONUNCIATION. 8. УРОК (Unit) УРОК (Unit) УРОК (Unit) УРОК (Unit) 4 80.

Syntactic systematicity in sentence processing with a recurrent self-organizing network

The College Board Redesigned SAT Grade 12

1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

Age Effects on Syntactic Control in. Second Language Learning

Underlying and Surface Grammatical Relations in Greek consider

BASIC ENGLISH. Book GRAMMAR

Derivational and Inflectional Morphemes in Pak-Pak Language

Developing Grammar in Context

An ICT environment to assess and support students mathematical problem-solving performance in non-routine puzzle-like word problems

On the Notion Determiner

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Context Free Grammars. Many slides from Michael Collins

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Evolution of Symbolisation in Chimpanzees and Neural Nets

Which verb classes and why? Research questions: Semantic Basis Hypothesis (SBH) What verb classes? Why the truth of the SBH matters

Author: Justyna Kowalczys Stowarzyszenie Angielski w Medycynie (PL) Feb 2015

Good Enough Language Processing: A Satisficing Approach

Written by: YULI AMRIA (RRA1B210085) ABSTRACT. Key words: ability, possessive pronouns, and possessive adjectives INTRODUCTION

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

BULATS A2 WORDLIST 2

Writing a composition

AN EXPERIMENTAL APPROACH TO NEW AND OLD INFORMATION IN TURKISH LOCATIVES AND EXISTENTIALS

Word learning as Bayesian inference

Construction Grammar. University of Jena.

Senior Stenographer / Senior Typist Series (including equivalent Secretary titles)

An Introduction to the Minimalist Program

The Effect of Written Corrective Feedback on the Accuracy of English Article Usage in L2 Writing

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Probabilistic principles in unsupervised learning of visual structure: human data and a model

Evidence for Reliability, Validity and Learning Effectiveness

Correspondence between the DRDP (2015) and the California Preschool Learning Foundations. Foundations (PLF) in Language and Literacy

Emmaus Lutheran School English Language Arts Curriculum

Good-Enough Representations in Language Comprehension

Using a Native Language Reference Grammar as a Language Learning Tool

ROSETTA STONE PRODUCT OVERVIEW

Rote rehearsal and spacing effects in the free recall of pure and mixed lists. By: Peter P.J.L. Verkoeijen and Peter F. Delaney

A Bootstrapping Model of Frequency and Context Effects in Word Learning

AN ANALYSIS OF GRAMMTICAL ERRORS MADE BY THE SECOND YEAR STUDENTS OF SMAN 5 PADANG IN WRITING PAST EXPERIENCES

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

School Competition and Efficiency with Publicly Funded Catholic Schools David Card, Martin D. Dooley, and A. Abigail Payne

Guidelines for Writing an Internship Report

Natural Language Processing. George Konidaris

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Probability estimates in a scenario tree

1 3-5 = Subtraction - a binary operation

REVIEW OF CONNECTED SPEECH

Linking the Common European Framework of Reference and the Michigan English Language Assessment Battery Technical Report

Integrating simulation into the engineering curriculum: a case study

Creating Travel Advice

First Grade Curriculum Highlights: In alignment with the Common Core Standards

NAME: East Carolina University PSYC Developmental Psychology Dr. Eppler & Dr. Ironsmith

Sources of difficulties in cross-cultural communication and ELT: The case of the long-distance but in Chinese discourse

Calculators in a Middle School Mathematics Classroom: Helpful or Harmful?

DOES RETELLING TECHNIQUE IMPROVE SPEAKING FLUENCY?

Ch VI- SENTENCE PATTERNS.

Learning By Asking: How Children Ask Questions To Achieve Efficient Search

Campus Academic Resource Program An Object of a Preposition: A Prepositional Phrase: noun adjective

Concept Acquisition Without Representation William Dylan Sabo

Running head: DELAY AND PROSPECTIVE MEMORY 1

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Procedia - Social and Behavioral Sciences 154 ( 2014 )

Revisiting the role of prosody in early language acquisition. Megha Sundara UCLA Phonetics Lab

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Dear Teacher: Welcome to Reading Rods! Reading Rods offer many outstanding features! Read on to discover how to put Reading Rods to work today!

Listening and Speaking Skills of English Language of Adolescents of Government and Private Schools

The History of Language Teaching

South Carolina English Language Arts

Formulaic Language and Fluency: ESL Teaching Applications

Transcription:

Language Learning ISSN 0023-8333 A in Sentence Processing Morten H. Christiansen Cornell University Maryellen C. MacDonald University of Wisconsin-Madison Most current approaches to linguistic structure suggest that language is recursive, that recursion is a fundamental property of grammar, and that independent performance constraints limit recursive abilities that would otherwise be infinite. This article presents a usage-based perspective on recursive sentence processing, in which recursion is construed as an acquired skill and in which limitations on the processing of recursive constructions stem from interactions between linguistic experience and intrinsic constraints on learning and processing. A connectionist model embodying this alternative theory is outlined, along with simulation results showing that the model is capable of constituent-like generalizations and that it can fit human data regarding the differential processing difficulty associated with center-embeddings in German and crossdependencies in Dutch. Novel predictions are furthermore derived from the model and corroborated by the results of four behavioral experiments, suggesting that acquired recursive abilities are intrinsically bounded not only when processing complex recursive constructions, such as center-embedding and cross-dependency, but also during processing of the simpler, right- and left-recursive structures. Introduction Ever since Humboldt (1836/1999, researchers have hypothesized that language makes infinite use of finite means. Yet the study of language had to wait nearly We thank Jerry Cortrite, Jared Layport, and Mariana Sapera for their assistance in data collection and Brandon Kohrt for help with the stimuli. We are also grateful to Christina, Behme, Shravan Vasishth, and two anonymous reviewers for their comments on an earlier version of this article. Correspondence concerning this article should be addressed to Morten H. Christiansen, Department of Psychology, Uris Hall, Cornell University, Ithaca, NY 14853. Internet: christiansen@ cornell.edu Language Learning 59:Suppl. 1, December 2009, pp. 126 161 126 C 2009 Language Learning Research Club, University of Michigan

a century before the technical devices for adequately expressing the unboundedness of language became available through the development of recursion theory in the foundations of mathematics (cf. Chomsky, 1965). Recursion has subsequently become a fundamental property of grammar, permitting a finite set of rules and principles to process and produce an infinite number of expressions. Thus, recursion has played a central role in the generative approach to language from its very inception. It now forms the core of the Minimalist Program (Boeckx, 2006; Chomsky, 1995) and has been suggested to be the only aspect of the language faculty unique to humans (Hauser, Chomsky, &, Fitch, 2002). Although generative grammars sanction infinitely complex recursive constructions, people s ability to deal with such constructions is quite limited. In standard generative models of language processing, the unbounded recursive power of the grammar is therefore typically harnessed by postulating extrinsic memory limitations (e.g., on stack depth; Church, 1982; Marcus, 1980). This article presents an alternative, usage-based view of recursive sentence structure, suggesting that recursion is not an innate property of grammar or an a priori computational property of the neural systems subserving language. Instead, we suggest that the ability to process recursive structure is acquired gradually, in an item-based fashion given experience with specific recursive constructions. In contrast to generative approaches, constraints on recursive regularities do not follow from extrinsic limitations on memory or processing; rather they arise from interactions between linguistic experience and architectural constraints on learning and processing (see also Engelmann & Vasishth, 2009; MacDonald & Christiansen, 2002), intrinsic to the system in which the knowledge of grammatical regularities is embedded. Constraints specific to particular recursive constructions are acquired as part of the knowledge of the recursive regularities themselves and therefore form an integrated part of the representation of those regularities. As we will see next, recursive constructions come in a variety of forms; but contrary to traditional approaches to recursion, we suggest that intrinsic constraints play a role not only in providing limitations on the processing of complex recursive structures, such as center-embedding, but also in constraining performance on the simpler right- and left-branching recursive structures albeit to a lesser degree. Varieties of Recursive Structure Natural language is typically thought to involve a variety of recursive constructions. 1 The simplest recursive structures, which also tend to be the most 127 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

common in normal speech, are either right-branching as in (1) or left-branching as in (2): (1) a. John saw the dog that chased the cat. b. John saw the dog that chased the cat that bit the mouse. (2) a. The fat black dog was sleeping. b. The big fat black dog was sleeping. In the above example sentences, (1a) can be seen as incorporating a single level of right-branching recursion in the form of the embedded relative clause that chased the cat. Sentence (1b) involves two levels of right-branching recursion because of the two embedded relative clauses that chased the cat and that bit the mouse. A single level of left-branching recursion is part of (2a) in the form of the adjective fat fronting black dog. In (2b) two adjectives, big and fat, iteratively front black dog, resulting in a left-branching construction with two levels of recursion. Because right- and left-branching recursion can be captured by iterative processes, we will refer to them together as iterative recursion (Christiansen & Chater, 1999). Chomsky (1956) showed that iterative recursion of infinite depth can be processed by a finite-state device. However, recursion also exists in more complex forms that cannot be processed in its full, unbounded generality by finite-state devices. The best known type of such complex recursion is center-embedding as exemplified in (3): (3) a. The dog that John saw chased the cat. b. The cat that the dog that John saw chased bit the mouse. These sentences provide center-embedded versions of the right-branching recursive constructions in (1). In (3a), the sentence John saw the dog is embedded as a relative clause within the main sentence the dog chased the cat, generating one level of center-embedded recursion. Two levels of center-embedded recursion can be observed in (3b), in which John saw the dog is embedded within the dog chased the cat, which, in turn, is embedded within the cat bit the mouse. The processing of center-embedded constructions has been studied extensively in psycholinguistics for more than half a century. These studies have shown, for example, that English sentences with more than one centerembedding [e.g., sentence (3b)] are read with the same intonation as a list of random words (Miller, 1962), cannot easily be memorized (Foss & Cairns, 1970; Miller & Isard, 1964), are difficult to paraphrase (Hakes & Foss, 1970; Larkin & Burns, 1977) and comprehend (Blaubergs & Braine, 1974; Hakes, Language Learning 59:Suppl. 1, December 2009, pp. 126 161 128

Evans, & Brannon, 1976; Hamilton & Deese, 1971; Wang, 1970), and are judged to be ungrammatical (Marks, 1968). These processing limitations are not confined to English. Similar patterns have been found in a variety of languages, ranging from French (Peterfalvi & Locatelli, 1971), German (Bach, Brown, & Marslen-Wilson, 1986), and Spanish (Hoover, 1992) to Hebrew (Schlesinger, 1975), Japanese (Uehara & Bradley, 1996) and Korean (Hagstrom & Rhee, 1997). Indeed, corpus analyses of Danish, English, Finnish, French, German, Latin, and Swedish (Karlsson, 2007) indicate that doubly center-embedded sentences are practically absent from spoken language. Moreover, it has been shown that using sentences with a semantic bias or giving people training can improve performance on such structures, but only to a limited extent (Blaubergs & Braine, 1974; Powell & Peters, 1973; Stolz, 1967). Symbolic models of sentence processing typically embody a rule-based competence grammar that permits unbounded recursion. This means that the models, unlike humans, can process sentences with multiple centerembeddings. Since Miller and Chomsky (1963), the solution to this mismatch has been to impose extrinsic memory limitations exclusively aimed at capturing the human performance limitations on doubly center-embedded constructions. Examples include limits on stack depth (Church, 1982; Marcus, 1980), limits on the number of allowed sentence nodes (Kimball, 1973) or partially complete sentence nodes in a given sentence (Stabler, 1994), limits on the amount of activation available for storing intermediate processing products as well as executing production rules (Just & Carpenter, 1992), the self-embedding interference constraint (Gibson & Thomas, 1996), and an upper limit on sentential memory cost (Gibson, 1998). No comparable limitations are imposed on the processing of iterative recursive constructions in symbolic models. This may due to the fact that even finite-state devices with bounded memory are able to process right- and leftbranching recursive structures of infinite length (Chomsky, 1956). It has been widely assumed that depth of recursion does not affect the acceptability (or processability) of iterative recursive structures in any interesting way (e.g., Chomsky, 1965; Church, 1982; Foss & Cairns, 1970; Gibson, 1998; Reich, 1969; Stabler, 1994). Indeed, many studies of center-embedding in English have used right-branching relative clauses as baseline comparisons and found that performance was better relative to the center-embedded stimuli (e.g., Foss & Cairns, 1970; Marks, 1968; Miller & Isard, 1964). A few studies have reported more detailed data on the effect of depth of recursion in right-branching constructions and found that comprehension also decreases as depth of recursion increases in these structures, although not too the same degree as with center-embedded 129 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

stimuli (e.g., Bach et al., 1986; Blaubergs & Braine, 1974). However, it is not clear from these results whether the decrease in performance is caused by recursion per se or is merely a byproduct of increased sentence length. In this article, we investigate four predictions derived from an existing connectionist model of the processing of recursive sentence structure (Christiansen, 1994; Christiansen & Chater, 1994). First, we provide a brief overview of the model and show that it is capable of constituent-based generalizations and that it can fit key human data regarding the processing of complex recursive constructions in the form of center-embedding in German and cross-dependencies in Dutch. The second half of the article describes four online grammaticality judgment experiments testing novel predictions, derived from the model, using a word-by-word self-paced reading task. Experiments 1 and 2 tested two predictions concerning iterative recursion, and Experiments 3 and 4 tested predictions concerning the acceptability of doubly center-embedded sentences using, respectively, semantically biased stimuli from a previous study (Gibson & Thomas, 1999) and semantically neutral stimuli. A Connectionist Model of Recursive Sentence Processing Our usage-based approach to recursion builds on a previously developed Simple Recurrent Network (SRN; Elman, 1990) model of recursive sentence processing (Christiansen, 1994; Christiansen & Chater, 1994). The SRN, as illustrated in Figure 1, is essentially a standard feed-forward network equipped with an extra layer of so-called context units. The hidden unit activations from the previous time step are copied back to these context units and paired with the Figure 1 The basic architecture of the SRN used here as well as in Christiansen (1994) and Christiansen and Chater (1994). Arrows with solid lines denote trainable weights, whereas the arrow with the dashed line denotes the copy-back connections. Language Learning 59:Suppl. 1, December 2009, pp. 126 161 130

S NP VP. NP N N PP N rel PossP N N and NP VP V i V t NP V o (NP) V c that S rel who VP who NP V t o PP prep N loc (PP) PossP (PossP) N Poss Figure 2 The context-free grammar used to generate training stimuli for the connectionist model of recursive sentence processing developed by Christiansen (1994) and Christiansen and Chater (1994). current input. This means that the current state of the hidden units can influence the processing of subsequent inputs, providing the SRN with an ability to deal with integrated sequences of input presented successively. The SRN was trained via a word-by-word prediction task on 50,000 sentences (mean length: 6 words; range: 3 15 words) generated by a context-free grammar (see Figure 2) with a 38-word vocabulary. 2 This grammar involved left-branching recursion in the form of prenominal possessive genitives, rightbranching recursion in the form of subject relative clauses, sentential complements, prepositional modifications of NPs, and NP conjunctions, as well as complex recursion in the form of center-embedded relative clauses. The grammar also incorporated subject noun/verb agreement and three additional verb argument structures (transitive, optionally transitive, and intransitive). The generation of sentences was further restricted by probabilistic constraints on the complexity and depth of recursion. Following training, the SRN performed well on a variety of recursive sentence structures, demonstrating that the SRN was able to acquire complex grammatical regularities. 3 Usage-Based Constituents A key question for connectionist models of language is whether they are able to acquire knowledge of grammatical regularities going beyond simple cooccurrence statistics from the training corpus. Indeed, Hadley (1994) suggested that connectionist models could not afford the kind of generalization abilities necessary to account for human language processing (see Marcus, 1998, for a similar critique). Christiansen and Chater (1994) addressed this challenge using the SRN from Christiansen (1994). In the training corpus, the noun boy had been prevented from ever occurring in a NP conjunction (i.e., NPs such 131 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

as John and boy and boy and John did not occur). During training, the SRN had therefore only seen singular verbs following boy. Nonetheless, the network was able to correctly predict that a plural verb must follow John and boy as prescribed by the grammar. Additionally, the network was still able to correctly predict a plural verb when a prepositional phrase was attached to boy as in John and boy from town. This suggests that the SRN is able to make nonlocal generalizations based on the structural regularities in the training corpus (see Christiansen & Chater, 1994, for further details). If the SRN relied solely on local information, it would not have been able to make correct predictions in either case. Here, we provide a more stringent test of the SRN s ability to make appropriate constituent-based generalizations, using the four different types of test sentences shown in (4): (4) a. Mary says that John and boy see. (known word) b. Mary says that John and zog see. (novel word) c. Mary says that John and near see. (illegal word) d. Mary says that John and man see. (control word) Sentence (4a) is similar to what was used by Christiansen and Chater (1994) to demonstrate correct generalization for the known word, boy,usedinanovel position. In (4b), a completely novel word, zog, which the SRN had not seen during training (i.e., the corresponding unit was never activated during training) is activated as part of the NP conjunction. As an ungrammatical contrast, (4c) involves the activation of a known word, near, used in a novel but illegal position. Finally, (4d) provides a baseline in which a known word, man, is used in a position in which it is likely to have occurred during training (although not in this particular sentence). Figure 3 shows the summed activation for plural verbs for each of the four sentence types in (4). Strikingly, both the known word in a novel position as well as the completely novel word elicited activations of the plural verbs that were just as high as for the control word. In contrast, the SRN did not activate plural verbs after the illegal word, indicating that it is able to distinguish between known words used in novel positions (which are appropriate given its distributionally defined lexical category) versus known words used in an ungrammatical context. Thus, the network demonstrated sophisticated generalization abilities, ignoring local word co-occurrence constraints while appearing to comply with structural information at the constituent level. It is important to note, however, that SRN is unlikely to have acquired constituency in a categorical form (Christiansen & Chater, 2003) but instead have acquired constituents Language Learning 59:Suppl. 1, December 2009, pp. 126 161 132

Figure 3 Activation of plural verbs after presentation of the sentence fragment Mary says that John and N...,whereN is either a known word in a known position (boy), anovelword(zog), a known word in an illegal position (near), or a control word that have previously occurred in this position (man). that are more in line with the usage-based notion outlined by Beckner and Bybee (this issue). Deriving Novel Predictions Simple Recurrent Networks have been employed successfully to model many aspects of psycholinguistic behavior, ranging from speech segmentation (e.g., Christiansen, Allen, & Seidenberg, 1998; Elman, 1990) and word learning (e.g., Sibley, Kello, Plaut, & Elman, 2008) to syntactic processing (e.g., Christiansen, Dale, & Reali, in press; Elman 1993; Rohde, 2002; see also Ellis & Larsen-Freeman, this issue) and reading (e.g., Plaut, 1999). Moreover, SRNs have also been shown to provide good models of nonlinguistic sequence learning (e.g., Botvinick & Plaut, 2004, 2006; Servan-Schreiber, Cleeremans, & McClelland, 1991). The human-like performance of the SRN can be attributed to an interaction between intrinsic architectural constraints (Christiansen & Chater, 1999) and the statistical properties of its input experience (MacDonald & Christiansen, 2002). By analyzing the internal states of SRNs before and after training with right-branching and center-embedded materials, Christiansen and Chater found that this type of network has a basic architectural bias toward locally bounded dependencies similar to those typically found in iterative 133 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

recursion. However, in order for the SRN to process multiple instances of iterative recursion, exposure to specific recursive constructions is required. Such exposure is even more crucial for the processing of center-embeddings because the network in this case also has to overcome its architectural bias toward local dependencies. Hence, the SRN does not have a built-in ability for recursion, but instead it develops its human-like processing of different recursive constructions through exposure to repeated instances of such constructions in the input. In previous analyses, Christiansen (1994) noted certain limitations on the processing of iterative and complex recursive constructions. In the following, we flesh out these results in detail using the Grammatical Prediction Error (GPE) measure of SRN performance (Christiansen & Chater, 1999; MacDonald & Christiansen, 2002). To evaluate the extent to which a network has learned a grammar after training, performance on a test set of sentences is measured. For each word in the test sentences, a trained network should accurately predict the next possible words in the sentence; that is, it should activate all and only the words that produce grammatical continuations of that sentence. Moreover, it is important from a linguistic perspective not only to determine whether the activated words are grammatical given prior context but also which items are not activated despite being sanctioned by the grammar. Thus, the degree of activation of grammatical continuations should correspond to the probability of those continuations in the training set. The GPE assesses all of these facets of SRN performance, taking correct activations of grammatical continuations, correct suppression of ungrammatical continuations, incorrect activations of ungrammatical continuations, and incorrect suppressions of grammatical continuations into account (see Appendix A for details). The GPE scores range between 0 and 1, providing a very stringent measure of performance. To obtain a perfect GPE score of 0, the SRN must not only predict all and only the next words prescribed by grammar but also be able to scale those predictions according to the lexical frequencies of the legal items. The GPE for an individual word reflects the difficulty that the SRN experienced for that word given the previous sentential context, and it can be mapped qualitatively onto word reading times, with low GPE values reflecting a prediction for short reading times and high values indicating long predicted reading times (MacDonald & Christiansen, 2002). The mean GPE averaged across a sentence expresses the difficulty that the SRN experienced across the sentence as a whole, and such GPE values have been found to correlate with sentence grammaticality ratings (Christiansen & Chater, 1999), with low mean GPE scores predicting low grammatical complexity ratings and high Language Learning 59:Suppl. 1, December 2009, pp. 126 161 134

Figure 4 An illustration of the dependencies between subject nouns and verbs (arrows below) and between transitive verbs and their objects (arrows above) in sentences with two center-embeddings (a) and two cross-dependencies (b). scores indicating a prediction for high complexity ratings. Next, we first use mean sentence GPE scores to fit data from human experiments concerning the processing of complex recursive constructions in German and Dutch, after which we derive novel predictions concerning human grammaticality ratings for both iterative and center-embedded recursive constructions in English and present four experiments testing these predictions. Center-Embedding Versus Cross-Dependency Center-embeddings and cross-dependencies have played an important role in the theory of language. Whereas center-embedding relations are nested within each other, cross-dependencies cross over one another (see Figure 4). As noted earlier, center-embeddings can be captured by context-free grammars, but cross-dependencies require a more powerful grammar formalism (Shieber, 1985). Perhaps not surprisingly, cross-dependency constructions are quite rare across the languages of the world, but they do occur in Swiss-German and Dutch. An example of a Dutch sentence with two cross-dependencies is shown in (5), with subscripts indicating dependency relations. (5) De mannen 1 hebben Hans 2 Jeanine 3 de paarden helpen 1 leren 2 voeren 3 Literal: The men have Hans Jeanine the horses help teach feed Gloss: The men helped Hans teach Jeanine to feed the horses Although cross-dependencies have been assumed to be more difficult to process than comparable center-embeddings, Bach et al. (1986) found that sentences with two center-embeddings in German were significantly harder to process than comparable sentences with two cross-dependencies in Dutch. In order to model the comparative difficulty of processing centerembeddings versus cross-dependencies, we trained an SRN on sentences generated by a new grammar in which the center-embedded constructions were replaced by cross-dependency structures (see Figure 5). The iterative 135 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

S NP VP. S cd N 1 N 2 V 1(t o) V 2(i). S cd N 1 N 2 N V 1(t o) V 2(t o). S cd N 1 N 2 N 3 V 1(t o) V 2(t o) V 3(i). S cd N 1 N 2 N 3 N V 1(t o) V 2(t o) V3 (t o). NP N N PP N rel PossP N N and NP VP V i V t NP V o (NP) V c that S rel who VP PP prep N loc (PP) PossP (PossP) N Poss Figure 5 The context-sensitive grammar used to generate training stimuli for the connectionist model of recursive sentence processing developed by Christiansen (1994). recursive constructions, vocabulary, and other grammar properties remained the same as in the original context-free grammar. Thus, only the complex recursive constructions differed across the two grammars. In addition, all training and network parameters were held constant across the two simulations. After training, the cross-dependency SRN achieved a level of general performance comparable to that of the center-embedding SRN (Christiansen, 1994). Here, we focus on the comparison between the processing of the two complex types of recursion at different depths of embedding. Bach et al. (1986) asked native German speakers to provide comprehensibility ratings of German sentences involving varying depths of recursion in the form of center-embedded constructions and corresponding right-branching paraphrases with the same meaning. Native Dutch speakers were tested using similar Dutch materials but with the center-embedded constructions replaced by cross-dependency constructions. The left-hand side of Figure 6 shows the Bach et al. results, with the ratings for the right-branching paraphrase sentences subtracted from the matching complex recursive test sentences to remove effects of processing difficulty due to length. The SRN results the mean sentence GPE scores averaged over 10 novel sentences are displayed on the right-hand side of Figure 6. For both humans and SRNs, there is no difference in processing difficulty for the two types of complex recursion at one level of embedding. However, for doubly embedded constructions, center-embedded structures Language Learning 59:Suppl. 1, December 2009, pp. 126 161 136

Figure 6 Human performance (from Bach et al., 1986) on center-embedded constructions in German and cross-dependency constructions in Dutch with one or two levels of embedding (left panel). SRN performance on similar complex recursive structures (right panel). (in German) are harder to process than comparable cross-dependencies (in Dutch). These simulation results thus demonstrate that the SRNs exhibit the same kind of qualitative processing difficulties as humans do on the two types of complex recursive constructions (see also Christiansen & Chater, 1999). Crucially, the networks were able to match human performance without needing complex external memory devices (such as a stack of stacks; Joshi, 1990). Next, we go beyond fitting existing data to explore novel predictions made by the center-embedding SRN for the processing of recursive constructions in English. Experiment 1: Processing Multiple Right-Branching Prepositional Phrases In most models of sentence processing, multiple levels of iterative recursion are represented by having the exact same structure occurring several times (e.g., multiple instances of a PP). In contrast, the SRN learns to represent each level of recursion slightly differently from the previous one (Elman, 1991). This leads to increased processing difficulty as the level of recursion grows because the network has to keep track of each level of recursion separately, suggesting that depth of recursion in iterative constructions should affect processing difficulty beyond a mere length effect. Based on Christiansen s (1994) original analyses, we derived specific predictions for sentences involving zero, one, or two levels of right-branching recursion in the form of PP modifications of an NP 4 as shown in (6): 137 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

(6) a. The nurse with the vase says that the flowers by the window resemble roses. (1 PP) b. The nurse says that the flowers in the vase by the window resemble roses. (2 PPs) c. The blooming flowers in the vase on the table by the window resemble roses. (3 PPs) Predictions were derived from the SRNs for these three types of sentences and tested with human participants using a variation of the stop making sense sentence-judgment paradigm (Boland, 1997; Boland, Tanenhaus, & Garnsey, 1990; Boland, Tanenhaus, Garnsey, & Carlson, 1995), with a focus on grammatical acceptability rather than semantic sensibility. Following the presentation of each sentence, participants rated the sentence for grammaticality on a 7-point scale; these ratings were then compared with the SRN predictions. Method Participants Thirty-six undergraduate students from the University of Southern California received course credit for participation in this experiment. All participants in this and subsequent experiments were native speakers of English with normal or corrected-to-normal vision. Materials Nine experimental sentences were constructed with 1 PP, 2 PPs, and 3 PPs versions as in (6). All items are from this and subsequent experiments are included in Appendix B. Each sentence version had the same form as (6a) (6c). The 1 PP sentence type began with a definite NP modified by a single PP (The nurse with the vase), followed by a sentential complement verb and a complementizer (says that), a definite NP modified by a second single PP (the flowers by the window), and a final transitive VP with an indefinite noun (resemble roses). The 2 PP sentence type began with the same definite NP as 1 PP stimuli, followed by the same sentential complement verb and complementizer, a definite NP modified by a recursive construction with 2 PPs (the flowers in the vase by the window), and the same final transitive VP as 1 PP stimuli. The 3 PP sentence type began with a definite NP including an adjective (The blooming flowers), modified by a recursive construction with 3 PPs (in the vase on the table by the window), and the same transitive VP as in the other two sentence types. Each sentence was 14 words long and always ended with the same final NP (the window) and VP (resemble roses). Language Learning 59:Suppl. 1, December 2009, pp. 126 161 138

The three conditions were counterbalanced across three lists. In addition, 9 practice sentences and 42 filler sentences were created to incorporate a variety of recursive constructions of equal complexity to the experimental sentences. Two of the practice sentences were ungrammatical as were nine of the fillers. Twenty-one additional stimulus items were sentences from other experiments and 30 additional fillers mixed multiple levels of different kinds of recursive structures. Procedure Participants read sentences on a computer monitor, using a word-by-word center presentation paradigm. Each trial started with a fixation cross at the center of the screen. The first press of the space bar removed the fixation cross and displayed the first word of the sentence, and subsequent presses removed the previous word and displayed the next word. For each word, participants decided whether what they had read so far was a grammatical sentence of English. Participants were instructed that both speed and accuracy were important in the experiment and to base their decisions on their first impression about whether a sentence was grammatical. If the sentence read so far was considered grammatical, the participants would press the space bar if not, they would press a NO key when the sentence became ungrammatical. The presentation of a sentence ceased when the NO was pressed. When participants finished a sentence, either by reading it all the way through with the space bar or by reading it part way and then pressing the NO key when it became ungrammatical, the screen was cleared and they would be asked to rate how good this sentence was. 5 The participants would respond by pressing a number between 1 and 7 on the keyboard, with 1 indicating that the sentence was perfectly good English and 7 indicating that it was really bad English. Participants were encouraged to use the numbers in between for intermediate judgments. The computer recorded the response of the participant. Participants were assigned randomly to three counterbalanced lists. Each participant saw a different randomization of experimental and filler items. SRN Testing The model was tested on three sets of sentences corresponding to the three types shown in (6). The determiner the and the adjective in (6c) (blooming) could not be included in test sentences because they were not found in the training grammar. Moreover, the actual lexical items used in the network simulations were different from those in the human experiment because of limitations imposed by the training vocabulary, but the lexical categories remained the 139 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

same. The three sentence types had the same length as in the experiment, save that (6c) was one word shorter. All sentences involved at least two PPs [although only in (6b) and (6c) were they recursively related]. The crucial factor differentiating the three sentence types is the number of PPs modifying the subject noun (flowers) before the final verb (resemble). The sentence types were created to include 1, 2, or 3 PPs in this position. In order to ensure that the sentences were equal in length, right-branching sentential complements (says that...) were used in (6a) and (6b) such that the three sentence types are of the same global syntactic complexity. Mean GPE scores were recorded for 10 novel sentences of each type. Results and Discussion SRN Predictions Although the model found the sentences relatively easy to process, there was a significant effect of depth of recursion on GPE scores, F(2, 18) = 13.41, p <.0001, independent of sentence length (see Table 1). Thus, the model predicted an effect of sentence type for human ratings, with 3 PPs (6c) rated substantially worse than 2 PPs (6b), which, in turn, should be rated somewhat worse than 1 PP (6a). Rejection Data The PP stimuli were generally grammatically acceptable to our participants, with only 6.48% (21 trials) rejected during the reading/judgment task. Only 4.63% of the 1 PP stimuli and 3.70% of the 2 PP stimuli were rejected, and the difference between the two rejection scores was not significant, χ 2 (1) < 0.1. In contrast, 11.11% of the items with 3 PPs were rejected an increase in rejection rate that was significant compared with the 2 PP condition, χ 2 (1) = 3.51, p <.05, but only marginally significant in comparison with the 1 PP condition, χ 2 (1) = 2.43, p =.0595. Thus, there was a tendency to perceive the 1 PP and 2 PP stimuli as more grammatical than the counterpart with 3 PPs. Figure 7 shows the cumulative profile of rejections across word position in the sentences, starting at the fourth word. Rejections across the three sentence types were more likely to occur toward the end of a sentence, with two thirds of the rejections occurring during the presentation of the last four words, and with only three sentences rejected before the presentation of the 10th word (i.e., by in Figure 7). The rejection profile for the 3 PP stimuli suggests that it is the occurrence of the third PP (by the window) that makes these stimuli less acceptable than the 1 PP and 2 PP stimuli. Language Learning 59:Suppl. 1, December 2009, pp. 126 161 140

Figure 7 The cumulative percentage of rejections for each PP condition at each word position is shown starting from the fourth word. Table 1 The processing difficulty of multiple PP modifications of NPs SRN predictions Human results No. of PPs Mean GPE SD Mean rating SD 1 PP 0.153 0.008 2.380 0.976 2 PPs 0.161 0.015 2.704 0.846 3 PPs 0.214 0.042 3.269 0.925 Note. NP = noun phrase; PP = prepositional phrase. Grammaticality Ratings The ratings for the three sentence types are shown in Table 1. As predicted by the connectionist model, there was a significant effect of sentence type, F 1 (2, 70) = 10.87, p <.0001; F 2 (2, 16) = 12.43, p <.001, such that the deeper the level of recursion, the worse the sentences were rated. The model also predicted that there should be only a small difference between the ratings for the 1 PP and the 2 PP stimuli but a significant difference between the stimuli with the 2 PPs and 3 PPs. The experiment also bears out this prediction t. The stimuli with the 2 PPs were rated only 13.62% worse than the 1 PP stimuli a difference that was only marginally significant, F 1 (1, 35) = 2.97, p =.094; F 2 (1, 8) = 4.56, p =.065. The items with 3 PPs elicited the worst ratings, which were 37.36% worse than the 1 PP items and 20.89% worse than the 2 PP items. The rating 141 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

difference between the sentences with 2 PPs and 3 PPs was significant, F 1 (1, 35) = 5.74, p <.005; F 2 (1, 8) = 10.90, p <.02). The human ratings thus confirmed the predictions from the connectionist model: Increasing the depth of right-branching recursion has a negative effect on processing difficulty that cannot be attributed to a mere length effect. As predicted by the model, the deeper the level of recursion across the three types of stimuli, the worse the sentences were rated by the participants. This result is not predicted by most other current models of sentence processing, in which right-branching recursion does not cause processing difficulties beyond potential length effects (although see Lewis & Vasishth, 2005). Experiment 2: Processing Multiple Left-Branching Possessive Genitives In addition to the effect of multiple instances of right-branching iterative recursion on processing as confirmed by Experiment 1, Christiansen (1994) also observed that the depth of recursion effect in left-branching structures varied in its severity depending on the sentential position in which such recursion occurs. When processing left-branching recursive structures involving multiple prenominal genitives, the SRN learns that it is not crucial to keep track of what occurs before the final noun. This tendency is efficient early in the sentence but creates a problem with recursion toward the end of sentence because the network becomes somewhat uncertain where it is in the sentence. We tested this observation in the context of multiple possessive genitives occurring in either subject (7a) or object (7b) positions in transitive constructions: (7) a. Jane s dad s colleague s parrot followed the baby all afternoon. (Subject) b. The baby followed Jane s dad s colleague s parrot all afternoon. (Object) Method SRN Testing The model was tested as in Experiment 1 on two sets of 10 novel sentences corresponding the two types of sentences in (7). Participants Thirty-four undergraduate students from the University of Southern California received course credit for participation in this experiment. Materials We constructed 10 experimental items with the same format as (7). As in (7a), the Subject stimuli started with three prenominal genitives, of which the first Language Learning 59:Suppl. 1, December 2009, pp. 126 161 142

always contained a proper name (Jane s dad s colleague s), followed by the subject noun (parrot), a transitive verb (followed), a simple object NP (the baby), and a duration adverbial (all afternoon). The Object stimuli reversed the order of the two NPs, placing the multiple prenominal genitives in the object position and the simple NP in the subject position, as illustrated by (7b). The conditions were counterbalanced across two lists, each containing five sentences of each type. Additionally, there were 9 practice items (including one ungrammatical), 29 filler items (of which 9 were ungrammatical), and 20 items from other experiments. Procedure Experiment 2 involved the same procedure as Experiment 1. Results and Discussion SRN Predictions Comparisons of mean sentence GPE for the two types of sentence materials predicted that having two levels of recursion in an NP involving left-branching prenominal genitives should be significantly less acceptable in an object position compared to a subject position, F(1, 9) = 110.33, p <.0001. Rejection Data Although the genitive stimuli seemed generally acceptable, participants rejected twice as many sentences (13.24%) as in Experiment 1. The rejection profiles for the two sentence types are illustrated in Figure 8, showing that the rejections are closely associated with the occurrence of the multiple prenominal genitives. However, there was no overall difference in the number of sentences rejected in the Subject (13.53%) and Object (12.94%) conditions, χ 2 (1) < 1. Grammaticality Ratings As predicted by the SRN model, the results in Table 2 show that multiple prenominal genitives were less acceptable in object position than in subject position, F 1 (1, 33) = 5.76, p <.03; F 2 (1, 9) = 3.48, p =.095. These results suggest that the position of multiple instances of recursion within a sentence affects its acceptability. Experiment 3: Processing Multiple Semantically Biased Center-Embeddings In contrast to iterative recursion, complex recursion in the form of centerembedding has often been used as an important source of information about 143 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

Figure 8 The cumulative percentage of rejections for sentences incorporating multiple prenominal genitives in subject or object positions. complexity effects in human sentence processing (e.g., Blaubergs & Braine, 1974; Foss & Cairns, 1970; Marks, 1968; Miller, 1962; Miller & Isard, 1964; Stolz, 1967). Of particular interest is a study by Gibson and Thomas (1999) investigating the role of memory limitations in the processing of doubly centerembedded object relative clause constructions. Consistent with the external memory limitation account of Gibson (1998), they found that when deleting themiddlevp[was cleaning every week in (8a)], the resulting ungrammatical sentence (8b) was rated no worse that the original grammatical version. (8) a. The apartment that the maid who the service had sent over was cleaning every week was well decorated. (3 VPs) Table 2 The processing difficulty of multiple possessive genitives SRN predictions Human results Genitive position Mean GPE SD Mean rating SD Sub NP 0.222 0.006 3.606 1.715 Obj NP 0.299 0.025 3.965 1.752 Note. Sub NP = subject noun phrase; Obj NP = object noun phrase. Language Learning 59:Suppl. 1, December 2009, pp. 126 161 144

b. The apartment that the maid who the service had sent over was well decorated. (2 VPs) In contrast, Christiansen (1994) noted that the SRN tended to expect that doubly center-embedded sentences would end when it had received two verbs, suggesting that (8b) should actually be rated better than (8a). Christiansen and Chater (1999) further demonstrated that this prediction is primarily due to intrinsic architectural limitations on the processing on doubly center-embedded material rather than insufficient experience with these constructions. 6 Gibson and Thomas results came from offline ratings, whereas in Experiment 3 we use the online method from the previous two experiments and predict that with this more sensitive measure, sentences such as the ungrammatical (8b) will actually be rated better than grammatical sentences like (8a). Method SRN Testing The model was tested as in Experiments 1 and 2 on two sets of 10 novel sentences corresponding to the sentence types in (8). Participants Thirty-six undergraduate students from the University of Southern California received course credit for participation in this experiment. Materials Six experimental items were selected from the Gibson and Thomas (1999) stimuli, focusing on the key grammatical (3 VP) versus ungrammatical (2 VPs) version of each sentence as in (8). The two conditions were counterbalanced across two lists, three of each type. In addition, there were 9 practice items (including 2 ungrammatical), 30 filler items (of which 9 were ungrammatical), and 27 items from other experiments. Procedure Experiment 3 involved the same procedure as Experiments 1 and 2. Results and Discussion SRN Predictions The mean GPE scores across the two types of sentences followed the preliminary findings by Christiansen and Chater (1999): The grammatical 3-VP sentences were rated significantly worse than the ungrammatical 2-VP sentences, F(1, 9) = 2892.23, p <.0001. 145 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

Figure 9 The cumulative percentage of rejections averaged across NP and VP regions in semantically biased center-embedded sentences with 2 VPs or 3 VPs. Rejection Data Because the stimuli from Gibson and Thomas (1999) were not equated for length, the number of rejections were averaged for each NP and VP region rather than for each word. The resulting cumulative rejection profile is shown in Figure 9, indicating that significantly more 3 VP sentences were rejected than 2 VP sentences [63% vs. 32.4%; χ 2 (1) = 20.21, p <.0001]. Grammaticality Ratings As predicted by the SRN model (Table 3), the grammatical 3 VP sentences were rated significantly worse than their ungrammatical 2 VP counterparts, F 1 (1, 35) = 15.55, p <.0001; F 2 (1, 5) = 6.85, p <.05. These results suggest Table 3 The processing difficulty of multiple semantically biased center-embeddings SRN predictions Human results No. of VPs Mean GPE SD Mean rating SD 2 VPs 0.307 0.010 4.778 1.268 3 VPs 0.404 0.005 5.639 1.037 Note. VPs = verb phrases. Language Learning 59:Suppl. 1, December 2009, pp. 126 161 146

humans share the SRN s processing preference for the ungrammatical 2 VP construction over the grammatical 3 VP version (for similar SRN results and additional data for German, see Engelmann & Vasishth, 2009). Experiment 4: Processing Multiple Semantically Neutral Center-Embeddings Two potential concerns about the Gibson and Thomas (1999) stimuli used in Experiment 3 are that (a) the results could be an artifact of length because the sentences were not controlled for overall length and (b) the stimuli included semantic biases [e.g., apartment/decorated, service/sent over in (6b)] that may have increased the plausibility of the 2 VP stimuli. In Experiment 4, we sought to replicate the results from Experiment 3 with semantically neutral stimuli adapted from Stolz (1967), in which adverbs replaced the missing verbs in 2 VP constructions to control for overall length as in (9): (9) a. The chef who the waiter who the busboy offended appreciated admired the musicians. (3 VPs) b. The chef who the waiter who the busboy offended frequently admired the musicians. (2 VPs) Method SRN Testing The training corpus on which the model was trained did not include semantic constraints (e.g., animacy). Instead, the difference between the centerembedded test items used to make SRN predictions for Experiments 3 and 4 was one of argument structure. The Gibson and Thomas (1999) stimuli in Experiment 3 used optionally transitive verbs, whereas the Experiment 4 stimuli contained transitive verbs. The model was tested as in the previous experiments on two sets of 10 novel sentences matching the structure of the two sentence types in (9). Participants Thirty-four undergraduate students from the University of Southern California received course credit for participation in this experiment. Materials Ten semantically neutral doubly center-embedded items were adapted from Stolz (1967), each with a 3 VP and 2 VP version as in(9). The conditions were counterbalanced across two lists, each containing five sentences of each 147 Language Learning 59:Suppl. 1, December 2009, pp. 126 161

Figure 10 The cumulative percentage of rejections averaged across word position in semantically neutral center-embedded sentences with 2 VPs or 3 VPs. type. Additionally, there were 9 practice items (including one ungrammatical), 29 filler items (of which 9 were ungrammatical), and 20 items from other experiments. Procedure Experiment 4 involved the same procedure as Experiments 1 3. Results and Discussion SRN Predictions As in Experiment 3, the mean GPE scores predicted that 3 VP sentences should be rated significantly worse than 2 VP sentences, F(1, 9) = 43.60, p <.0001. Rejection Data The cumulative rejection profile for the neutral items in the current experiment replicated that for the semantically biased stimuli in the previous experiment (Figure 10): significantly more 3 VP sentences (78.8%) were rejected than the corresponding 2 VP constructions [52.9%; χ 2 (1) = 25.33, p <.0001]. Grammaticality Ratings Again, in line with the SRN predictions (Table 4), the 3 VP sentences were rated significantly worse than their 2 VP counterparts, F 1 (1, 33) = 7.88, p <.01; F 2 (1, 9) = 27.46, p <.001. Thus, the ungrammatical 2 VP constructions Language Learning 59:Suppl. 1, December 2009, pp. 126 161 148

Table 4 The processing difficulty of multiple semantically neutral center-embeddings SRN predictions Human results No. of VPs Mean GPE SD Mean rating SD 2 VPs 0.360 0.015 5.553 1.251 3 VPs 0.395 0.008 6.165 0.672 Note. VPs = verb phrases. are preferred over the grammatical 3 VP versions even when controlling for overall length and the influence of semantic bias. General Discussion We have presented simulation results from a connectionist implementation of our usage-based approach to recursion, indicating that the model has sophisticated constituent-based generalization abilities and is able to fit human data regarding to the differential processing difficulty of center-embedded and crossdependency constructions. Novel predictions were then derived from the model and confirmed by the results of four grammaticality judgment experiments. Importantly, this model was not developed for the purpose of fitting these data but was, nevertheless, able to predict the patterns of human grammaticality judgments across three different kinds of recursive structure. Indeed, as illustrated by Figure 11, the SRN predictions not only provide a close fit with the human ratings within each experiment but also capture the increased complexity evident across Experiments 1 4. Importantly, the remarkably good fit between the model and the human data both within and across the experiments were obtained without changing any parameters across the simulations. In contrast, the present pattern of results provides a challenge for most other accounts of human sentence processing that rely on arbitrary, externally specified limitations on memory or processing to explain patterns of human performance. Like other implemented computational models, the specific instantiation of our usage-based approach to recursive sentence processing presented here is not without limitations. Although the model covers several key types of recursive sentence constructions, its overall coverage of English is limited in both vocabulary size and range of grammatical regularities. Another limiting factor is that the model predicts only the next word in a sentence. Despite mounting evidence highlighting the importance of prediction to learning, in general (Niv & Schoenbaum, 2008), and language processing, in particular (e.g., Federmeier, 2007; Levy, 2008; Hagoort, in press; Pickering & Garrod, 2007), incorporating 149 Language Learning 59:Suppl. 1, December 2009, pp. 126 161