Sentence comprehension. Event structures. Sentence Gestalt Model (St. John & McClelland, 1990) Sentence generation

Similar documents
1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

CS 598 Natural Language Processing

Context Free Grammars. Many slides from Michael Collins

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Construction Grammar. University of Jena.

Good-Enough Representations in Language Comprehension

A Usage-Based Approach to Recursion in Sentence Processing

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Derivational and Inflectional Morphemes in Pak-Pak Language

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Developing Grammar in Context

Chapter 4: Valence & Agreement CSLI Publications

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Copyright and moral rights for this thesis are retained by the author

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Inleiding Taalkunde. Docent: Paola Monachesi. Blok 4, 2001/ Syntax 2. 2 Phrases and constituent structure 2. 3 A minigrammar of Italian 3

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Natural Language Processing. George Konidaris

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Using computational modeling in language acquisition research

Argument structure and theta roles

ELD CELDT 5 EDGE Level C Curriculum Guide LANGUAGE DEVELOPMENT VOCABULARY COMMON WRITING PROJECT. ToolKit

Dear Teacher: Welcome to Reading Rods! Reading Rods offer many outstanding features! Read on to discover how to put Reading Rods to work today!

Adjectives tell you more about a noun (for example: the red dress ).

Words come in categories

California Department of Education English Language Development Standards for Grade 8

The Four Principal Parts of Verbs. The building blocks of all verb tenses.

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Constraining X-Bar: Theta Theory

Minimalism is the name of the predominant approach in generative linguistics today. It was first

Universal Grammar 2. Universal Grammar 1. Forms and functions 1. Universal Grammar 3. Conceptual and surface structure of complex clauses

An Interactive Intelligent Language Tutor Over The Internet

BASIC ENGLISH. Book GRAMMAR

The College Board Redesigned SAT Grade 12

Written by: YULI AMRIA (RRA1B210085) ABSTRACT. Key words: ability, possessive pronouns, and possessive adjectives INTRODUCTION

Good Enough Language Processing: A Satisficing Approach

An Empirical and Computational Test of Linguistic Relativity

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

An Introduction to the Minimalist Program

Compositional Semantics

Thornhill Primary School - Grammar coverage Year 1-6

Phenomena of gender attraction in Polish *

Adjectives In Paragraphs

Language Acquisition by Identical vs. Fraternal SLI Twins * Karin Stromswold & Jay I. Rifkin

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Underlying and Surface Grammatical Relations in Greek consider

More ESL Teaching Ideas

Participate in expanded conversations and respond appropriately to a variety of conversational prompts

Today we examine the distribution of infinitival clauses, which can be

The Structure of Multiple Complements to V

Aging and the Use of Context in Ambiguity Resolution: Complex Changes From Simple Slowing

Unit 8 Pronoun References

The Role of the Head in the Interpretation of English Deverbal Compounds

Common Core ENGLISH GRAMMAR & Mechanics. Worksheet Generator Standard Descriptions. Grade 2

Language Learning and Development. ISSN: (Print) (Online) Journal homepage:

Morphosyntactic and Referential Cues to the Identification of Generic Statements

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

The Acquisition of English Grammatical Morphemes: A Case of Iranian EFL Learners

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

2017 national curriculum tests. Key stage 1. English grammar, punctuation and spelling test mark schemes. Paper 1: spelling and Paper 2: questions

Syntactic Ambiguity Resolution in Sentence Processing: New Evidence from a Morphologically Rich Language

Citation for published version (APA): Veenstra, M. J. A. (1998). Formalizing the minimalist program Groningen: s.n.

L1 and L2 acquisition. Holger Diessel

Control and Boundedness

Frequency and pragmatically unmarked word order *

Grammars & Parsing, Part 1:

Ambiguity in the Brain: What Brain Imaging Reveals About the Processing of Syntactically Ambiguous Sentences

Ch VI- SENTENCE PATTERNS.

On the Notion Determiner

Generation of Referring Expressions: Managing Structural Ambiguities

Parsing of part-of-speech tagged Assamese Texts

BULATS A2 WORDLIST 2

Advanced Grammar in Use

Writing a composition

Ambiguities and anomalies: What can eye-movements and event-related potentials reveal about second language sentence processing?

In search of ambiguity

Some Principles of Automated Natural Language Information Extraction

Sight Word Assessment

What the National Curriculum requires in reading at Y5 and Y6

Language acquisition: acquiring some aspects of syntax.

The Structure of Relative Clauses in Maay Maay By Elly Zimmer

Arizona s English Language Arts Standards th Grade ARIZONA DEPARTMENT OF EDUCATION HIGH ACADEMIC STANDARDS FOR STUDENTS

Opportunities for Writing Title Key Stage 1 Key Stage 2 Narrative

The Smart/Empire TIPSTER IR System

English for Life. B e g i n n e r. Lessons 1 4 Checklist Getting Started. Student s Book 3 Date. Workbook. MultiROM. Test 1 4

Taught Throughout the Year Foundational Skills Reading Writing Language RF.1.2 Demonstrate understanding of spoken words,

Analyzing Linguistically Appropriate IEP Goals in Dual Language Programs

SEMAFOR: Frame Argument Resolution with Log-Linear Models

Stephen Craint and Donald Shankweilert. 1. Introduction

FOREWORD.. 5 THE PROPER RUSSIAN PRONUNCIATION. 8. УРОК (Unit) УРОК (Unit) УРОК (Unit) УРОК (Unit) 4 80.

CEFR Overall Illustrative English Proficiency Scales

Pseudo-Passives as Adjectival Passives

SENTENCE PARTS AND PATTERNS

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

Transcription:

Traditional view of language Language knowledge largely consists of an explicit grammar that determines what sentences are part of a language Isolated from or types of knowledge pragmatic, semantic, lexical(?) Language learning involves identifying single, correct grammar of language Grammar induction is underconstrained by linguistic input given lack of explicit negative evidence Impossible under near-arbitrary positive-only presentation (Gold, 1967) Language learning requires strong innate linguistic constraints to narrow range of possible grammars considered A connectionist approach to sentence processing (Elman, 1991) S NP VI. NP VT NP. NP N N RC RC who VI who VT NP who NP VT N boy girl cat dog Mary John boys girls cats dogs VI barks sings walks bites eats bark sing walk bite eat VT chases feeds walks bites eats chase feed walk bite eat Simple recurrent network trained to predict next word in English-like sentences Context-free grammar, number agreement, variable verb argument structure, multiple levels of embedding 75% of sentences had at least one relative clause; average length of 6 words. e.g., Girls who cat who lives chases walk dog who feeds girl who cats walk. 1 / 22 After 20 sweeps through 4 sets of 10,000 sentences, mean absolute error for new set of 10,000 sentences was 0.177 (cf. initial: 12.45; uniform: 1.92) 3 / 22 Statistical view of language Principal Components Analysis Language environment has rich distributional regularities May not provide correction but is certainly not adversarial (cf. Gold, 1967) Language learning requires only that knowledge across speakers converges sufficiently to support effective communication No sharp division between linguistic vs. extra-linguistic knowledge Effectiveness of learning depend both on structure of input and on existing knowledge (linguistic and extra-linguistic) Distributional information can provide implicit negative evidence Example: implicit prediction of upcoming input Sufficient for language learning when combined with domain-general biases Boy chases boy who chases boy who chases boy. Principal Components Analysis (PCA) of network s internal representations Largest amount of variance (PC-1) reflects word class (noun, verb, function word) Separate dimension of variation (PC-11) encodes syntactic role (agent/patient) for nouns and level of embedding for verbs 2 / 22 4 / 22

Sentence comprehension Event structures Traditional perspective Linguistic knowledge as grammar, separate from semantic/pragmatic influences on performance (Chomsky, 1957) Psychological models with initial syntactic parse that is insensitive to lexical/semantic constraints (Ferreira & Clifton, 1986; Frazier, 1986) Problem: Interdependence of syntax and semantics The spy saw policeman with a revolver The spy saw policeman with binoculars The bird saw birdwatcher with binoculars The pitcher threw ball The container held apples/cola The boy spread jelly on bread Alternative: Constraint satisfaction Sentence comprehension involves integrating multiple sources of information (both semantic and syntactic) to construct most plausible interpretation of a sentence (MacDonald et al., 1994; Seidenberg, 1997; Tanenhaus & Trueswell, 1995) 5 / 22 14 active frames, 4 passive frames, 9 matic roles Total of 120 possible events (varying in likelihood) 7 / 22 Sentence Gestalt Model (St. John & McClelland, 1990) Sentence generation Trained to generate matic role assignments of event described by single-clause sentence Sentence constituents ( phrases) presented one at a time After each constituent, network updates internal representation of sentence meaning ( Sentence Gestalt ) Current Sentence Gestalt trained to generate full set of role/filler pairs (by successive probes ) Must predict information based on partial input and learned experience, but must revise if incorrect 6 / 22 Given a specific event, probabilistic choices of Which matic roles are explicitly mentioned What word describes each constituent Active/passive voice Example: busdriver eating steak with knife -adult ate -food with-a-utensil -steak was-consumed-by -person someone ate something Total of 22,645 sentence-event pairs 8 / 22

Acquisition Sentence types Active syntactic: Passive syntactic: Regular semantic: Irregular semantic: Online updating and backtracking busdriver kissed teacher teacher was kissed by busdriver busdriver ate steak busdriver ate soup Active voice learned before passive voice Syntactic constraints learned before semantic constraints Final network tested on 55 randomly generated unambiguous sentences Correct on 1699/1710 (99.4%) of role/filler assignments 9 / 22 Implied constituents 11 / 22 Semantic-syntactic interactions Lexical ambiguity Concept instantiation 10 / 22 Implied constituents 12 / 22

Noun similarities Summary: St. John and McClelland (1990) Syntactic and semantic constraints can be learned and brought to bear in an integrated fashion to perform online sentence comprehension Approach stands in sharp contrast to linguistic and psycholinguistic ories espousing a clear separation of grammar from rest of cognition 13 / 22 15 / 22 Verb similarities Sentence comprehension and production (Rohde) 14 / 22 Extends approach of Sentence Gestalt model to multi-clause sentences Trained to generate learned message representation and to predict successive words in sentences when given varying degrees of prior context 16 / 22

Training language Message encoder Multiple verb tenses e.g., ran, was running, runs, is running, will run, will be running Passives Relative clauses (normal and reduced) Prepositional phrases Dative shift e.g., gave flowers to girl, gave girl flowers Singular, plural, and mass nouns 12 noun stems, 12 verb stems, 6 adjectives, 6 adverbs Examples The boy drove. An apple will be stolen by dog. Mean cops give John dog that was eating some food. John who is being chased by fast cars is stealing an apple which was had with pleasure. 17 / 22 Methods Triples presented in sequence For each triple, all presented triples queried three ways (given two elements, generate third) Trained on 2 million sentence meanings Full language Triples correct: 91.9% Components correct: 97.2% Units correct: 99.9% Reduced language ( 10 words): Triples correct: 99.9% 19 / 22 Encoding messages with triples The boy who is being chased by fast dogs stole some apples in park. Training: Comprehension (and prediction) Methods No context on half of trials Context was weak clamped (25% strength) on or half Initial state of message layer clamped with varying strength Correct query responses with comprehended message: Without context: 96.1% With context: 97.9% 18 / 22 20 / 22

Testing: Comprehension of relative clauses Single embedding: Center- vs. Right-branching; Subject- vs. Object-relative CS: A dog [who chased John] ate apples. RS: John chased a dog [who ate apples]. CO: A dog [who John chased] ate apples. RO: John ate a dog [who apples chased]. Empirical Data Model Testing: Production Methods Message initialized to correct value and weak clamped (25% strength) Most actively predicted word selected for production No explicit training 86.5% of sentences correctly produced. 21 / 22 22 / 22