Word Grammar. by Richard Hudson. Universität Tübingen, Word Grammar. Nika Strem, Iuliia Kocharina. Overview. The Cognitive Network

Similar documents
A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

LING 329 : MORPHOLOGY

Derivational and Inflectional Morphemes in Pak-Pak Language

Parallel Evaluation in Stratal OT * Adam Baker University of Arizona

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

CS 598 Natural Language Processing

AQUA: An Ontology-Driven Question Answering System

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

An Interactive Intelligent Language Tutor Over The Internet

Developing a TT-MCTAG for German with an RCG-based Parser

Parsing of part-of-speech tagged Assamese Texts

Dependency, licensing and the nature of grammatical relations *

Minimalism is the name of the predominant approach in generative linguistics today. It was first

Argument structure and theta roles

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

English Language and Applied Linguistics. Module Descriptions 2017/18

Proof Theory for Syntacticians

NAME: East Carolina University PSYC Developmental Psychology Dr. Eppler & Dr. Ironsmith

Abstractions and the Brain

The Acquisition of English Grammatical Morphemes: A Case of Iranian EFL Learners

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Citation for published version (APA): Veenstra, M. J. A. (1998). Formalizing the minimalist program Groningen: s.n.

Linking Task: Identifying authors and book titles in verbose queries

Constraining X-Bar: Theta Theory

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Some Principles of Automated Natural Language Information Extraction

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

The College Board Redesigned SAT Grade 12

Using dialogue context to improve parsing performance in dialogue systems

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Language Acquisition by Identical vs. Fraternal SLI Twins * Karin Stromswold & Jay I. Rifkin

An Introduction to the Minimalist Program

Basic concepts: words and morphemes. LING 481 Winter 2011

FOREWORD.. 5 THE PROPER RUSSIAN PRONUNCIATION. 8. УРОК (Unit) УРОК (Unit) УРОК (Unit) УРОК (Unit) 4 80.

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Compositional Semantics

SOME MINIMAL NOTES ON MINIMALISM *

On-Line Data Analytics

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

AN INTRODUCTION (2 ND ED.) (LONDON, BLOOMSBURY ACADEMIC PP. VI, 282)

The Acquisition of Person and Number Morphology Within the Verbal Domain in Early Greek

1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

Possessive have and (have) got in New Zealand English Heidi Quinn, University of Canterbury, New Zealand

Concept Acquisition Without Representation William Dylan Sabo

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

COMPUTATIONAL COMPLEXITY OF LEFT-ASSOCIATIVE GRAMMAR

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

The Effect of Extensive Reading on Developing the Grammatical. Accuracy of the EFL Freshmen at Al Al-Bayt University

Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology. Michael L. Connell University of Houston - Downtown

Control and Boundedness

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Chapter 4: Valence & Agreement CSLI Publications

Universal Grammar 2. Universal Grammar 1. Forms and functions 1. Universal Grammar 3. Conceptual and surface structure of complex clauses

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Evolution of Symbolisation in Chimpanzees and Neural Nets

Program Matrix - Reading English 6-12 (DOE Code 398) University of Florida. Reading

Context Free Grammars. Many slides from Michael Collins

Correspondence between the DRDP (2015) and the California Preschool Learning Foundations. Foundations (PLF) in Language and Literacy

Natural Language Processing. George Konidaris

ELA/ELD Standards Correlation Matrix for ELD Materials Grade 1 Reading

A Framework for Customizable Generation of Hypertext Presentations

The Effect of Discourse Markers on the Speaking Production of EFL Students. Iman Moradimanesh

Writing a composition

Inleiding Taalkunde. Docent: Paola Monachesi. Blok 4, 2001/ Syntax 2. 2 Phrases and constituent structure 2. 3 A minigrammar of Italian 3

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Prediction of Maximal Projection for Semantic Role Labeling

Chapter 3: Semi-lexical categories. nor truly functional. As Corver and van Riemsdijk rightly point out, There is more

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

Aspectual Classes of Verb Phrases

Loughton School s curriculum evening. 28 th February 2017

Chapter 9 Banked gap-filling

Language Acquisition Chart

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

THE INTERNATIONAL JOURNAL OF HUMANITIES & SOCIAL STUDIES

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

5. UPPER INTERMEDIATE

Intermediate Academic Writing

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

What is a Mental Model?

Procedia - Social and Behavioral Sciences 154 ( 2014 )

Multiple case assignment and the English pseudo-passive *

Lexical phonology. Marc van Oostendorp. December 6, Until now, we have presented phonological theory as if it is a monolithic

Linguistics. Undergraduate. Departmental Honors. Graduate. Faculty. Linguistics 1

Underlying and Surface Grammatical Relations in Greek consider

Developing True/False Test Sheet Generating System with Diagnosing Basic Cognitive Ability

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

LEXICAL COHESION ANALYSIS OF THE ARTICLE WHAT IS A GOOD RESEARCH PROJECT? BY BRIAN PALTRIDGE A JOURNAL ARTICLE

cambridge occasional papers in linguistics Volume 8, Article 3: 41 55, 2015 ISSN

Good-Enough Representations in Language Comprehension

Approaches to Teaching Second Language Writing Brian PALTRIDGE, The University of Sydney

Word Stress and Intonation: Introduction

Highlighting and Annotation Tips Foundation Lesson

Program in Linguistics. Academic Year Assessment Report

What the National Curriculum requires in reading at Y5 and Y6

On the Notion Determiner

Intensive English Program Southwest College

Florida Reading Endorsement Alignment Matrix Competency 1

Transcription:

by Richard Hudson Universität Tübingen, 2017 1 / 61

2 / 61

3 / 61

The Notion of Word grammar (WG) is a general theory of language structure WG is a branch of cognitive linguistics The main consideration is to be true to the facts of language structure WG assumption is that language can be analyzed and explained in the same way as other kinds of knowledge or behavior 4 / 61

Word as the Central Concept in WG The central unit of analysis is the word Grammar. Words are the only units of syntax, the largest and the smallest. Words provide the only point of contact between syntax and semantics Situation. Words are the basic units for contextual analysis Words are the nodes that hold the language part of the cognitive network together 5 / 61

Word as the Central Concept in WG 6 / 61

7 / 61

Language as Part of a General All knowledge is represented via a network of memory connections The sub-network of words is just a part of the total vast set of associations Thus, our language is a network of associations closely integrated with the rest of our knowledge 8 / 61

Language as Part of a General Traditionally, only the lexicon is viewed as a network in language (Aitchison 2012), storing only irregularities, while regularities are stored in a different way as rules (Pinker and Prince 1988). In WG view, exceptional and general patterns can both be accommodated in the same network due to default inheritance The only difference between these rules lies in two places: a verb versus come, and its stem followed by -ed versus came. Similarly, they can both be incorporated into the same network 9 / 61

Language as Part of a General 10 / 61

Language as Part of a General Vocabulary items are related in a network of phonological, syntactic and semantic links In language there is nothing but a network - no rules or principles or parameters or processes, except those that are expressed in terms of the network 11 / 61

Language as Part of a General The nodes of a WG network are atoms without any internal structure, so a language is not a network of complex information-packages such as lexical entries, constructions, schemas, or signs The information in each such package must be unpacked so that it can be integrated into the general network 12 / 61

Language as Part of a General 13 / 61

Labelled Links In cognitive psychology, there is a long tradition of associative network theories where all links have the same status In the WG view, links are classified and labelled stem, shape, sense, referent, subject, adjunct, etc. The classifying categories range from the most general (the isa link) to very specific categories 14 / 61

Labelled Links This approach allows named links to be used as functions, which yield a variable e. g. the referent of the subject of the verb defines one unique concept for each verb Every concept is uniquely defined by its links to other concepts, so labels are redundant. On the contrary, a network with unlabeled links is a mere associative network 15 / 61

Labelled Links Example: John saw Mary Labels serve to classify the links as same or different, so if we remove the label we lose information Each concept in a network is represented just once, with multiple links to other concepts rather than copies of the concept itself 16 / 61

Labelled Links 17 / 61

Modularity Is there a distinct module of the mind dedicated exclusively to language (or its layers)? In WG, the language network is a small part of a much larger network Language pathology: selective impairment of abilities Language network does not occupy a distinct part of the human mind or brain, but is intimately embedded in the general cognitive network (R. Hudson 2007) 18 / 61

19 / 61

Generalizations and exceptions Characteristics of a general category are inherited by instances of that category by default only when not overridden by a specific case (e. g. past tense) 20 / 61

is carried by the fundamental isa relation (e. g., because snores isa verb it automatically inherits all the known characteristics of a verb ) A fact is automatically blocked by any other fact which conflicts and is more specific 21 / 61

Multiple Multiple inheritance is an extension of default inheritance in which one concept inherits from two different concepts simultaneously: a cat inherits characteristics from mammal and from pet Ungrammaticality of *I aren t (You aren t vs. Aren t I?) due to conflicting inheritance from aren t (the negative present of BE) and am (the I-form of BE), without any way for either to override the other 22 / 61

Multiple Multiple default inheritance, defined by these facts about a concept A which isa B: : Normally A inherits all the characteristics of B and any other nodes on the isa chain leading up from B (i. e. any node which A transitive-isa) inheritance: But it does not inherit values for relations which already have a value Multiple inheritance: If A transitive-isa any other concept, it inherits from this in the same way as from B 23 / 61

24 / 61

Under WG, language is a network of concepts Language is part of the same general conceptual network that contains many concepts which are not part of language Language is a module in that the links among words are denser than those between words and other kinds of concepts, but it is not encapsulated or distinguished by special formal characteristics 25 / 61

A word may have a variety of links to other words and concepts: to its morphological structure via the stem and shape links to its semantics by the sense and referent links to its syntax by dependencies and word classes 26 / 61

27 / 61

Concepts are classified exclusively by means of inheritance hierarchies, which leaves no place for feature-descriptions: to classify a word as a verb in WG we give it an isa link to verb, not a feature-description which contains that of verb Multiple inheritance allows words to be classified on two different dimensions : as lexemes (DOG, LIKE, IF, etc.) and as inflections (plural, past, etc.) 28 / 61

29 / 61

The language network is a collection of words and word-parts (speech-sounds and morphemes) which are linked to each other and to the rest of cognition in a variety of ways, of which the most important is the isa relationship which classifies them and allows default inheritance 30 / 61

31 / 61

A WG analysis of an utterance is an extension of the permanent cognitive network, in which the relevant word tokens comprise a fringe of temporary concepts attached by isa links, so the utterance network has just the same formal characteristics as the permanent network A WG grammar can generate representations of actual utterances, in contrast with most other kinds of grammar which generate only idealized utterances or sentences 32 / 61

Word tokens must have different names from their types: identical labels imply identity of concept, whereas tokens and types are clearly distinct concepts. The WG convention is to reserve conventional names for types, with tokens labelled w1, w2, etc. Thus, I agree consists of w1 and w2, which isa I and AGREE:pres 33 / 61

Word types and tokens must have the same kinds of characteristics to allow inheritance. WG accommodates deviant input because the link between tokens and types is the Best Fit Principle (Miller 1992) 34 / 61

35 / 61

36 / 61

-free syntax Jo snores. NOT Jo tense snore. Exception: words that are both parts of other words and part of the syntactic structure Example: clitics Paul en mange beaucoup. Although the smallest units of syntax are always words, some words are not the smallest units of syntax but clusters of smaller words. Alternatively, we might distinguish phonological and syntactic words (Rosta 1997) 37 / 61

Inflectional morphology is responsible for mapping stems to shapes. Every lexeme has a stem, and by default a word s stem is also its shape Exceptions are due to inflectional morphology or to cliticization For example, BEHIND <behind>, DOG <dog> 38 / 61

39 / 61

40 / 61

In WG view, phrase structure is rejected In WG, syntactic structure consists of dependencies between pairs of single words A syntactic dependency is a relationship between two words that are connected by a syntactic rule The word-word dependencies form chains linking every word ultimately to the head of the phrase or sentence 41 / 61

42 / 61

A surface dependency analysis can always be translated into a phrase structure Dependency analysis is much more restrictive than phrase-structure analysis The extra richness of dependency analysis lies in the labelled dependency links, as well as in the possibility of multiple dependencies 43 / 61

Structure-sharing Structure-sharing is found when one word depends on more than one other word - i.e. when it is shared as a dependent It is the main device in WG which allows phrases to be discontinuous 44 / 61

45 / 61

In a surface structure, each dependency is licensed by the grammar network In case of structure-sharing, just one of the dependencies is drawn above the words Any of the competing dependencies could be chosen, but only one choice satisfies a well-formed surface structure (free of tangling and dangling ) The totality of dependencies constitutes the surface structure 46 / 61

Coordination Conjuncts must share their external dependencies The structure of the coordination itself ( conjuncts and coordinators ) is analyzed using word-strings simple strings of words built up by ordinary dependencies 47 / 61

48 / 61

49 / 61

WG has a compositional semantics in which each word in a sentence contributes some structure stored as its meaning These meanings are concepts defined by a network of links to other concepts: APPLE and PEAR EAT 50 / 61

Words of all word classes contribute the same kind of semantic structure, which in WG is divided into sense (general categories) and referent (the most specific individual or category referred to) similar to the contrast in morphology between stem and shape 51 / 61

52 / 61

The way in which the meanings of the words in a sentence are combined is guided by the syntax, but the semantic links are provided by the senses themselves 53 / 61

54 / 61

A word s basic sense (the one inherited from its lexeme) is modified by the word s dependents, which produces a second sense, more specific than the basic sense but more general than the referent This intermediate sense contains the meaning of the head word plus its dependent, so in effect it is the meaning of that phrase 55 / 61

In contrast with the syntax, the semantic structure contains a node for each phrase, as well as nodes for the individual words i. e., a phrase structure The step-wise composition of word meanings is called semantic phrasing 56 / 61

57 / 61

The advantage of WG for a parser is the lack of invisible words, but the dependency basis also helps by allowing each incoming word to be integrated with the words already processed without the need to build (or rebuild) higher syntactic nodes A simple algorithm guides the search for dependencies in a way that guarantees a well-formed surface structure 58 / 61

Short sentences make good examples. 1. w1 = short. No progress:- w1. 2. w2 = sentences. Capture:- w1 w2, 3. w3 = make. Capture:- W1 6 w2 w3, 4. w4 = good. No progress:- w1 w2 w3, w4. 5. w5 = examples, Capture:- w4 e w5, 6. Submit:- w1 w2 w3 (w4 ) w5 59 / 61

Hudson s theory of syntactic complexity builds on this incremental parsing model The aim of the parser is to link each word as a dependent to some other word This link can most easily be established while both words are still active in working memory Once a word has become inactive it can be reconstructed (but this is costly) 60 / 61

Thus, short links are always preferred to long ones, which enables easy calculation of the processing load for a sentence: the mean dependency distance (the number of other words between a word and the word on which it depends) In English texts, dependency links tend to be very short (70% of words are adjacent to the word on which they depend, with 10% variation in either direction) 61 / 61

Aitchison, Jean (2012). Words in the mind: An introduction to the mental lexicon. John Wiley & Sons. Brown, Dunstan et al. (1996). Russian noun stress and network morphology. In: Linguistics 34.1, pp. 53 108. Covington, Michael A (2001). A fundamental algorithm for dependency parsing. In: Proceedings of the 39th annual ACM southeast conference. Citeseer, pp. 95 102. Fraser, Norman M (1989). Parsing and dependency grammar. In: Carston, R obyn (ed.) UCL W or k-in g P a pers in L in g uistics 1, p. 29. Fraser, Norman M and Richard A Hudson (1992). in word grammar. In: Computational Linguistics 18.2, pp. 133 158. Hudson, Richard (2010). An introduction to word grammar. Cambridge University Press. (2007). Language networks: the new. Oxford University Press. 61 / 61

Hudson, Richard (1998). Word grammar. Miller, JE (1992). Richard Hudson, English. Oxford: Basil Blackwell, 1990. Pp. 0+ 445.. In: Journal of Linguistics 28.02, pp. 500 505. Pinker, Steven and Alan Prince (1988). On language and connectionism: Analysis of a parallel distributed processing model of language acquisition. In: Cognition 28.1, pp. 73 193. Rosta, Andrew (1997). English syntax and theory. PhD thesis. University College London (University of London). 61 / 61