Psych229: Language Acquisition

Similar documents
Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Using computational modeling in language acquisition research

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Language Development: The Components of Language. How Children Develop. Chapter 6

A Case Study: News Classification Based on Term Frequency

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Evolution of Symbolisation in Chimpanzees and Neural Nets

Word learning as Bayesian inference

NAME: East Carolina University PSYC Developmental Psychology Dr. Eppler & Dr. Ironsmith

CS 598 Natural Language Processing

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

Abstractions and the Brain

Speech Recognition at ICSI: Broadcast News and beyond

Degeneracy results in canalisation of language structure: A computational model of word learning

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Language acquisition: acquiring some aspects of syntax.

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Constraining X-Bar: Theta Theory

1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Revisiting the role of prosody in early language acquisition. Megha Sundara UCLA Phonetics Lab

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

POLA: a student modeling framework for Probabilistic On-Line Assessment of problem solving performance

Python Machine Learning

Probabilistic Latent Semantic Analysis

Lecture 10: Reinforcement Learning

Language Acquisition Chart

The History of Language Teaching

Scholastic Leveled Bookroom

The Strong Minimalist Thesis and Bounded Optimality

Lecture 2: Quantifiers and Approximation

CHILDREN S POSSESSIVE STRUCTURES: A CASE STUDY 1. Andrew Radford and Joseph Galasso, University of Essex

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Age Effects on Syntactic Control in. Second Language Learning

The Good Judgment Project: A large scale test of different methods of combining expert predictions

English Language and Applied Linguistics. Module Descriptions 2017/18

Good-Enough Representations in Language Comprehension

FOREWORD.. 5 THE PROPER RUSSIAN PRONUNCIATION. 8. УРОК (Unit) УРОК (Unit) УРОК (Unit) УРОК (Unit) 4 80.

A Stochastic Model for the Vocabulary Explosion

Proof Theory for Syntacticians

a) analyse sentences, so you know what s going on and how to use that information to help you find the answer.

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Natural Language Processing. George Konidaris

A Usage-Based Approach to Recursion in Sentence Processing

Learning Methods for Fuzzy Systems

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

Unraveling symbolic number processing and the implications for its association with mathematics. Delphine Sasanguie

Arizona s English Language Arts Standards th Grade ARIZONA DEPARTMENT OF EDUCATION HIGH ACADEMIC STANDARDS FOR STUDENTS

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Construction Grammar. University of Jena.

Artificial Neural Networks written examination

Psychology and Language

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

AN ANALYSIS OF GRAMMTICAL ERRORS MADE BY THE SECOND YEAR STUDENTS OF SMAN 5 PADANG IN WRITING PAST EXPERIENCES

Houghton Mifflin Reading Correlation to the Common Core Standards for English Language Arts (Grade1)

Can Human Verb Associations help identify Salient Features for Semantic Verb Classification?

Computerized Adaptive Psychological Testing A Personalisation Perspective

UC Berkeley L2 Journal

Which verb classes and why? Research questions: Semantic Basis Hypothesis (SBH) What verb classes? Why the truth of the SBH matters

Generative Second Language Acquisition & Foreign Language Teaching Winter 2009

Using focal point learning to improve human machine tacit coordination

Quantitative analysis with statistics (and ponies) (Some slides, pony-based examples from Blase Ur)

AQUA: An Ontology-Driven Question Answering System

Software Maintenance

An Empirical and Computational Test of Linguistic Relativity

The Acquisition of English Grammatical Morphemes: A Case of Iranian EFL Learners

Learning Disability Functional Capacity Evaluation. Dear Doctor,

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Control and Boundedness

Lecture 1: Basic Concepts of Machine Learning

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

CS Machine Learning

Children s Acquisition of Syntax: Simple Models are Too Simple

Neuro-Symbolic Approaches for Knowledge Representation in Expert Systems

Context Free Grammars. Many slides from Michael Collins

Developing a TT-MCTAG for German with an RCG-based Parser

Modeling user preferences and norms in context-aware systems

A heuristic framework for pivot-based bilingual dictionary induction

How to analyze visual narratives: A tutorial in Visual Narrative Grammar

An Introduction to the Minimalist Program

Taught Throughout the Year Foundational Skills Reading Writing Language RF.1.2 Demonstrate understanding of spoken words,

Word Segmentation of Off-line Handwritten Documents

On the Formation of Phoneme Categories in DNN Acoustic Models

Generative models and adversarial training

The Smart/Empire TIPSTER IR System

Lecture 1: Machine Learning Basics

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Data Structures and Algorithms

MYCIN. The MYCIN Task

Derivational and Inflectional Morphemes in Pak-Pak Language

Linking object names and object categories: Words (but not tones) facilitate object categorization in 6- and 12-month-olds

Stages of Literacy Ros Lugg

Minimalism is the name of the predominant approach in generative linguistics today. It was first

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Ling/Span/Fren/Ger/Educ 466: SECOND LANGUAGE ACQUISITION. Spring 2011 (Tuesdays 4-6:30; Psychology 251)

Transcription:

Psych229: Language Acquisition The Standard Theory, according to Chomsky Big Questions of Language Acquisition: What constitutes knowledge of language? Lecture 18 Poverty of the Stimulus & Modeling How is this knowledge acquired? How is this knowledge used? Knowledge of language, according to Chomsky Properties of language, according to Chomsky Knowledge of language = grammar Grammar = complex set of rules and constraints that gives speakers intuitions that some sentences belong in the language while others do not Competence Hypothesis: Grammar is separate from performance factors, like dysfluencies (she said um..wrote that), errors (I bringed it), memory capacity (The boy that the dog that the cat chased bit ran home.), and statistical properties of language (frequency of transitive (Sarah ate the peach) vs. intransitive use (Sarah ate)) I think we are forced to conclude that probabilistic models give no particular insight into some of the basic problems of syntactic structure - Chomsky, 1957 Grammar is generative: it can be used to produce and comprehend an infinite number of sentences Grammar involves abstract structures: information that speakers unconsciously used is not overtly available in the observable data Grammar is modular: there are separate components with different types of representations governed by different principles Grammar is domain-specific: language exhibits properties not seen in other areas of cognition, so it cannot be the product of our general ability to think and learn 1

Language acquisition, according to Chomsky The induction problem, according to Chomsky How does a child acquire a grammar that has those properties (generative, involving abstract structures, modular, domain-specific)? Poverty of the stimulus problem: Available data insufficient to determine all these properties of the grammar. Therefore, children must bring innate knowledge to the language learning problem that guides them to the correct instantiation of grammar. Available data properties leading to this inductive problem: noisy (degenerate): sometimes there are incorrect examples in the input variable: no child s input is the same as another s, but all converge no reliable negative evidence: no labeled examples of what s not in the language no positive evidence for some generalizations: yet children still converge on them The input is too poor : what people know extends far beyond the sample of utterances in their input The input is too rich : the available data can be covered by a number of generalizations, but only some of them are the right ones (yes/no questions auxiliary inversion) Conclusion: Without innate biases, generalizations of language are unlearnable from the available data. How language is used, according to Chomsky How is the grammar used to produce and comprehend utterances in real time? Neural networks Designed to solve tasks, provide input-output mapping based on data Learning: gradual changes to the weights between units in the network that determine patterns of activation Not the focus of the generative theory. Not a grammar Parameters: learning rule that adjusts weights, network structure Grammar = higher level generalization about network behavior, abstracts away from actual implementation Grammar = computational level, network = algorithmic + implementational level 2

Neural networks Property: Can derive structural regularities from relatively noisy input. (This comes from the gradual learning capability.) Realistic learning input. Property: A network that has learned can then process novel forms. It has generative capacity. (Ex: word pronunciation) Implication: Poverty of the stimulus may not be the induction problem originally thought? We saw We saw her We saw her duck 3

Seidenberg s point: Statistical properties determine language use and neural nets provide a way to explicitly encode, acquire, and exploit this information. Children can encode statistical properties of language (Jusczyk 1997 = properties of sounds, Saffran et al. 1996 = transitional probabilities of syllables) Seidenberg s point: Acquisition is about learning to use the language, which means paying attention to its statistical properties and learning from them. Another point: Connectionist networks formalize the implementation of bootstrapping - extracting regularity from the data (used for word segmentation, word meaning, grammatical category, syntactic constructions) Big point of Seidenberg: [connectionism] attempts to explain language in terms of how is it acquired and used rather than an idealized competence grammar. The idea is not merely that competence grammar needs to incorporate statistical and probabilistic information; rather it is that the nature of language is determined by how it is acquired and used and therefore needs to be explained in terms of these functions and the brain mechanisms that support them. Such performance theories are not merely the competence theory plus some additional assumptions about acquisition and processing; the approaches begin with different goals and end up with different explanations for why languages have the properties they have. Connectionism in Action: An example where it could help Correlations between verb meaning and verb usage Hoggle loaded jewels into his bag. Hoggle loaded his bag with jewels. Hoggle poured jewels into his bag. *Hoggle poured his bag with jewels. *Hoggle filled the jewels into his bag. Hoggle filled his bag with jewels. Input is irregular - children do not get explicit examples of all of these, yet somehow come to know this paradigm. 4

Clue clusters of verbs with similar properties (if children realize this, learning is easier) load, pile, cram, spray, scatter pour, drip, slop, slosh fill, blanket, cover, coat Problem: How would the child know to cluster these verbs together if they never hear all the verbs in all the possible syntactic frames? Semantically, they re very similar. However This is a constraint satisfaction problem, which neural nets are really good at solving. Information available on groupings load, pile, cram, spray, scatter pour, drip, slop, slosh fill, blanket, cover, coat 1) How much the semantics of each verb overlaps with any other verb 2) Correlations between syntactic frames verbs appear in and the exact semantics of the verb 3) Item-specific idiosyncracies (due to language change) Connectionist net can learn the right subgroups (Allen 1997) from this information and then much easier to notice that there are syntactic usage generalizations for the groups. Therefore, this can be learned. Which is good, since it s a language-specific property. But what about learning more abstract things (like syntax) and languageindependent things that are hard (or impossible) to observe? future work for connectionist models. And innate knowledge? Innate capacities may take the form of biases or sensitivities toward particular types of information inherent in environmental events such as language, rather than a priori knowledge of grammar itself. Brain organization therefore constrains how language is learned, but the principles that govern the acquisition, representation, and use of language are not specific to this type of knowledge 5