NATURAL LANGUAGE PROCESSING. Dr. G. Bharadwaja Kumar

Similar documents
2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

BULATS A2 WORDLIST 2

Ch VI- SENTENCE PATTERNS.

Developing Grammar in Context

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Words come in categories

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Unit 8 Pronoun References

CS 598 Natural Language Processing

Context Free Grammars. Many slides from Michael Collins

BASIC ENGLISH. Book GRAMMAR

Thornhill Primary School - Grammar coverage Year 1-6

Written by: YULI AMRIA (RRA1B210085) ABSTRACT. Key words: ability, possessive pronouns, and possessive adjectives INTRODUCTION

First Grade Curriculum Highlights: In alignment with the Common Core Standards

Word Stress and Intonation: Introduction

Taught Throughout the Year Foundational Skills Reading Writing Language RF.1.2 Demonstrate understanding of spoken words,

Reading Grammar Section and Lesson Writing Chapter and Lesson Identify a purpose for reading W1-LO; W2- LO; W3- LO; W4- LO; W5-

ELD CELDT 5 EDGE Level C Curriculum Guide LANGUAGE DEVELOPMENT VOCABULARY COMMON WRITING PROJECT. ToolKit

An Evaluation of POS Taggers for the CHILDES Corpus

Houghton Mifflin Reading Correlation to the Common Core Standards for English Language Arts (Grade1)

Training and evaluation of POS taggers on the French MULTITAG corpus

Campus Academic Resource Program An Object of a Preposition: A Prepositional Phrase: noun adjective

Adjectives tell you more about a noun (for example: the red dress ).

Construction Grammar. University of Jena.

Dear Teacher: Welcome to Reading Rods! Reading Rods offer many outstanding features! Read on to discover how to put Reading Rods to work today!

SAMPLE. Chapter 1: Background. A. Basic Introduction. B. Why It s Important to Teach/Learn Grammar in the First Place

Sample Goals and Benchmarks

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Part I. Figuring out how English works

Probabilistic Latent Semantic Analysis

Oakland Unified School District English/ Language Arts Course Syllabus

Indian Institute of Technology, Kanpur

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Emmaus Lutheran School English Language Arts Curriculum

Heuristic Sample Selection to Minimize Reference Standard Training Set for a Part-Of-Speech Tagger

Programma di Inglese

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Intensive English Program Southwest College

Formulaic Language and Fluency: ESL Teaching Applications

Constraining X-Bar: Theta Theory

Lecture 10: Reinforcement Learning

Grammars & Parsing, Part 1:

5 th Grade Language Arts Curriculum Map

4 th Grade Reading Language Arts Pacing Guide

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

Grade 7. Prentice Hall. Literature, The Penguin Edition, Grade Oregon English/Language Arts Grade-Level Standards. Grade 7

Opportunities for Writing Title Key Stage 1 Key Stage 2 Narrative

Chapter 9 Banked gap-filling

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Introducing the New Iowa Assessments Language Arts Levels 15 17/18

AN ANALYSIS OF GRAMMTICAL ERRORS MADE BY THE SECOND YEAR STUDENTS OF SMAN 5 PADANG IN WRITING PAST EXPERIENCES

Senior Stenographer / Senior Typist Series (including equivalent Secretary titles)

What the National Curriculum requires in reading at Y5 and Y6

Loughton School s curriculum evening. 28 th February 2017

Writing a composition

Argument structure and theta roles

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Advanced Grammar in Use

California Department of Education English Language Development Standards for Grade 8

ELA/ELD Standards Correlation Matrix for ELD Materials Grade 1 Reading

- Period - Semicolon - Comma + FANBOYS - Question mark - Exclamation mark

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

FOREWORD.. 5 THE PROPER RUSSIAN PRONUNCIATION. 8. УРОК (Unit) УРОК (Unit) УРОК (Unit) УРОК (Unit) 4 80.

A Computational Evaluation of Case-Assignment Algorithms

DIRECT AND INDIRECT SPEECH

GRADE 1 GRAMMAR REFERENCE GUIDE Pre-Unit 1: PAGE 1 OF 21

Derivational and Inflectional Morphemes in Pak-Pak Language

Myths, Legends, Fairytales and Novels (Writing a Letter)

The College Board Redesigned SAT Grade 12

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

Copyright 2002 by the McGraw-Hill Companies, Inc.

Mercer County Schools

Parsing of part-of-speech tagged Assamese Texts

The Evolution of Random Phenomena

a) analyse sentences, so you know what s going on and how to use that information to help you find the answer.

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

L1 and L2 acquisition. Holger Diessel

LEXICAL COHESION ANALYSIS OF THE ARTICLE WHAT IS A GOOD RESEARCH PROJECT? BY BRIAN PALTRIDGE A JOURNAL ARTICLE

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

South Carolina College- and Career-Ready Standards for Mathematics. Standards Unpacking Documents Grade 5

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

Set up your desk: Do Now Share-Out 1. Do Now Share-Out 2. Rule the World Essay 10/11/2012

Defragmenting Textual Data by Leveraging the Syntactic Structure of the English Language

Correlated GRADE. Congratulations on your purchase of some of the finest teaching materials in the world. to State Standards

A Comparison of Two Text Representations for Sentiment Analysis

Participate in expanded conversations and respond appropriately to a variety of conversational prompts

Appendix D IMPORTANT WRITING TIPS FOR GRADUATE STUDENTS

Lesson objective: Year: 5/6 Resources: 1a, 1b, 1c, 1d, 1e, 1f, Examples of newspaper orientations.

Today we examine the distribution of infinitival clauses, which can be

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Virtually Anywhere Episodes 1 and 2. Teacher s Notes

TABE 9&10. Revised 8/2013- with reference to College and Career Readiness Standards

Course Outline for Honors Spanish II Mrs. Sharon Koller

More ESL Teaching Ideas

arxiv:cmp-lg/ v1 7 Jun 1997 Abstract

Transcription:

NATURAL LANGUAGE PROCESSING Dr. G. Bharadwaja Kumar

PARTS OF SPEECH The parts of speech explain how a word is used in a sentence. Based on their usage and functionality words are categorized into several types or parts of speech.

Words having major parts of speech contribute for meaning to a greater extent, and hence are sometimes called content words Nouns, verbs, adjectives, and adverbs are content parts of speech. Function words are words that exist to explain or create grammatical or structural relationships into which the content words may fit. Pronouns, prepositions, conjunctions, determiners, qualifiers/intensifiers, and interrogatives are some function parts of speech.

Use of POS Tagging

Nouns This part of a speech refers to words that are used to name persons, things, animals, places, ideas, or events. Tom Hanks is very versatile. The italicized noun refers to a name of a person. Dogs can be extremely cute. In this example, the italicized word is considered a noun because it names an animal. It is my birthday. The word birthday is a noun which refers to an event.

Noun Subcategories Proper proper nouns always start with a capital letter and refers to specific names of persons, places, or things. Examples: Volkswagen Beetle, Shakey s Pizza, Game of Thrones Common common nouns are the opposite of proper nouns. These are just generic names of persons, things, or places. Examples: car, pizza parlor, TV series Concrete this kind refers to nouns which you can perceive through your five senses. Examples: folder, sand, board Abstract- unlike concrete nouns, abstract nouns are those which you can t perceive through your five senses. Examples: happiness, grudge, bravery

Count it refers to anything that is countable, and has a singular and plural form. Examples: kitten, video, ball Mass this is the opposite of count nouns. Mass nouns are also called noncountable nouns, and they need to have counters to quantify them. Examples of Counters: kilo, cup, meter Examples of Mass Nouns: rice, flour, garter Collective refers to a group of persons, animals, or things. Example: faculty (group of teachers), class (group of students), pride (group of lions)

Pronoun A pronoun is a part of a speech which functions as a replacement for a noun. Some examples of pronouns are: I, it, he, she, mine, his, hers, we, they, theirs, and ours. Sample Sentences: Janice is a very stubborn child. She just stared at me and when I told her to stop. The largest slice is mine. We are number one.

Adjectives This part of a speech is used to describe a noun or a pronoun. Adjectives can specify the quality, the size, and the number of nouns or pronouns. The carvings are intricate. The italicized word describes the appearance of the noun carvings. I have two hamsters. The italicized word two, is an adjective which describes the number of the noun hamsters. Wow! That doughnut is huge! The italicized word is an adjective which describes the size of the noun doughnut.

Conjuctions The conjunction is a part of a speech which joins words, phrases, or clauses together. Examples of Conjunctions: and, yet, but, for, nor, or, and so Sample Sentences: This cup of tea is delicious and very soothing. Kiyoko has to start all over again because she didn t follow the professor s instructions. Homer always wanted to join the play, but he didn t have the guts to audition. The italicized words in the sentences above are some examples of conjunctions.

Verbs This is the most important part of a speech, for without a verb, a sentence would not exist. Simply put, this is a word that shows an action (physical or mental) or state of being of the subject in a sentence. Examples of State of Being Verbs : am, is, was, are, and were Sample Sentences: As usual, the Stormtroopers missed their shot. The italicized word expresses the action of the subject Stormtroopers. They are always prepared in emergencies. The verb are refers to the state of being of the pronoun they, which is the subject in the sentence.

Adverb Just like adjectives, adverbs are also used to describe words, but the difference is that adverbs describe adjectives, verbs, or another adverb. The different types of adverbs are: Adverb of Manner this refers to how something happens or how an action is done. Example: Annie danced gracefully. The word gracefully tells how Annie danced. Adverb of Time- this states when something happens or when it is done. Example: She came yesterday. The italicized word tells when she came.

Adverb of Place this tells something about where something happens or where something is done. Example: Of course, I looked everywhere! The adverb everywhere tells where I looked. Adverb of Degree this states the intensity or the degree to which a specific thing happens or is done. Example: The child is very talented. The italicized adverb answers the question, To what degree is the child talented?

Prepositions This part of a speech basically refers to words that specify location or a location in time. Examples of Prepositions: above, below, throughout, outside, before, near, and since Sample Sentences: Micah is hiding under the bed. The italicized preposition introduces the prepositional phrase under the bed, and tells where Micah is hiding. During the game, the audience never stopped cheering for their team. The italicized preposition introduces the prepositional phrase during the game, and tells when the audience cheered.

Interjections This part of a speech refers to words which express emotions. Since interjections are commonly used to convey strong emotions, they are usually followed by an exclamation point. Sample Sentences: Ouch! That must have hurt. Hurray, we won! Hey! I said enough!

Corpus Alignment

Sample rules N-IP rule: A tag N (noun) cannot be followed by a tag IP (interrogative pronoun)... man who man: {N} who: {RP, IP} --> {RP} relative pronoun ART-V rule: A tag ART (article) cannot be followed by a tag V (verb)...the book the: {ART} book: {N, V} --> {N}

How TBL Rules are Learned We will assume that we have a tagged corpus. Brill s TBL algorithm has three major steps. Tag the corpus with the most likely tag for each (unigram model) Choose a transformation that deterministically replaces an existing tag with a new tag such that the resulting tagged training corpus has the lowest error rate out of all transformations. Apply the transformation to the training corpus. These steps are repeated until a stopping criterion is reached. The result (which will be our tagger) will be: First tags using most-likely tags Then apply the learned transformations

Strengths of transformation-based tagging exploits a wider range of lexical and syntactic regularities can look at a wider context condition the tags on preceding/next words not just preceding tags. can use more context than bigram or trigram. transformation rules are easier to understand

Stochastic POS tagging Stochastic (=Probabilistic) tagging Assume that a word s tag only depends on the previous tags (not following ones) Use a training set (manually tagged corpus) to: learn the regularities of tag sequences learn the possible tags for a word model this info through a Markov process

Hidden Markov Model For Markov chains, the output symbols are the same as the states. See sunny weather: we re in state sunny But in part-of-speech tagging (and other things) The output symbols are words But the hidden states are part-of-speech tags So we need an extension! A Hidden Markov Model is an extension of a Markov chain in which the output symbols are not the same as the states. This means we don t know which state we are in. Lecture 1, 7/21/2005 Natural Language Processing 33

Hidden Markov Model (HMM) Taggers Goal: maximize P(word tag) x P(tag previous n tags) Lexical information Syntagmatic information P(word tag) word/lexical likelihood probability that given this tag, we have this word NOT probability that this word has this tag modeled through language model (word-tag matrix) P(tag previous n tags) tag sequence likelihood probability that this tag follows these previous tags modeled through language model (tag-tag matrix)

Efficient Tagging How to find the most likely sequence of tags for a sequence of words Given the contextual and lexical estimates, we can use the Viterbi algorithm to avoid using the brute force method, which for N tags and T words examines N T sequences. 35

For "Flies like a flower", there are four words and four possible tags, giving 256 sequences depicted below. In a brute force method, all of them would be examined. 36

Viterbi Notation To track the probability of the best sequence leading to each possible tag at each position, the algorithm uses, an N n array, where N is the number of tags and n is the number of words in the sentence. t (t i ) records the probability of the best sequence up to position t that ends with the tag, t i. To record the actual best sequence, it suffices to record only the one preceding tag for each tag and position. Hence, another array, an N n array, is used. t (t i ) indicates for the tag t i in position t which tag at position t-1 is in the best sequence. 37

Viterbi Algorithm Given the word sequence w 1,n, the lexical tags t 1,N, the lexical probabilities P(w t t t ), and the bigram probabilities P(t i t j ), find the most likely sequence of lexical tags for the word sequence. Initialization Step: For i= 1 to N do // For all tag states t 1,N 1 (t i ) = P(w 1 t i ) P(t i ø) 1 (t i ) = 0 // Starting point 38

Viterbi Algorithm Iteration Step: For f=2 to n // next word index For i= 1 to N // tag states t 1,N f (t i ) = max j=1,n ( f-1 (t j ) P(t i t j )) P(w f t i )) f (t i ) = argmax j=1,n ( f-1 (t j ) P(t i t j )) P(w f t i )) //index that gave max Sequence Identification Step: X n = argmax j=1,n n (t j ) For i = n-1 to 1 do X i = i+1 (X i+1 ) // Get the best ending tag state for w n // Get the rest // Use the back pointer from subsequent state P(X 1,, X n ) = max j=1,n n (t j ) 39

Example 40

Second Iteration Step 41

Final Iteration Now we have to backtrack to get the best sequence Flies N like V a ART flower N 42

HMM Training

HMM Learning: Supervised