Two hours. Question ONE is COMPULSORY UNIVERSITY OF MANCHESTER SCHOOL OF COMPUTER SCIENCE. Date: Friday 25th January 2013 Time: 14:00-16:00

Similar documents
CS 598 Natural Language Processing

Parsing of part-of-speech tagged Assamese Texts

Natural Language Processing. George Konidaris

Cross Language Information Retrieval

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Derivational: Inflectional: In a fit of rage the soldiers attacked them both that week, but lost the fight.

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Context Free Grammars. Many slides from Michael Collins

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Books Effective Literacy Y5-8 Learning Through Talk Y4-8 Switch onto Spelling Spelling Under Scrutiny

1/20 idea. We ll spend an extra hour on 1/21. based on assigned readings. so you ll be ready to discuss them in class

Prediction of Maximal Projection for Semantic Role Labeling

Compositional Semantics

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Constraining X-Bar: Theta Theory

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Derivational and Inflectional Morphemes in Pak-Pak Language

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

Linking Task: Identifying authors and book titles in verbose queries

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Grammars & Parsing, Part 1:

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Universiteit Leiden ICT in Business

A Bayesian Learning Approach to Concept-Based Document Classification

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Some Principles of Automated Natural Language Information Extraction

A Domain Ontology Development Environment Using a MRD and Text Corpus

Memory-based grammatical error correction

What the National Curriculum requires in reading at Y5 and Y6

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

LING 329 : MORPHOLOGY

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

A Graph Based Authorship Identification Approach

a) analyse sentences, so you know what s going on and how to use that information to help you find the answer.

Ensemble Technique Utilization for Indonesian Dependency Parser

Accurate Unlexicalized Parsing for Modern Hebrew

The stages of event extraction

The Role of the Head in the Interpretation of English Deverbal Compounds

Using dialogue context to improve parsing performance in dialogue systems

CEFR Overall Illustrative English Proficiency Scales

Inleiding Taalkunde. Docent: Paola Monachesi. Blok 4, 2001/ Syntax 2. 2 Phrases and constituent structure 2. 3 A minigrammar of Italian 3

The Evolution of Random Phenomena

The Smart/Empire TIPSTER IR System

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

An Introduction to the Minimalist Program

Control and Boundedness

Modeling full form lexica for Arabic

Ch VI- SENTENCE PATTERNS.

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

Developing a TT-MCTAG for German with an RCG-based Parser

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

Analysis of Probabilistic Parsing in NLP

Words come in categories

Citation for published version (APA): Veenstra, M. J. A. (1998). Formalizing the minimalist program Groningen: s.n.

LTAG-spinal and the Treebank

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

The Interface between Phrasal and Functional Constraints

Construction Grammar. University of Jena.

Three New Probabilistic Models. Jason M. Eisner. CIS Department, University of Pennsylvania. 200 S. 33rd St., Philadelphia, PA , USA

AQUA: An Ontology-Driven Question Answering System

National Literacy and Numeracy Framework for years 3/4

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

A Case Study: News Classification Based on Term Frequency

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

ScienceDirect. Malayalam question answering system

"f TOPIC =T COMP COMP... OBJ

Underlying and Surface Grammatical Relations in Greek consider

1.2 Interpretive Communication: Students will demonstrate comprehension of content from authentic audio and visual resources.

Argument structure and theta roles

Specifying a shallow grammatical for parsing purposes

Basic Syntax. Doug Arnold We review some basic grammatical ideas and terminology, and look at some common constructions in English.

Heuristic Sample Selection to Minimize Reference Standard Training Set for a Part-Of-Speech Tagger

RANKING AND UNRANKING LEFT SZILARD LANGUAGES. Erkki Mäkinen DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF TAMPERE REPORT A ER E P S I M S

Corpus Linguistics (L615)

Unit 8 Pronoun References

ARNE - A tool for Namend Entity Recognition from Arabic Text

BULATS A2 WORDLIST 2

Theoretical Syntax Winter Answers to practice problems

Impact of Controlled Language on Translation Quality and Post-editing in a Statistical Machine Translation Environment

LNGT0101 Introduction to Linguistics

Houghton Mifflin Reading Correlation to the Common Core Standards for English Language Arts (Grade1)

Part III: Semantics. Notes on Natural Language Processing. Chia-Ping Chen

The College Board Redesigned SAT Grade 12

Beyond the Pipeline: Discrete Optimization in NLP

Applications of memory-based natural language processing

Writing a composition

Performance Analysis of Optimized Content Extraction for Cyrillic Mongolian Learning Text Materials in the Database

Feature-Based Grammar

Transcription:

Two hours Question ONE is COMPULSORY UNIVERSITY OF MANCHESTER SCHOOL OF COMPUTER SCIENCE Natural Language Systems Date: Friday 25th January 2013 Time: 14:00-16:00 Please answer Question ONE in Section A and TWO Questions from Section B. This is a CLOSED book examination The use of electronic calculators is permitted provided they are not programmable and do not store text. [PTO]

Section A You should answer question 1: each part of this question carries 5 marks 1. a) How do dictionary-based tagging and affix-based tagging work? [3 marks] How would you use backing-off as a strategy for combining them? [2 marks] b) Explain the terms precision, recall and F-measure. [3 marks] Suppose you had two taggers, where one assigned the tag NN to every word and the other assigned the tag DT to the word the and left every other word untagged. What can you say about the precision and recall of these two taggers? [2 marks] c) Regular expressions can be used to describe simple grammatical patterns, but they cannot be used for describing recursive structures. Explain briefly why it is not possible to write a regular expression to capture the fact that an NP can form part of a determiner phrase (as in the old man s wife s father, where the old man is part of the determiner phrase the old man s, and the old man s wife is part of the determiner phrase the old man s wife s ). [4 marks] How does using cascaded regular expressions help you get round this problem? [2 marks] d) What is the term frequency-inverse document frequency (TF-IDF) score of a word? Why is it useful? e) Explain why lexical ambiguity is a greater problem than structural ambiguity for machine translation systems. [2 marks] What kind of information might you be able to extract from a corpus to help with lexical ambiguity? [3 marks] f) What is assimilation? Is it a greater problem for speech synthesis or speech recognition? Page 2 of 7

Section B Answer two questions from this section. Each question carries 35 marks. 2. a) Describe the difference between inflectional and derivational morphology, illustrating your answer with examples of each. [6 marks] Which of these would you be more concerned with if you were developing a machine translation system? [4 marks] b) The presence of spelling changes that reflect changes in pronunciation causes problems for programs that carry out morphological analysis. Explain briefly why the plural of the noun kiss is written as kisses, and why the past tense form of change is changed. [6 marks] c) Consider the following English spelling rules, given in the format used in the lectures: [] ==> [e] : [c0] _ [v0]. [e] ==> []: [c0, c1] _ [c2]. [c0] ==> [] : [v0] _ [c0]. Which of the following surface forms are related to their constituent parts by these rules (where a surface form is derived by one of the rules you must specify which one): kisses ==> kiss + s, changing ==> change + ing, seen ==> see + en, clapping ==> clap + ing, calling ==> call + ing, watches ==> watch + s, lived ==> live + ed? [7 marks] d) Show how the categorial descriptions of roots and affixes in Figure 1 support analyses of the words chanter, chantez, chantons, chanterez, chanterons, chantiez, chantions given the basic rules of categorial combination in Figure2. [7 marks] chant ==> a/b. i ==> b/c. er ==> b. ez ==> c. ==> b/c. ons ==> c. er ==> b/c. Figure 1: Roots and affixes Show how the extended categorial rules in Figure3 will let you analyse these words from left to right. What are the advantages of carrying out morphological analysis in a strictly left to right order? [5 marks] Page 3 of 7 [PTO]

X ==> X/Y, Y. X ==> Y, X\Y. Figure 2: Basic categorial rules X/Z ==> X/Y, Y/Z. X\Z ==> Y\Z, X\Y. Figure 3: Extended categorial rules 3. a) Almost all natural language systems include a component which tries to find the relations between words a parser. Traditional systems use a hand-crafted grammar, which is intended to capture the constraints that allow a native speaker to decide whether some sequence of words is a well-formed sentence of their language. More recent systems attempt to infer these rules from sample data, usually in the form of collections of dependency trees. Discuss the advantages and disadvantages of both approaches. [9 marks] b) Describe the major data structures and the four main operations used in Nivre s deterministic parsing algorithm. [7 marks] Give a brief account of the time complexity of this algorithm. [4 marks] c) Show the sequence of operations that would obtain the dependency tree in Figure 4 from the sentence I saw a man in it. [10 marks] in it man saw a I Figure 4: Dependency tree for I saw a man in it d) Explain how machine learning can be used to obtain rules which will choose which action to perform at each stage. Show a rule that might be extracted from the sequence of actions that you carried out in part (3c) for deciding to attach a determiner to a noun. [5 marks] Page 4 of 7

4. a) What is a vector space model of lexical meaning? [5 marks] What is the role of the context in such models? What kinds of things can be used as contexts? [5 marks] b) Consider the following passages: using a window of the two preceding and two following words as the context, determine whether mouse is more similar to hockey or to rodent. [8 marks]. What are the advantages and disadvantages of using this notion of context when compared to using the entire document in which a word appears and to using its syntactic parent. [10 marks] A mouse is a small mammal belonging to the order of rodents, characteristically having a pointed snout, small rounded ears, and a long naked or almost hairless tail. The best known mouse species is the common house mouse. In some places, certain kinds of field mice are also common. Field hockey is a team sport in which a team of players attempts to score goals by hitting, pushing or flicking a ball into an opposing team s goal using sticks. In some countries, it is known simply as hockey ; however, the name field hockey is used in countries in which the word hockey is generally reserved for another form of hockey, such as ice hockey, street hockey or roller hockey. c) How might you use a vector space model to realise that a word had multiple meanings, and how could you then decide which interpretation was intended in a given context? [7 marks] Page 5 of 7 [PTO]

5. a) Describe phone-based and diphone-based speech synthesis, and explain why diphonebased synthesisers usually produce better quality output than phone-based synthesisers. [8 marks] What parameters do you need to control in order to produce natural sounding speech, and how would you find appropriate values to choose for these parameters? [7 marks] b) Most speech recognition systems use a non-recursive context-free grammar (NR- CFG) to constrain what can possibly be said, and a hidden Markov model (HMM) to work out the most likely interpretation of a given speech sequence. Describe the relationship between the NR-CFG and the HMM, and explain what emission probabilities and transition probabilities are. You should use the NR-CFG in Figure 5 and the HMM in Figure 6 (overleaf) to illustrate your answer. [10 marks] $npsubj = he she; $verb = loves hates; $npobj = it them; $vp = $verb $npobj; ($npsubj $vp) Figure 5: Simple CFG c) Find the most likely path through the HMM in Figure 6 (answers that approximate calculations to one significant figure are acceptable, e.g. using 0.9 0.7 0.6 rather than 0.9 0.7 = 0.63). [10 marks] Page 6 of 7

0.7 0.7 he (0.0) 0.7 loves (0.0) 0.8 it (0.0) she (0.0) 0.3 hates (0.0) 0.2 them (0.0) 0.3 0.3 Figure 6: Simple HMM Page 7 of 7 END OF EXAMINATION