difference in parsing accuracy Sambhav Jain, Dipti M Sharma and Rajeev Sangal Language Technologies Research Center IIIT-Hyderabad

Similar documents
Two methods to incorporate local morphosyntactic features in Hindi dependency

Ensemble Technique Utilization for Indonesian Dependency Parser

Grammar Extraction from Treebanks for Hindi and Telugu

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Using dialogue context to improve parsing performance in dialogue systems

Linking Task: Identifying authors and book titles in verbose queries

Prediction of Maximal Projection for Semantic Role Labeling

Indian Institute of Technology, Kanpur

Applications of memory-based natural language processing

Learning Computational Grammars

Survey on parsing three dependency representations for English

The stages of event extraction

Memory-based grammatical error correction

Cross-Lingual Dependency Parsing with Universal Dependencies and Predicted PoS Labels

Experiments with a Higher-Order Projective Dependency Parser

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Named Entity Recognition: A Survey for the Indian Languages

Parsing of part-of-speech tagged Assamese Texts

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

University of Alberta. Large-Scale Semi-Supervised Learning for Natural Language Processing. Shane Bergsma

LTAG-spinal and the Treebank

Accurate Unlexicalized Parsing for Modern Hebrew

A deep architecture for non-projective dependency parsing

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

Context Free Grammars. Many slides from Michael Collins

CS 598 Natural Language Processing

Developing a TT-MCTAG for German with an RCG-based Parser

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

Beyond the Pipeline: Discrete Optimization in NLP

The Strong Minimalist Thesis and Bounded Optimality

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

SEMAFOR: Frame Argument Resolution with Log-Linear Models

An Introduction to the Minimalist Program

A Graph Based Authorship Identification Approach

Hindi-Urdu Phrase Structure Annotation

The Smart/Empire TIPSTER IR System

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

The Discourse Anaphoric Properties of Connectives

Assignment 1: Predicting Amazon Review Ratings

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

arxiv: v1 [cs.cl] 2 Apr 2017

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Vocabulary Usage and Intelligibility in Learner Language

CS 446: Machine Learning

CS Machine Learning

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

ARNE - A tool for Namend Entity Recognition from Arabic Text

A Bayesian Learning Approach to Concept-Based Document Classification

Calibration of Confidence Measures in Speech Recognition

Heuristic Sample Selection to Minimize Reference Standard Training Set for a Part-Of-Speech Tagger

A Case Study: News Classification Based on Term Frequency

How to analyze visual narratives: A tutorial in Visual Narrative Grammar

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

A Vector Space Approach for Aspect-Based Sentiment Analysis

Towards a MWE-driven A* parsing with LTAGs [WG2,WG3]

Online Updating of Word Representations for Part-of-Speech Tagging

Modeling full form lexica for Arabic

A Simple Surface Realization Engine for Telugu

A Computational Evaluation of Case-Assignment Algorithms

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

A Dataset of Syntactic-Ngrams over Time from a Very Large Corpus of English Books

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

HinMA: Distributed Morphology based Hindi Morphological Analyzer

Extracting Verb Expressions Implying Negative Opinions

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Distant Supervised Relation Extraction with Wikipedia and Freebase

Building a Semantic Role Labelling System for Vietnamese

Some Principles of Automated Natural Language Information Extraction

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Control and Boundedness

AQUA: An Ontology-Driven Question Answering System

An Evaluation of POS Taggers for the CHILDES Corpus

What the National Curriculum requires in reading at Y5 and Y6

Natural Language Processing. George Konidaris

Three New Probabilistic Models. Jason M. Eisner. CIS Department, University of Pennsylvania. 200 S. 33rd St., Philadelphia, PA , USA

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

The Effect of Multiple Grammatical Errors on Processing Non-Native Writing

Training and evaluation of POS taggers on the French MULTITAG corpus

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

Domain Adaptation for Parsing

DEVELOPMENT OF A MULTILINGUAL PARALLEL CORPUS AND A PART-OF-SPEECH TAGGER FOR AFRIKAANS

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Leveraging Sentiment to Compute Word Similarity

Argument structure and theta roles

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

What Can Neural Networks Teach us about Language? Graham Neubig a2-dlearn 11/18/2017

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization

THE VERB ARGUMENT BROWSER

Transcription:

Two semantic features make all the difference in parsing accuracy Akshar Bharati, Samar Husain, Bharat Ambati, Sambhav Jain, Dipti M Sharma and Rajeev Sangal Language Technologies Research Center IIIT-Hyderabad ICON - 2008

Outline Motivation Hindi Dependency Treebank Dependency parsers Experiments General observations Conclusion ICON - 2008

Motivation Urgent need of a broad coverage Hindi parser Parser required for almost all Natural Language applications Availability of a Hindi Treebank annotated with dependency relations Results can indirectly check the consistency of treebank annotation

Hindi Dependency Treebank Hindi is a verb final language g with free word order and rich case marking. Experiments have been performed on a subset of Hyderabad Dependency Treebank (Begum et al., 2008) for Hindi. No. of sentences :1800 Average length :19.85 (words/sentence) Unique tokens :6585

Example rama ne mohana ko puswaka xi (rama ne) (mohana ko) (puswaka) (xi) Ram ERG Mohan DAT book gave Ram gave a book to Mohan xi k1 k4 k2 rama mohana puswaka

Dependency Parsers Grammar-driven parsers Parsing as a constraint-satisfaction problem Data-driven parsers Use corpus for building probabilistic models Parsers Explored (Data-driven) M ltp MaltParser MSTParser

MaltParser Version 1.0.1: Parsing algorithms: Nivre (2003) (arc-eager, arc-standard) Covington (2001) (projective, non-projective) Learning algorithms: MBL (TIMBL) SVM (LIBSVM) Feature models: Combinations of part-of-speech features, dependency type features and lexical features

MST Parser Version 0.4b: Parsing algorithms: Chu-Liu-Edmonds (1967) (non-projective) Eisner (1996) (projective) Learning algorithms: Online Large Margin Learning Feature models: Combinations of part-of-speech features, dependency type features and lexical features

Data Training set Development set Test set : 1178 Sentences : 352 Sentences : 363 Sentences Input format CONLL format Input Data Chunk heads Tagset Core tagset (24 tags), a subset of complete tagset (38 tags)

Example rama ne mohana ko puswaka xi (rama ne) (mohana ko) (puswaka) (xi) Ram ERG Mohan DAT book gave Ram gave a book to Mohan CONLL format: ID FORM LEMMA CPOSTAG POSTAG FEATS HEAD DEPREL PHEAD PDEPREL 1 rama rama NP NNP _ 4 K1 2 mohana mohana NP NNP _ 4 K4 3 puswaka puswaka NP NN _ 4 K2 4 xi xe VGF VM _ 0 main

Baselines Baseline1 A B C D E ROOT Baseline2 A B V1 C V2 ROOT NLP Winterschool 2008, IIIT-H

Results Unlabeled Attachment* Baseline1 Baseline2 46.56% 60.89% *Correct head-dependent pair, labels on the arc not considered NLP Winterschool 2008, IIIT-H

Experiments We begin with two basic hypotheses which we test while tuning the parser. For morphologically rich, free word order languages a) High performance can be achieved using vibhakti b) Subject, Object (k1, k2) can be learnt with high accuracy using vibhakti vibhakti: Generic term for preposition, postposition and suffix eg: ne in rama ne

Experiments (cont ) Experiments-I: Tuning the parser Experiments I: Tuning the parser Parsing Algorithms Model Parameters Morpho-syntactic features (FEATS) Feature set

Tuning the parser: Parsing Algorithms Malt Nivre Arc eager Arc standard Covington MST Projective Non-projective Chu-Liu-Edmonds (non-projective) Eisner (projective)

Tuning the parser: Model Parameters MST Different training-k (k highest scoring trees) K=1 to k<=20 K=5 gave the best results. Od Order Malt 1 and 2 Order 2 gave better results than order 1 Tuning SVM model was difficult. Tried various parameters but could not find any pattern CONLL shared task 2007 settings used by same parser for various languages. Turkish settings performed better than others

Results UA (UC) LA (LC) L MST Default 83.19 (40.50) 59.25 (8.54) 62.26 Non-Projective 83.94 60.72 63.3333 Algorithm, k-5 (43.53) (8.81) Default 84.44 59.10 Malt (44.63) (9.09) Arc-eager Turkish SVM settings 85.02 (42.98) 59.03 (9.09) 61.22 60.97 UA Unlabeled Attachment LA Labeled Attachment L Labeled Accuracy UC Unlabeled Complete LC Labeled Complete

Tuning the parser: Morpho-syntactic features (FEATS) F1 no feature (default) F2 TAM (Tense, Aspect and Modality) labels for verbs and postpositions for nouns F3 TAM class for verbs and postpositions for nouns.

Example rama ne mohana ko puswaka xi (rama ne) (mohana ko) (puswaka) (xi) Ram ERG Mohan DAT book gave Ram gave a book to Mohan CONLL format: ID FORM LEMMA CPOSTAG POSTAG FEATS HEAD DEPREL PHEAD PDEPREL 1 rama rama NP NNP ne 4 K1 2 mohana mohana NP NNP ko 4 K4 3 puswaka puswaka NP NN 0 4 K2 4 xi xe VGF VM ya 0 main

Tuning the parser: Feature set Default MALTParser FORM POSTAG DEPREL MSTParser Basic Uni-gram Features Parent FORM/POSTAG Child FORM/POSTAG Basic Bi-gram Features FORM/POSTAG of parent + FORM/POSTAG of child Basic Uni-gram Features + DEPREL Extended FEATS Basic Uni-gram Features + parent FEATS Basic Uni-gram Features + child FEATS Conjoined parent FEATS + DEPREL child FEATS + DEPREL

Results UA (UC) LA (LC) L MST Malt Default 83.94 (43.53) 60.72 (8.81) 63.33 Extended d 88.71 64.27 66.6767 Features (53.72) (9.09) Conjoined 88.67 69.64 Features (55.10) (14.60) Default 85.02 (42.98) 59.03 (9.09) 72.62 60.97 Extended d 87.56 67.99 70.39 Features (54.55) (14.88)

Observations Using vibhakti as a feature helps enormously Almost 10% jump in LA and 5% jump in UA Very low performance for subject, object Around 50% of data do not have explicit vibhakti. k1 k2 MST P 74.49 53.33 R 75.35 63.3838 Malt P 75.53 53.01 R 75.82 61.82

Experiments-II: New Hypothesis Subject-Object confusion cannot arise in a language for human speakers Exploring right devices which humans use p g g to disambiguate subject-object GNP (Gender, Number, Person) information Minimal semantics

GNP-1 GNP for each lexical item using morphological analyzer Appended it to FEATS column of F3. F4: F3 + GNP UA (UC) LA (LC) L MST Malt F3 88.67 (55.10) 69.64 (14.60) F4 88.03 69.10 (52.89) (14.60) F3 87.56 (54.55) 67.99 (14.88) 72.62 72.40 70.39 F4 86.92 67.17 69.82 (51.79) (16.25)

GNP-1 (cont ) GNP is important for agreement Does help in indentifying relations But agreement in Hindi is not straight- forward. Eg: verb agrees with object if subject has a post-position, position it might sometime take the default GNP Machine could not learn the selective agreement patterns k1,k2 are worst hit by this feature

GNP-2 To prove the importance of this feature in disambiguation We provide the agreement feature explicitly by marking each nodes which agrees with the verb (F5) MST Malt UA (UC) LA (LC) F3 88.67 69.6464 72.62 (55.10) (14.60) F5 88.67 70.93 (55.65) (16.80) F3 87.56 (54.55) 67.99 (14.88) F5 87.63 69.32 (55.10) (18.18) 8) L 73.98 70.39 71.86

Minimal Semantics Two basic semantic features can disambiguate majority of subject-object confusion The semantic features are human-nonhuman animate-inanimate

Minimal Semantics Data with k1 and k2 merged (k1-k2) Parsers (Malt, MST) Classifier (CRF, basic semantics as feature) Di bi t Disambiguated Parsed Output

Minimal Semantics: Results UA (UC) LA (LC) F3 88.67 69.64 MST (55.10) (14.60) FM1 89.03 (56.20) 69.93 (14.05) 72.62 72.83 F3 87.56 67.99 70.39 Malt (54.55) (14.88) FM1 87.56 (53.44) 69.28 (14.88) 71.94 L

Minimal Semantics: Results F3 F3M1 k1 k2 k1 K2 MST P 74.49 53.33 80.79 48.47 R 75.35 63.38 74.28 65.71 Malt P 75.53 53.01 76.46 48.98 R 75.82 61.82 80.42 62.6060 Captures argument structure of the verb Certain verbs take only human subjects etc.,

Conclusion Isolating crucial clues present in language which help in parsing Different Features vibhakti GNP information Minimal semantics Some hard to learn linguistic constructions

Thank you

MaltParser Malt uses arc-eager parsing algorithm. History-based feature models are used for predicting the next parser action. Support vector machines are used for mapping histories to parser actions. It uses graph transformation to handle non-projective trees.

MST Parser Formalizes dependency parsing as search for Maximum Spanning Tree (MST) in weighted directed graphs. Constructs a weighted complete directed graph for the sentence and connects all nodes to dummy root node. Constructs an MST out of this using Chu- Liu Edmonds algorithm. (Chu and Liu, 1965; Edmonds, 1967) MSTparser uses online large margin learning as the learning algorithm.

MST Parser John saw Mary. NLP Winterschool 2008, IIIT-H

MST Parser NLP Winterschool 2008, IIIT-H

General Observations MST outperforms Malt in similar conditions. Best LA result for MST and Malt are 69.64% and 67.99% respectively Malt performance can be improved by tuning the SVM parameters MST performs consistently well, in identifying the root of the tree and conjunct relations. MST is far better in identifying longer dependency arcs, whereas Malt does better with shorter ones. for distance >7 the f-measure for Malt is ~62%, for MST it is ~65%.

General Observations Overall performance of both the parsers for LA is low, 67.99% and 69.64% 64% for MST and Malt respectively. Reasons could be Training size 1200 sentences for training But training size alone is not good criteria for low performance (Hall et al., 2007) Type of tags Syntactico-semantic tags Learning such tags is difficult (Nivre et al., 2007a) Non-projectivity Data has around 10% non-projective trees.

General Observations Investigate learning issues in building a Hindi parser using a dependency treebank Discover/revalidate useful learning features To bring out specific problems (ambiguity) Can some of the problems be solved?

TAM Class TAM labels in Hindi constraint the postpositions which can appear on the subject or object TAM labels which apply similar constraints can be grouped into a class wa_he, 0_rahA_hE he Vs wa_he, ya