Incremental Input Stream Segmentation for Real-time NLP Applications

Similar documents
Linking Task: Identifying authors and book titles in verbose queries

Role of Pausing in Text-to-Speech Synthesis for Simultaneous Interpretation

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

AQUA: An Ontology-Driven Question Answering System

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment

Ensemble Technique Utilization for Indonesian Dependency Parser

Applications of memory-based natural language processing

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Beyond the Pipeline: Discrete Optimization in NLP

The Smart/Empire TIPSTER IR System

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

CS 598 Natural Language Processing

Atypical Prosodic Structure as an Indicator of Reading Level and Text Difficulty

Using dialogue context to improve parsing performance in dialogue systems

arxiv: v1 [cs.cl] 2 Apr 2017

Prediction of Maximal Projection for Semantic Role Labeling

Parsing of part-of-speech tagged Assamese Texts

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Natural Language Processing. George Konidaris

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Towards a Machine-Learning Architecture for Lexical Functional Grammar Parsing. Grzegorz Chrupa la

Grammars & Parsing, Part 1:

Accurate Unlexicalized Parsing for Modern Hebrew

Proof Theory for Syntacticians

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Compositional Semantics

Speech Recognition at ICSI: Broadcast News and beyond

CS Machine Learning

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

An Introduction to the Minimalist Program

On Human Computer Interaction, HCI. Dr. Saif al Zahir Electrical and Computer Engineering Department UBC

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Rule Learning With Negation: Issues Regarding Effectiveness

The Discourse Anaphoric Properties of Connectives

An Interactive Intelligent Language Tutor Over The Internet

The stages of event extraction

Word Stress and Intonation: Introduction

Software Maintenance

Developing a TT-MCTAG for German with an RCG-based Parser

BYLINE [Heng Ji, Computer Science Department, New York University,

Improved Reordering for Shallow-n Grammar based Hierarchical Phrase-based Translation

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

The Interface between Phrasal and Functional Constraints

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Rule Learning with Negation: Issues Regarding Effectiveness

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

The Strong Minimalist Thesis and Bounded Optimality

Some Principles of Automated Natural Language Information Extraction

What Can Neural Networks Teach us about Language? Graham Neubig a2-dlearn 11/18/2017

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

A Framework for Customizable Generation of Hypertext Presentations

Word Segmentation of Off-line Handwritten Documents

Organizing Comprehensive Literacy Assessment: How to Get Started

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization

The presence of interpretable but ungrammatical sentences corresponds to mismatches between interpretive and productive parsing.

The Role of the Head in the Interpretation of English Deverbal Compounds

The MSR-NRC-SRI MT System for NIST Open Machine Translation 2008 Evaluation

LTAG-spinal and the Treebank

Probabilistic Latent Semantic Analysis

First Grade Curriculum Highlights: In alignment with the Common Core Standards

Multilingual Sentiment and Subjectivity Analysis

Learning Methods in Multilingual Speech Recognition

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

Copyright and moral rights for this thesis are retained by the author

Learning Computational Grammars

DRAFT VERSION 2, 02/24/12

Loughton School s curriculum evening. 28 th February 2017

What is a Mental Model?

A Vector Space Approach for Aspect-Based Sentiment Analysis

Noisy SMS Machine Translation in Low-Density Languages

TINE: A Metric to Assess MT Adequacy

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Extracting Verb Expressions Implying Negative Opinions

Multi-Lingual Text Leveling

How to analyze visual narratives: A tutorial in Visual Narrative Grammar

Myths, Legends, Fairytales and Novels (Writing a Letter)

arxiv: v1 [cs.cv] 10 May 2017

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Highlighting and Annotation Tips Foundation Lesson

Building a Semantic Role Labelling System for Vietnamese

Adapting Stochastic Output for Rule-Based Semantics

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

University of Alberta. Large-Scale Semi-Supervised Learning for Natural Language Processing. Shane Bergsma

A First-Pass Approach for Evaluating Machine Translation Systems

What the National Curriculum requires in reading at Y5 and Y6

A Domain Ontology Development Environment Using a MRD and Text Corpus

MYP Language A Course Outline Year 3

Agnès Tutin and Olivier Kraif Univ. Grenoble Alpes, LIDILEM CS Grenoble cedex 9, France

Transcription:

Incremental Input Stream Segmentation for Real-time NLP Applications Mahsa Yarmohammadi Streaming NLP for Big Data Class SBU Computer Science Department 9/29/2016

Outline Introduction Simultaneous speech-to-speech translation (SST) as an example of a real-time NLP application Current input stream segmentation strategies Incorporating syntax into input stream segmentation A novel partial parsing approach Hedge parsing Impact of hedge parsing in MT and SST

Introduction Simultaneous speech-to-speech translation (SST): Listening to source language speech, and at the same time, producing target language speech. Challenges of SST, and real-time systems: no later revisions of mistakes little latency in delivering the output after receiving the input process parts of the input, even before it has been completed segment continuous stream input data to the appropriate units

SST Pipeline

Input Segmentation SST requires segments of the stream input that: are separated at appropriate positions are non-overlapping could be processed sequentially Granularity of segments impacts translation latency/acc. shorter segments are typically delivered more quickly shorter segments are typically processed more quickly shorter segments will likely result in inferior translation accuracy

Input Segmentation Previous work on SST mainly focused on Pauses in the speech The location of comma or period in the transcribed text Combined punctuation-based and length-based methods Joint segmentation and translation optimization

Input Segmentation Fϋgen et al. (2007) Baseline: sentence boundaries 36.6% BLEU score by translating ASR reference transcripts, 33.4% by translating ASR hypotheses avg sentence length: 30 words Automatically predicted punctuation marks similar BLEU scores as above, avg segment length: 9 words Every n words n=7, 30.1% BLEU for ASR reference, 27.5% BLEU for ASR hypothesis can destroy semantic context Non-speech duration of 0.3 seconds 32.6% BLEU score for ASR hypotheses + lexical features 32.9% BLEU score, avg segment length: 9 words

Input Segmentation Rangarajan Sridhar et al. (2013) Non-linguistic and linguistic segmentation strategies Every n words larger n values: good translation accuracy, but high latency Optimal word alignment occurs only within segments poor translation due to short segments (2-4 words) Sentences, or comma-separated segments automatically predicted by an SVM classifier performs the best, but the classifier introduces a significant delay Four segment types of noun, verb, particle, and adverbial poor translation, mainly due to short segments

Input Segmentation Matusov et al. (2007) automatic sentence boundary and sub-sentence punctuation prediction the best translation achieved when boundary detection algorithms were directly optimized for translation quality Cettolo and Federico (2006) punctuation-based, length-based, and combined text segmentation criteria the best performance achieved by combining both linguistic and input length constraints

Syntax-based Segmentation and Annotation Human interpreters depend on info. of a structural nature the input segmentation follows mainly syntactical principles Syntactic annotations in the input segments could potentially improve the performance of SST syntactic annotations can be helpful in regular (non-incremental) translation (Mi et al., 2008;Liu et al., 2011;Zhang et al., 2011;Tamura et al., 2013)

Incremental Syntactic Analysis Applying syntactic info in real-time scenarios is challenging Conventional full syntactic parsing: is not directly applicable to sub-sentential segments builds fully connected structures over the entire string is generally computationally expensive A fast partial syntactic parsing of the input should be considered

A Novel Partial Parsing Approach Propose a novel partial parsing method for fast and incremental syntactic analysis of the input that: 1) less computationally demanding than a full parser but more effective than a shallow parser 2) allows for syntax-based segmentation, and 3) incorporates some degree of syntax without requiring the entire sentence

Full Syntactic Parsing Full parsing gives a complex complete parse tree of the sentence hierarchically embedded structures, recursive phrase construction great expressive power but computationally expensive

Partial Syntactic Parsing Shallow parsing (chunking) identifies flat, non-overlapping constituents the chunks lack hierarchical structures very fast and efficient, but not powerful enough to define recursive phrases

Syntactic Parsing Shallow parsing Full parsing efficiency complexity flat bracketing structures fully recursive structures

Syntactic Parsing Shallow parsing Some partial parsing? Full parsing efficiency complexity flat bracketing structures portions of recursive structures fully recursive structures

Syntactic Parsing Shallow parsing Hedge parsing Full parsing flat bracketing structures fully recursive structures fully recursive structures for constituents covering < L words

Hedge Transform Preserving every constituent of length up to some span L

Hedge Transform Preserving every constituent of length up to some span L span=2 span=3

Hedge Transform Preserving every constituent of length up to some span L span=2 span=5 span=3

Hedge Transform Preserving every constituent of length up to some span L span=1 span=1 span=2 span=5 span=3

Hedge Transform Constituents of span > L are recursively removed, children are attached to the parent example: L=4

Hedge Transform Constituents of span > L are recursively removed, children are attached to the parent example: L=4

Hedge Transform Constituents of span > L are recursively removed, children are attached to the parent example: L=4

Hedge Transform Hedges are sequentially connected to the top-most node, allowing for sentence segmentation before parsing

Hedge Parsing in MT Impact of hedge parsing in machine translation (MT): (1) How does augmenting a translation model with hedge syntax affect a regular (non-incremental) translation? compared to no syntax shallow syntax full syntax (2) How does hedge segmentation of the input affect the latency/acc trade-off in an incremental translation? compared to raw segments non-linguistic syntax shallow syntax

Hedge Parsing in MT In summary, the results show: significant improvement in translation quality by using hedgesyntax on the target side of the translation model compared to shallow- or no-syntax comparable to the performance of a full-syntax model hedge-syntax on the source side of the translation model falls behind full syntax although again outperforms shallow syntax hedge parsing of the inputs resulted in an acceptable accuracy/ latency trade-off in simultaneous translation, notably outperforming shallow syntax

Thank You! Questions?

SST Pipeline