Joint Modeling of Content and Discourse Relations in Dialogues

Similar documents
Assignment 1: Predicting Amazon Review Ratings

A Vector Space Approach for Aspect-Based Sentiment Analysis

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Speech Emotion Recognition Using Support Vector Machine

Probabilistic Latent Semantic Analysis

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Lecture 1: Machine Learning Basics

Linking Task: Identifying authors and book titles in verbose queries

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Generative models and adversarial training

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Dialog Act Classification Using N-Gram Algorithms

Modeling function word errors in DNN-HMM based LVCSR systems

CS Machine Learning

Python Machine Learning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Human Emotion Recognition From Speech

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Modeling function word errors in DNN-HMM based LVCSR systems

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Switchboard Language Model Improvement with Conversational Data from Gigaword

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

A study of speaker adaptation for DNN-based speech synthesis

Knowledge based expert systems D H A N A N J A Y K A L B A N D E

Beyond the Pipeline: Discrete Optimization in NLP

arxiv: v1 [cs.cv] 10 May 2017

CS 598 Natural Language Processing

A Comparison of Two Text Representations for Sentiment Analysis

Learning Methods in Multilingual Speech Recognition

BYLINE [Heng Ji, Computer Science Department, New York University,

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Detecting English-French Cognates Using Orthographic Edit Distance

Using dialogue context to improve parsing performance in dialogue systems

Computerized Adaptive Psychological Testing A Personalisation Perspective

Reducing Features to Improve Bug Prediction

Regression for Sentence-Level MT Evaluation with Pseudo References

Ensemble Technique Utilization for Indonesian Dependency Parser

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Australian Journal of Basic and Applied Sciences

HLTCOE at TREC 2013: Temporal Summarization

Mining Topic-level Opinion Influence in Microblog

CSL465/603 - Machine Learning

Truth Inference in Crowdsourcing: Is the Problem Solved?

Multilingual Sentiment and Subjectivity Analysis

Active Learning. Yingyu Liang Computer Sciences 760 Fall

AQUA: An Ontology-Driven Question Answering System

Prediction of Maximal Projection for Semantic Role Labeling

Lecture 1: Basic Concepts of Machine Learning

SARDNET: A Self-Organizing Feature Map for Sequences

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

A heuristic framework for pivot-based bilingual dictionary induction

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

The stages of event extraction

Introduction to Simulation

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

BODY LANGUAGE ANIMATION SYNTHESIS FROM PROSODY AN HONORS THESIS SUBMITTED TO THE DEPARTMENT OF COMPUTER SCIENCE OF STANFORD UNIVERSITY

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Corrective Feedback and Persistent Learning for Information Extraction

Georgetown University at TREC 2017 Dynamic Domain Track

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Evolutive Neural Net Fuzzy Filtering: Basic Description

Multi-Lingual Text Leveling

Experts Retrieval with Multiword-Enhanced Author Topic Model

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Transfer Learning Action Models by Measuring the Similarity of Different Domains

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Evolution of Random Phenomena

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Speech Recognition at ICSI: Broadcast News and beyond

Multilingual Document Clustering: an Heuristic Approach Based on Cognate Named Entities

Language Independent Passage Retrieval for Question Answering

Atypical Prosodic Structure as an Indicator of Reading Level and Text Difficulty

Rule-based Expert Systems

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

CS 446: Machine Learning

Action Recognition and Video

Clickthrough-Based Translation Models for Web Search: from Word Models to Phrase Models

The Smart/Empire TIPSTER IR System

Guru: A Computer Tutor that Models Expert Human Tutors

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

A Neural Network GUI Tested on Text-To-Phoneme Mapping

GACE Computer Science Assessment Test at a Glance

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Statewide Framework Document for:

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Named Entity Recognition: A Survey for the Indian Languages

arxiv: v2 [cs.cv] 30 Mar 2017

arxiv: v1 [cs.lg] 15 Jun 2015

Transcription:

Joint Modeling of Content and Discourse Relations in Dialogues Kechen Qin 1, Lu Wang 1, and Joseph Kim 2 1 College of Computer and Information Science Northeastern University 2 Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology qin.ke@husky.neu.edu

Motivation 2

Questions How to pinpoint and extract salient content from a meeting? How to understand and model the effectiveness of a meeting? 3

Related Work Extract salient content from the meeting - Jones (1999) - Fernandez et al. (2008) - Riedhammer et al. (2010) - Wang and Cardie (2012) Leverage discourse information to extract important information from meetings - Murray et al. (2006) - Galley (2006) 4

Content and discourse are intertwined A: I was just wondering if we want to have a rubber cover instead of a plastic one. B: Yeah. Positive C: So instead of the fascia that comes off being plastic, the fascia that comes off would be the rubber. Positive Elaboration Positive D: Alright. that could be a good idea. Elaboration E: Alright. that could be a good idea. Elaboration E: It would be comfortable to hold on also. B: Well that's been really popular with mobile phones so I don't see why not. 5

Unfortunately Discourse parsing in dialogues is still a challenging problem 6

Contributions Propose a framework to model the interaction between discourse and content in the meeting Model the consistency of understanding to learn the effectiveness of the meeting 7

Outline Introduction Methodology Corpus and Annotation Evaluation Consistency of Understanding Conclusion 8

Notations D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. B: Is a kinetic one going to be able to supply enough power? 9

Notations D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. x: discourse Unit B: Is a kinetic one going to be able to supply enough power? on argument level 10

Notations D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. Uncertain B: Is a kinetic one going to be able to supply enough power? d: Discourse Relation d = uncertain 11

Notations D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. c: Candidate Phrases c D = unimportant, unimportant, important c B = important B: Is a kinetic one going to be able to supply enough power? 12

Generic Framework Important candidate phrase Discourse relation P c, d x, w Input: discourse unit Model parameters 13

Generic Framework A log-linear model P c, d x, w = exp w φ c, d, x σ c,d exp w φ c, d, x exp w φ c, d, x Model Parameters Features 14

Generic Framework A log-linear model P c, d x, w exp w φ c, d, x exp w c φ c c, x + w d φ d d, x + w cd φ cd c, d, x Content features Discourse features Joint features E.g., whether the head word of the phrase was mentioned in preceding turn E.g., similarity between two discourse units 15

Generic Framework A log-linear model P c, d x, w exp w φ c, d, x exp w c φ c c, x + w d φ d d, x + w cd φ cd c, d, x Joint features E.g., whether phrases are salient when an elaboration relation is surrounded by two sentences with high similarity 16

Joint Learning SampleRank (Rohanimanesh et al., 2011) - Sampling-based search algorithm - Construct a sequence of configurations for sample labels as a Markov chain Monte Carlo (MCMC) chain - No limitations on the feature set Goyal and Eisenstein (2016) - On news articles summarization with Rhetorical Structure Theory (RST) - On sentence level - With simple summary features 17

SampleRank D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. Uncertain B: Is a kinetic one going to be able to supply enough power? 18

SampleRank D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. Elaboration B: Is a kinetic one going to be able to supply enough power? Initialization: Salient content phrases label: [unimportant, important, unimportant, important] Discourse relation label: [Elaboration] 19

SampleRank D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. Uncertain B: Is a kinetic one going to be able to supply enough power? Old samples (initialization): [unimportant, important, unimportant, important] [Elaboration] New samples: Sampled salient content phrases: [unimportant, unimportant, important, unimportant] Sampled discourse: [Uncertain] 20

SampleRank D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. Uncertain B: Is a kinetic one going to be able to supply enough power? Old samples (initialization): [unimportant, important, unimportant, important] [Elaboration] New samples: [unimportant, unimportant, important, unimportant] [Uncertain] Accept the new samples, if it improves the scoring function if score new score old > 0 old_samples new_samples 21

SampleRank D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. Uncertain B: Is a kinetic one going to be able to supply enough power? Old samples: [unimportant, important, unimportant, important] [Elaboration] New samples: [unimportant, unimportant, important, unimportant] [Uncertain] Accept the new samples, if it improves the scoring function if score new score old > 0 old_samples new_samples Update the parameters of the model based on old and new samples 22

Joint Inference Infer discourse and salient content iteratively - Dynamic Programming Infer discourse relation - Integer Linear Programming Infer salient phrase candidate 23

Joint Model with latent discourse Discourse relation as latent variable Important candidate phrase Discourse relation P c x, w d P c, d x, w Input: Discourse unit Model parameters 24

SampleRank D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. x B: Is a kinetic one going to be able to supply enough power? Old samples: [unimportant, important, unimportant, important] [discourse_type_1] New samples: [unimportant, unimportant, important, unimportant] [discourse_type_2] Accept the new samples, if it improves the scoring function if score new score old > 0 old_samples new_samples Update the parameters of the model based on old and new samples 25

SampleRank D: Two different types of batteries. Um can either use a hand dynamo, or the kinetic type ones. x B: Is a kinetic one going to be able to supply enough power? Old samples: [unimportant, important, unimportant, important] [discourse_type_1] New samples: [unimportant, unimportant, important, unimportant] [discourse_type_2] Accept the new samples, if it improves the scoring function if score new score old > 0 old_samples new_samples Update the parameters of the model based on old and new samples 26

Outline Introduction Methodology Corpus and Annotation Evaluation Consistency of Understanding Conclusion 27

Meeting Corpora AMI meetings (Carletta et al., 2006) - Annotated with abstractive summaries, argumentative discourse units, and argumentative discourse relations (Twente Argumentation schema by Rienks et al. 2005) ICSI meetings (Janin et al., 2003) - Annotated with salient content label 28

Outline Introduction Methodology Corpus and Annotation Evaluation Consistency of Understanding Conclusion 29

Evaluation Content selection - Extractive summarizer Discourse relation prediction - Discourse parser 30

Baselines and Comparisons: Summarization Longest Dialogue Act Centroid Dialogue Act Support Vector Machine (SVM) Keyword Extractive Approach (Liu et al., 2016) - Heuristic method using linguistic features - For fair comparison, we change it to keyphrase - State-of-the-art 31

Extractive Summary Length of Summary ROUGE_1_F1 ROUGE_SU4_F1 Longest Dialogue Act 30.9 23.1 15.3 Centroid Dialogue Act 17.5 20.8 11.3 SVM Baseline 49.8 27.5 11.8 Keyword Extraction (Liu et al., 2016) 62.4 36.2 13.5 Joint Model 66.6 41.1 20.9 Joint Model with Latent Discourse 85.9 42.4 21.3 Rouge_1: Unigrams Rouge_SU4: Skip-bigrams with at most 4 words in between 32

Discourse Relation Prediction 9 discourse relations in predefined discourse relations set from Twente Argumentation schema by Rienks et al. (2005) 33

Baselines and Comparisons: Discourse Accuracy F1 Majority Label 51.2 7.5 SVM Baseline 51.2 22.8 Support Vector Machine (SVM) - 5-fold Cross Validation - With the same feature set as our joint model 34

Baselines and Comparisons: Discourse Accuracy F1 Majority Label 51.2 7.5 SVM Baseline 51.2 22.8 Neural Language Model (Ji et al., 2016) 54.2 21.4 Neural Language Model (Ji et al., 2016) - State-of-the-art - Propose a novel latent variable recurrent neural network architecture for jointly modeling sequences of words and discourse relations 35

Baselines and Comparisons: Discourse Accuracy F1 Majority Label 51.2 7.5 SVM Baseline 51.2 22.8 Neural Language Model (Ji et al., 2016) 54.2 21.4 Joint Model 59.2 23.4 36

Outline Introduction Methodology Corpus and Annotation Evaluation Consistency of Understanding Conclusion 37

Consistency of Understanding Compare participant summaries to determine whether participants report the same decisions (Kim et al., 2016) Binary Classification Task - consistent vs. inconsistent 38

Our Model - Features Consistency Probability Probability ofconsistent understanding: max c,d P c, d x, w consistent - Probability of inconsistent understanding: max c,d P(c, d x, w inconsistent ) 39

Our Model - Features Consistency Probability Discourse Relation - Based on our study, there is high correlation between discourse information and consistency of the meeting - Positive, Negative (+) - Request, Specialization (-) - Unigram and bigram discourse relations 40

Our Model - Features Consistency Probability Discourse Relation Word Entrainment (Nenkova et al., 2008) - People tend to use similar words as the meeting proceeds - This phenomenon is very likely to be detected in effective meetings, when participants are on the same page in the meeting 41

Baselines and Comparisons Support Vector Machine (SVM) - Leave-one-out - Unigram and bigrams Hidden Markov Model (HMM) (Kim et al., 2016) - State-of-the-art - Discourse and head gesture 42

Results Accuracy F1 Majority Label 66.7 40.0 SVM Baseline 51.2 50.6 Hidden Markov Model (Kim et al., 2016) Oracle Discourse Relation Oracle Word Entrainment 60.5 50.5 69.8 62.7 61.2 57.8 43

Results Accuracy F1 Majority Label 66.7 40.0 SVM Baseline 51.2 50.6 Hidden Markov Model (Kim et al., 2016) Oracle Discourse Relation Oracle Word Entrainment 60.5 50.5 69.8 62.7 61.2 57.8 Our Model 68.2 63.1 44

Conclusion Propose a flexible framework to jointly model content and discourse. We achieve good performance on discourse recognition and salient content extraction tasks By using the outputs of our model, our system is able to learn the consistency prediction task 45

Future Work How to model idea flows among participants? - Which fraction of ideas are discussed and what is the outcome? - Which fraction of ideas are not discussed thoroughly and why? How can we leverage the discourse to capture the idea generation process? 46

Resources Project website (code & data): http://www.ccs.neu.edu/home/kechenqin/paper/acl20 17.html Consistency data download: http://people.csail.mit.edu/joseph_kim/data/cou_ami. zip Contact: qin.ke@husky.neu.edu 47

All for survival! 48

Thank you! Any Questions? 49