Identifying Localization in Reviews of Argument Diagrams

Similar documents
AQUA: An Ontology-Driven Question Answering System

CS Machine Learning

The stages of event extraction

Writing a Basic Assessment Report. CUNY Office of Undergraduate Studies

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Linking Task: Identifying authors and book titles in verbose queries

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Discriminative Learning of Beam-Search Heuristics for Planning

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Rule Learning With Negation: Issues Regarding Effectiveness

CSC200: Lecture 4. Allan Borodin

Lecture 1: Machine Learning Basics

(Sub)Gradient Descent

Learning Methods in Multilingual Speech Recognition

Using dialogue context to improve parsing performance in dialogue systems

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Ensemble Technique Utilization for Indonesian Dependency Parser

Radius STEM Readiness TM

arxiv: v1 [cs.cl] 2 Apr 2017

Modeling user preferences and norms in context-aware systems

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

What Different Kinds of Stratification Can Reveal about the Generalizability of Data-Mined Skill Assessment Models

Writing Research Articles

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Building a Semantic Role Labelling System for Vietnamese

Beyond the Pipeline: Discrete Optimization in NLP

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Visual CP Representation of Knowledge

Rule Learning with Negation: Issues Regarding Effectiveness

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

A Case Study: News Classification Based on Term Frequency

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Prediction of Maximal Projection for Semantic Role Labeling

Julia Smith. Effective Classroom Approaches to.

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Literature and the Language Arts Experiencing Literature

Learning From the Past with Experiment Databases

Case study Norway case 1

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

re An Interactive web based tool for sorting textbook images prior to adaptation to accessible format: Year 1 Final Report

Assignment 1: Predicting Amazon Review Ratings

BEETLE II: a system for tutoring and computational linguistics experimentation

On-Line Data Analytics

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Probabilistic Latent Semantic Analysis

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

Postprint.

Calibration of Confidence Measures in Speech Recognition

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

Shockwheat. Statistics 1, Activity 1

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Reduce the Failure Rate of the Screwing Process with Six Sigma Approach

COMPUTER-ASSISTED INDEPENDENT STUDY IN MULTIVARIATE CALCULUS

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Detecting English-French Cognates Using Orthographic Edit Distance

A Framework for Customizable Generation of Hypertext Presentations

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Fragment Analysis and Test Case Generation using F- Measure for Adaptive Random Testing and Partitioned Block based Adaptive Random Testing

Automating the E-learning Personalization

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

MADERA SCIENCE FAIR 2013 Grades 4 th 6 th Project due date: Tuesday, April 9, 8:15 am Parent Night: Tuesday, April 16, 6:00 8:00 pm

Disambiguation of Thai Personal Name from Online News Articles

Learning to Rank with Selection Bias in Personal Search

Multi-Lingual Text Leveling

Interactive Whiteboard

Outline. Web as Corpus. Using Web Data for Linguistic Purposes. Ines Rehbein. NCLT, Dublin City University. nclt

Person Centered Positive Behavior Support Plan (PC PBS) Report Scoring Criteria & Checklist (Rev ) P. 1 of 8

The Role of the Head in the Interpretation of English Deverbal Compounds

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Cross Language Information Retrieval

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Optimizing to Arbitrary NLP Metrics using Ensemble Selection

A Domain Ontology Development Environment Using a MRD and Text Corpus

Degree Qualification Profiles Intellectual Skills

Tap vs. Bottled Water

FIGURE IT OUT! MIDDLE SCHOOL TASKS. Texas Performance Standards Project

Learning a Cross-Lingual Semantic Representation of Relations Expressed in Text

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

The Smart/Empire TIPSTER IR System

A Coding System for Dynamic Topic Analysis: A Computer-Mediated Discourse Analysis Technique

Chapter 2 Rule Learning in a Nutshell

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

A Bayesian Learning Approach to Concept-Based Document Classification

Online Updating of Word Representations for Part-of-Speech Tagging

PAGE(S) WHERE TAUGHT If sub mission ins not a book, cite appropriate location(s))

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

Prentice Hall Literature: Timeless Voices, Timeless Themes Gold 2000 Correlated to Nebraska Reading/Writing Standards, (Grade 9)

Disciplinary Literacy in Science

South Carolina English Language Arts

Constructing Parallel Corpus from Movie Subtitles

University of Alberta. Large-Scale Semi-Supervised Learning for Natural Language Processing. Shane Bergsma

Innovative Methods for Teaching Engineering Courses

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Australian Journal of Basic and Applied Sciences

A Vector Space Approach for Aspect-Based Sentiment Analysis

Statewide Framework Document for:

Prentice Hall Literature: Timeless Voices, Timeless Themes, Platinum 2000 Correlated to Nebraska Reading/Writing Standards (Grade 10)

Transcription:

Identifying Localization in Reviews of Argument Diagrams Huy Nguyen 1 Diane Litman 1,2 1 Computer Science Department 2 Learning Research and Development Center at University of Pittsburgh

ArgumentPeer Project * Source texts Use diagram to help writing Author creates argument diagram AI guides reviewing Peers review argument diagram Phase I: Argument Diagramming Author revises argument diagram Phase II: Writing Author revises paper Peers review paper Author writes paper * NSF Grant No. 1122504 2

Outline Introduction Corpus Location Pattern Algorithm Experiments Discussion and future work Integrating LPA into SWoRD Outline 3

Peer reviews with SWoRD (Cho and Schunn, 2007) Web-based reciprocal peer review system to facilitate writing and reviewing practices for students Manage typical activity cycles such as writing, reviewing, back-evaluating, and rewriting However, SWoRD lacks intelligence for detecting and responding to problems with student reviewing s performance Introduction 4

Argument diagram with LASAD (Scheuer et. al, 2009) Support the learning of argumentation skills through graphical representations Argument diagrams with nodes represent statements and arcs represent argumentative or rhetorical relations By combining SWoRD and LASAD, student argument diagrams are distributed to student reviewers for comment Introduction 5

Review localization Pinpointing the source or location of a problem and/or solution (Nelson and Schunn, 2009) Significantly related to feedback implementation of peer paper review (Nelson and Schunn, 2009) and peer argument diagram review (Lippman et al., 2012) Paper review localization was proved predictable using NLP and ML techniques (Xiong and Litman, 2011) We address review localization of peer argument diagram review Introduction 6

Research goals Overall: Adapting and applying Natural Language Processing and Machine Learning techniques to help peer reviewers review the diagram and/or writing of others based on automatic detection of effective review comment features This study: Automatically identifying review localization in student argument diagram reviews Introduction 7

Argument diagram review corpus Context: Research Method Lab, Fall 2011 Students created argument, student reviewers then provided written feedback and grades Instructor-defined ontology 4 node types: current study, hypothesis, claim, and citation 4 arc types: comparison, undefined, supports, and opposes Comments were segmented into 1104 idea units (contiguous feedback referring to a single topic) On average, each diagram was reviewed by 3 peers with 19 comment units Corpus 8

An example argument diagram (localization is highlighted) Not localized The citations presented are solid evidence but are not presented in the best way possible. The justification is understandable but not convincing. Also the con-argument for the time of day hypothesis is not sufficient. Citation 15 does not oppose the claim. Corpus Localized 9

Annotation Two annotators coded 1104 comments for issue types: praise, summary, problem, solution, problem and solution (both), or uncodeable 590 comments having types of praise, problem, or both were further coded for localization with label = {yes, no} Inter-rater reliability (kappa) is high: 0.87 for issue type 0.84 for localization Corpus 10

Diagram Review Localization: Observation Paper review vs. diagram review Graph structure of argument diagrams makes it more convenient to include location information Xiong and Litman (2011) reported 53% of reviews localized Our corpus has 74% of reviews localized The way that localization is realized in diagram review differs from that in paper review Location Patterns Algorithm 11

Location Patterns Numbered ontology type A diagram component is identified by referring to its node/arc type followed by ID/order number hypothesis 1 support arc 15 Location Patterns Algorithm 12

Location Patterns Textual component content: text in diagram node/arc are made concise Reviewers use textual content in conjunction with node/arc type gender hypothesis claim that women are more polite than men Location Patterns Algorithm 13

Location Patterns Connected component: referring to a line of argumentation Identify connection between components support for the gender hypothesis claim node in between the opposes and support arcs 26 and 32 Location Patterns Algorithm 14

Location Patterns Unique Component: identifying the unique node/arc of a given type The opposing arc Typical numerical expressions are used to express localization The second hypothesis, H2 [14] (claim node), #22 (support arc) Location Patterns Algorithm 15

Localization Pattern Algorithm (LPA) Location information must involve diagram component keyword surrounded by supporting words A diagram component keyword: The words node or arc Node/arc type from the ontology (parsed automatically) Supporting words are in proximity of a keyword which help locate the component Localization Pattern Algorithm 16

Localization Pattern Algorithm Supporting words are selected from common words between review and node/arc content (stemmed already) Identified accordingly to 5 localization pattern (applied to review sentences that have common words) Numbered ontology type: supporting words are number/list of numbers right after keyword Textual component content: Supporting words occur right before keyword Or after keyword with distance less than 3 Localization Pattern Algorithm 17

Localization pattern algorithm Unique component: count number of node/arc of each type while parsing argument diagrams Connected component: extend node/arc text by the textual content of the other node/arc that it connects to Supporting words must be in the extended content Typical numerical expressions: use held-out development data to learn regular expressions Localization Pattern Algorithm 18

Features used in paper review localization Xiong and Litman 2011: studied syntactic features from the parsed dependency tree of sentence Domain word count (dw_cnt) dictionary of domain word is learned automatically from set of argument diagrams So_domain: indicates whether domain word appear between subject and object of review Det_count: counts number of demonstrative determiners in comment Overlapping window features: Compute the maximal overlapping window Report window size (wnd_size) and number of common words (overlap_num) Paper review localization 19

Experimental results Two baseline models Majority model (simply assign every instance label of the most common class) plocalization model using only paper review features Syntactic features vs. structural patterns Two proposed models: LPA: use only output of LPA to identify the labels Combined: add LPA binary output as a feature into plocalization Models are learned using decision tree (Weka J48) Evaluated via 10-fold cross validation Experiments 20

Experimental results Metric Accuracy (%) Kappa Weighted precision Weighted recall Majority 74.07 0 0.55 0.74 *: significantly better than Majority plocalization 73.98 < 0.01 0.55 0.74 plocalization does not outperform Majority LPA alone is significantly better than baselines LPA can predict efficiently the minor class Combined model yields the best results of all LPA 80.34 * 0.54 * 0.83 * 0.80 * Combined 83.78 * 0.56 * 0.84 * 0.84 * Experiments 21

Learned decision tree Experiments 22

Integrating LPA into SWoRD Textual comment Dimension Integrating LPA into SWoRD 23

Screenshot of system intervention Reviewer makes decision System guides reviewer Integrating LPA into SWoRD 24

Conclusion and future work LPA algorithm for identifying localization in peer review of argument diagrams Outperforms a model developed for paper review localization Combining the two approaches work best of all Deployed in SWoRD in June 2013 In future, automatically learn patterns and regular expressions Test on new corpus with different ontology Apply lesson learned from developing LPA back to paper review localization model Conclusion and future work 25

THANK YOU Questions and Comments QA 26

Numerical rating footer 27

Selected examples Type + ID/function Type + content Connection path 28