Stanford NLP. Evan Jaffe and Evan Kozliner

Similar documents
Ensemble Technique Utilization for Indonesian Dependency Parser

Unsupervised Learning of Narrative Schemas and their Participants

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

Python Machine Learning

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Developing a TT-MCTAG for German with an RCG-based Parser

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

AQUA: An Ontology-Driven Question Answering System

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

CS Machine Learning

(Sub)Gradient Descent

Second Exam: Natural Language Parsing with Neural Networks

Online Updating of Word Representations for Part-of-Speech Tagging

Rule Learning With Negation: Issues Regarding Effectiveness

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

arxiv: v1 [cs.cl] 2 Apr 2017

BYLINE [Heng Ji, Computer Science Department, New York University,

SEMAFOR: Frame Argument Resolution with Log-Linear Models

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Basic Parsing with Context-Free Grammars. Some slides adapted from Julia Hirschberg and Dan Jurafsky 1

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

Using dialogue context to improve parsing performance in dialogue systems

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Rule Learning with Negation: Issues Regarding Effectiveness

Distant Supervised Relation Extraction with Wikipedia and Freebase

A Version Space Approach to Learning Context-free Grammars

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Semantic and Context-aware Linguistic Model for Bias Detection

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

University of Alberta. Large-Scale Semi-Supervised Learning for Natural Language Processing. Shane Bergsma

A Dataset of Syntactic-Ngrams over Time from a Very Large Corpus of English Books

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Parsing of part-of-speech tagged Assamese Texts

A Comparison of Two Text Representations for Sentiment Analysis

Learning Methods for Fuzzy Systems

Grammars & Parsing, Part 1:

A Graph Based Authorship Identification Approach

The stages of event extraction

Word Segmentation of Off-line Handwritten Documents

Towards a MWE-driven A* parsing with LTAGs [WG2,WG3]

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

The Smart/Empire TIPSTER IR System

Short Text Understanding Through Lexical-Semantic Analysis

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Business Analytics and Information Tech COURSE NUMBER: 33:136:494 COURSE TITLE: Data Mining and Business Intelligence

Mining Association Rules in Student s Assessment Data

What Can Neural Networks Teach us about Language? Graham Neubig a2-dlearn 11/18/2017

CS 598 Natural Language Processing

Probabilistic Latent Semantic Analysis

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

stateorvalue to each variable in a given set. We use p(x = xjy = y) (or p(xjy) as a shorthand) to denote the probability that X = x given Y = y. We al

Switchboard Language Model Improvement with Conversational Data from Gigaword

Linking Task: Identifying authors and book titles in verbose queries

arxiv: v1 [cs.cv] 10 May 2017

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

Beyond the Pipeline: Discrete Optimization in NLP

A Vector Space Approach for Aspect-Based Sentiment Analysis

Fragment Analysis and Test Case Generation using F- Measure for Adaptive Random Testing and Partitioned Block based Adaptive Random Testing

South Carolina English Language Arts

THE world surrounding us involves multiple modalities

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Discriminative Learning of Beam-Search Heuristics for Planning

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Driving Author Engagement through IEEE Collabratec

Humboldt-Universität zu Berlin

Applications of memory-based natural language processing

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Detecting English-French Cognates Using Orthographic Edit Distance

Modeling function word errors in DNN-HMM based LVCSR systems

Lecture 1: Machine Learning Basics

Natural Language Processing. George Konidaris

ScienceDirect. Malayalam question answering system

MYCIN. The MYCIN Task

arxiv: v1 [cs.lg] 15 Jun 2015

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Lecture 1: Basic Concepts of Machine Learning

AN INTRODUCTION (2 ND ED.) (LONDON, BLOOMSBURY ACADEMIC PP. VI, 282)

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

Probability and Statistics Curriculum Pacing Guide

Human-like Natural Language Generation Using Monte Carlo Tree Search

Unsupervised Dependency Parsing without Gold Part-of-Speech Tags

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Indian Institute of Technology, Kanpur

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Resolving Complex Cases of Definite Pronouns: The Winograd Schema Challenge

Welcome to ACT Brain Boot Camp

Extracting Social Networks and Biographical Facts From Conversational Speech Transcripts

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

A heuristic framework for pivot-based bilingual dictionary induction

CSL465/603 - Machine Learning

Transcription:

Stanford NLP Evan Jaffe and Evan Kozliner

Some Notable Researchers Chris Manning Statistical NLP, Natural Language Understanding and Deep Learning Dan Jurafsky sciences Percy Liang Natural Language Understanding, conversational speech & dialog and NLP for the social Semantic parsing, probabilistic models for NLP

Manning Article A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task ACL 2016 Outstanding Paper Awards Reading comprehension task Google s Deep Mind researchers built the data set based on CNN and Daily Mail Question and answer dataset (based on bullet points) Coreference chains replaced with markers (ensures the machine is reading the text and not just using a language mode to answer the question) Small number of features such as the frequency of a potential answer Classifiers examined: LSTM and conventional ensemble of decision trees Accuracy around 75% (potentially the best possible due to the data prep)

Jurafsky Article Predicting the Rise and Fall of Scientific Topics from Trends in their Rhetorical Framing Concept of rhetorical scholarly functions of topics (as opposed to sentences) Examines the relationship between the rhetorical role of topics (for example if a topic is being used as a result) and the popularity of the articles they belong to Interesting system involving several pieces: Algorithm to tag the rhetorical roles of topics in articles where they are untagged, creates a tuple of topics to rhetorical roles Use this and the popularity of the article at that time to create time series of the rhetorical functions and their popularity in that time Cluster time series in unsupervised way Classify time series with logistic regression based on their cluster

Improving Coreference Resolution by Learning Entity-Level Representations Clark and Manning, 2016 ACL Identity Coreference: structured prediction that links mentions that have the same referent The city council denied the demonstrators a permit because they feared violence. they advocated violence. (Winograd 1972)

Improving Coreference Resolution by Learning Entity-Level Representations Clark and Manning, 2016 ACL Identity Coreference: structured prediction that links mentions that have the same referent The city council denied the demonstrators a permit because they feared violence. they advocated violence. (Winograd 1972)

Their Approach Easy-first agglomerative clustering Start state: all mentions in their own cluster Merge clusters until final partition reached. Deciding whether to merge or not depends on the score of the resulting cluster vs. the score of the mention being a singleton.

A Fast and Accurate Parser Using Neural Networks Chen and Manning, 2014 EMNLP Universal Dependencies, Marie-Catherine de Marneffe (OSU LING) Transition-based parsing: buffer, stack, linear scan Contrast with bottom-up parsing: speed, theoretical issues

Transition-based Parsing Stack: partial parse, including words currently being processed Buffer: words yet to be processed Parser applies transitions to its state until its buffer is empty and the dependency graph is completed. Initial state: all words in order on the buffer, with a single dummy ROOT node on the stack. Transitions: LEFT-ARC: marks the second item on the stack as a dependent of the first item, and removes the second item from the stack (if the stack contains at least two items). RIGHT-ARC: marks the first item on the stack as a dependent of the second item, and removes the first item from the stack (if the stack contains at least two items). SHIFT: removes a word from the buffer and pushes it onto the stack (if the buffer is not empty).

Transition-based Parsing