WordSleuth: Deducing Social Connotations from Syntactic Clues. Shannon Stanton UROP May 14, Shannon Stanton

Similar documents
Lecture 1: Machine Learning Basics

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

CS Machine Learning

Python Machine Learning

Switchboard Language Model Improvement with Conversational Data from Gigaword

A Case Study: News Classification Based on Term Frequency

Learning From the Past with Experiment Databases

Rule Learning With Negation: Issues Regarding Effectiveness

Linking Task: Identifying authors and book titles in verbose queries

Experience College- and Career-Ready Assessment User Guide

Word Segmentation of Off-line Handwritten Documents

Australian Journal of Basic and Applied Sciences

Speech Emotion Recognition Using Support Vector Machine

(Sub)Gradient Descent

Probabilistic Latent Semantic Analysis

Rule Learning with Negation: Issues Regarding Effectiveness

CS 446: Machine Learning

National Literacy and Numeracy Framework for years 3/4

MADERA SCIENCE FAIR 2013 Grades 4 th 6 th Project due date: Tuesday, April 9, 8:15 am Parent Night: Tuesday, April 16, 6:00 8:00 pm

Reducing Features to Improve Bug Prediction

Interactive Whiteboard

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

The stages of event extraction

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Detecting English-French Cognates Using Orthographic Edit Distance

Sight Word Assessment

Probability and Statistics Curriculum Pacing Guide

J j W w. Write. Name. Max Takes the Train. Handwriting Letters Jj, Ww: Words with j, w 321

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

A Comparison of Two Text Representations for Sentiment Analysis

Activity Recognition from Accelerometer Data

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Ryerson University Sociology SOC 483: Advanced Research and Statistics

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Mining Association Rules in Student s Assessment Data

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

A Coding System for Dynamic Topic Analysis: A Computer-Mediated Discourse Analysis Technique

Evaluation of Teach For America:

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Testing for the Homeschooled High Schooler: SAT, ACT, AP, CLEP, PSAT, SAT II

CS4491/CS 7265 BIG DATA ANALYTICS INTRODUCTION TO THE COURSE. Mingon Kang, PhD Computer Science, Kennesaw State University

Multi-Lingual Text Leveling

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Assignment 1: Predicting Amazon Review Ratings

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

AQUA: An Ontology-Driven Question Answering System

A heuristic framework for pivot-based bilingual dictionary induction

Online Updating of Word Representations for Part-of-Speech Tagging

Teachers: Use this checklist periodically to keep track of the progress indicators that your learners have displayed.

Introduction to the Practice of Statistics

Analyzing sentiments in tweets for Tesla Model 3 using SAS Enterprise Miner and SAS Sentiment Analysis Studio

Atypical Prosodic Structure as an Indicator of Reading Level and Text Difficulty

A Web Based Annotation Interface Based of Wheel of Emotions. Author: Philip Marsh. Project Supervisor: Irena Spasic. Project Moderator: Matthew Morgan

Affective Classification of Generic Audio Clips using Regression Models

Generative models and adversarial training

RESPONSE TO LITERATURE

Research Design & Analysis Made Easy! Brainstorming Worksheet

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Genre classification on German novels

Summary / Response. Karl Smith, Accelerations Educational Software. Page 1 of 8

Feature Selection based on Sampling and C4.5 Algorithm to Improve the Quality of Text Classification using Naïve Bayes

Verbal Behaviors and Persuasiveness in Online Multimedia Content

4 Almost always mention the topic and the overall idea of simple. 3 Oftentimes mention the topic and the overall idea of simple

American Journal of Business Education October 2009 Volume 2, Number 7

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Review in ICAME Journal, Volume 38, 2014, DOI: /icame

Multivariate k-nearest Neighbor Regression for Time Series data -

PREP S SPEAKER LISTENER TECHNIQUE COACHING MANUAL

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

PAGE(S) WHERE TAUGHT If sub mission ins not a book, cite appropriate location(s))

Unit 2. A whole-school approach to numeracy across the curriculum

Language Acquisition Chart

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

Learning Methods in Multilingual Speech Recognition

Teacher: Mlle PERCHE Maeva High School: Lycée Charles Poncet, Cluses (74) Level: Seconde i.e year old students

Modeling function word errors in DNN-HMM based LVCSR systems

Welcome to ACT Brain Boot Camp

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Blank Table Of Contents Template Interactive Notebook

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

MYP Language A Course Outline Year 3

Semantic and Context-aware Linguistic Model for Bias Detection

Acquiring Competence from Performance Data

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Using Hashtags to Capture Fine Emotion Categories from Tweets

Applications of data mining algorithms to analysis of medical data

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Model Ensemble for Click Prediction in Bing Search Ads

EXAMINING THE DEVELOPMENT OF FIFTH AND SIXTH GRADE STUDENTS EPISTEMIC CONSIDERATIONS OVER TIME THROUGH AN AUTOMATED ANALYSIS OF EMBEDDED ASSESSMENTS

Applications of memory-based natural language processing

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Writing Research Articles

Modeling function word errors in DNN-HMM based LVCSR systems

Impact of Cluster Validity Measures on Performance of Hybrid Models Based on K-means and Decision Trees

1. READING ENGAGEMENT 2. ORAL READING FLUENCY

Speech Recognition at ICSI: Broadcast News and beyond

Transcription:

WordSleuth: Deducing Social Connotations from Syntactic Clues Shannon Stanton Shannon Stanton UROP May 14, 2011 1

Plan I. Research Question II. WordSleuth A. Game-play B. Taboo list III. Machine Learning A. Data representation B. Classification Algorithms IV. Future Possibilities V. Question and Answer 2

I. Question Can humans derive complex social ideas from simple text? - intention: deception, persuasion - attitude: formality, politeness, rudeness - emotion: embarrassment, confidence 57%-71% (Pearl and Steyvers 2010)...Can computers? 3

Example Social connotations include: confidence disbelief persuading rudeness deception embarrassment politeness formality Example Text Input: I don't care if Nancy laughs at my outfit I think I look good! 4

II. WordSleuth Problem: Where to get the data? Solution: Create WordSleuth, a Game-With-A-Purpose (GWAP) to encourage people to annotate data. GWAP: Game created specifically to obtain data related to a particular research area. (von Ahn 2006) 5

II. WordSleuth: My Role To make improvements to the game: A. Enable online functionality B. Taboo-list functionality 6

Result II. A: Online Game App www.gwap.ss.uci.edu The message was: You know that the new findings at the symposium prove my theory and I can list at least 20 papers to disprove you before you even finish reading the titles. You guessed: confidence The answer: persuading 7

II. A. The Online Game Application Completing the web application of the game Currently 2,185 Annotated Messages with 8,941 annotations, Up from 1,167 Annotated Messages with 3,198 annotations 187% increase in messages, 280% increase in annotations 8

II. B. Online Game App Are people any good at it? Yes! target confidence 84.4 2.0 2.0 0.8 1.0 6.1 2.3 1.3 deception 4.5 74.3 4.3 2.4 1.1 7.8 3.2 2.4 disbelief 2.7 4.1 80.7 3.3 1.3 1.9 2.7 3.3 embarrassment 0.4 3.0 5.6 83.0 2.1 1.1 2.7 2.1 formality 1.4 0.0 0.7 1.0 70.5 2.4 22.4 1.7 persuading 6.1 5.1 0.8 0.6 3.0 80.2 3.0 1.2 politeness 1.6 2.2 0.6 1.8 13.8 3.4 75.4 1.2 rudeness 2.1 1.2 3.1 1.9 1.6 2.9 1.0 86.1 guesses Baseline: 1/8 = 12.5% Average: 80.4% 9

II. B. Taboo List 10

II. B. Taboo List - By discouraging use of words already wellrepresented in the data, we encourage breadth and variety of data. - Makes the game a bit more challenging for players. - Makes the job of the classifier algorithms harder, as unigrams will have less direct correlation with class. 11

II. B. Taboo List - Taboo Words calculated using Mutual Information - Mutual Information: A measure of correlation Example: If category confidence has 10 instances of Nancy, and no other category does, the mutual information will be high If all categories have the same number of a common word (such as the ) the mutual information will be low. 12

Results II. B: Taboo List > rudeness: popped, unprofessional, spotty > disbelief: jumped, megaphone, twenty > persuading: fast, alcohol, pay > deception: still, blonde, reality > embarrassment: accidentally, deodorant, surprising > formality: abuse, calm, soldier > politeness: yelled, scores, nices > confidence: nancy, modest, respectable 13

III. Machine Learning: A. Data Representation How to make use of the data? We can't just feed strings of English directly to the learning algorithms. Message ID : MessageText : Target Cue: Creator : Guesses/Category 1049 This is a very nice house you have here, Mrs. Smith, and such good coffee. formality labsubjectcl0 1 1 0 0 0 0 4 0 0 0 14

III. Machine Learning A. Data Representation So what features do we use anyway? Originally: - Vocabulary (that appears more than once in the data) - Bigrams/Trigrams (word sequences) - punctuation count - types:tokens ratio (unique words : total words) Added: - interrobangs?! -! :? ratio - sub clause analysis...over 4000 features and counting! 15

III. Machine Learning: A. Data Representation Solution: Feature Extraction Represent data as a list of ordered triples with a category (MessageID : FeatureID : Feature Value) Target Cue Sparsity: Allows us to ignore features not present for a given example. 16

III. Machine Learning What do we do with all that data anyway? Detective Data 17

III. Machine Learning B. Classification Algorithms - Previously used: SMLR (Sparse Multinomial Logistic Regression): 59% (Pearl and Steyvers 2010) - KNN (K Nearest Neighbors) - Transductive Clustering 18

III. Machine Learning B. Classification Algorithms 10-fold-cross-validation: - Train/Transduce algorithm on 90% of the data, test it on 10% Base line for Machine Learners: 13.5% (most common category) 19

III. Machine Learning B. Classification Algorithms KNN K nearest neighbors: Preliminary Success: 75.7% test accuracy Blue or yellow? 20

III. Machine Learning B. Classification Algorithms Transductive Clustering vs KNN Blue or yellow? Intuition:? KNN: blue Clustering: yellow 21

III. Machine Learning B. Classification Algorithms Transductive Agglomerative Clustering Blue or yellow? 22

III. B. Agglomerative Clustering Mean accuracy: 12.99% (deviation 0.00618) remember, baseline is 13.5% Why so poor? Unlabeled patterns take the label of the cluster with which they are joined. It never joins clusters with different labels. Thus, very near clusters and imperfect clusters become problems. 23

III. Machine Learning B. Classification Algorithms Transductive Clustering: Graph Cutter Blue or yellow? 24

III. B. Transductive Graph Cutter Mean Accuracy: 97.8% But, possibly over-fitting 25

III. Machine Learning B. Summary Algorithm Success SMLR 59% KNN 75.7% Transductive Agglomerative 12.99% Transductive Graph Cutting 97.8% 26

IV. Future Extensions Machine Learning Approaches: Additional Classification algorithms - Bagging the good ones - Encode the underlying assumption that each data entry of same ID should be classified the same. Applications: - In the way of a spell checker, an attitude checker - Computational modeling of human cognition 27

Summary I. Can computers learning social ques in text? Yes! II. How do we obtain data? WordSleuth a. Lots of data? WordSleuth online b. Good data? Taboo list III. How does a machine learn? KNN, Transduction IV. What's left to do approaches and applications 28

References and Acknowledgments Pearl, L. & Steyvers, M. (2010). Identifying Emotions, Intentions, & Attitudes in Text Using a Game with a Purpose. Proceedings of NAACL-HLT 2010 Workshop on Computational Approaches to Analysis and Generation of Emotion in Text. Los Angeles, CA: NAACL. von Ahn, L. 2006. Games With A Purpose. IEEE Computer Magazine, June 2006: 96-98. Waffles code repository: http://waffles.sourceforge.net 29

Questions? 30

Mutual Information Mutual Information = log ( p(x y) / p(x) ) For each word in the dataset p(x) = the frequency of word x (in the data set) p(y) = the frequency of social category y (in the dataset) p(x y) = the frequency of x in y 31

32