Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying the Polarity of Opinion Sentences

Similar documents
Multilingual Sentiment and Subjectivity Analysis

A Case Study: News Classification Based on Term Frequency

Linking Task: Identifying authors and book titles in verbose queries

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Lecture 1: Machine Learning Basics

On document relevance and lexical cohesion between query terms

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Switchboard Language Model Improvement with Conversational Data from Gigaword

Netpix: A Method of Feature Selection Leading. to Accurate Sentiment-Based Classification Models

Reducing Features to Improve Bug Prediction

Software Maintenance

Postprint.

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Rule Learning With Negation: Issues Regarding Effectiveness

The stages of event extraction

Leveraging Sentiment to Compute Word Similarity

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

What Different Kinds of Stratification Can Reveal about the Generalizability of Data-Mined Skill Assessment Models

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Emotions from text: machine learning for text-based emotion prediction

Rule Learning with Negation: Issues Regarding Effectiveness

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Australian Journal of Basic and Applied Sciences

Using Games with a Purpose and Bootstrapping to Create Domain-Specific Sentiment Lexicons

A Comparison of Two Text Representations for Sentiment Analysis

Multi-Lingual Text Leveling

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Robust Sense-Based Sentiment Classification

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

arxiv: v1 [cs.cl] 2 Apr 2017

Movie Review Mining and Summarization

Distant Supervised Relation Extraction with Wikipedia and Freebase

Vocabulary Usage and Intelligibility in Learner Language

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Loughton School s curriculum evening. 28 th February 2017

AQUA: An Ontology-Driven Question Answering System

Cross Language Information Retrieval

CS 446: Machine Learning

Speech Recognition at ICSI: Broadcast News and beyond

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

The Ups and Downs of Preposition Error Detection in ESL Writing

LQVSumm: A Corpus of Linguistic Quality Violations in Multi-Document Summarization

Search right and thou shalt find... Using Web Queries for Learner Error Detection

A Bayesian Learning Approach to Concept-Based Document Classification

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Beyond the Pipeline: Discrete Optimization in NLP

arxiv:cmp-lg/ v1 22 Aug 1994

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

HLTCOE at TREC 2013: Temporal Summarization

Re-evaluating the Role of Bleu in Machine Translation Research

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

SEMAFOR: Frame Argument Resolution with Log-Linear Models

Methods for the Qualitative Evaluation of Lexical Association Measures

CS Machine Learning

Integrating Semantic Knowledge into Text Similarity and Information Retrieval

Memory-based grammatical error correction

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

The Smart/Empire TIPSTER IR System

Using dialogue context to improve parsing performance in dialogue systems

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Disambiguation of Thai Personal Name from Online News Articles

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Truth Inference in Crowdsourcing: Is the Problem Solved?

Predicting Students Performance with SimStudent: Learning Cognitive Skills from Observation

Determining the Semantic Orientation of Terms through Gloss Classification

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2

Columbia University at DUC 2004

METHODS FOR EXTRACTING AND CLASSIFYING PAIRS OF COGNATES AND FALSE FRIENDS

A Re-examination of Lexical Association Measures

The taming of the data:

Using Small Random Samples for the Manual Evaluation of Statistical Association Measures

Extracting Opinion Expressions and Their Polarities Exploration of Pipelines and Joint Models

Detecting Online Harassment in Social Networks

South Carolina English Language Arts

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Comparison of network inference packages and methods for multiple networks inference

Word Segmentation of Off-line Handwritten Documents

Cross-Lingual Text Categorization

Formulaic Language and Fluency: ESL Teaching Applications

Universiteit Leiden ICT in Business

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Probabilistic Latent Semantic Analysis

Mandarin Lexical Tone Recognition: The Gating Paradigm

Assignment 1: Predicting Amazon Review Ratings

Analyzing sentiments in tweets for Tesla Model 3 using SAS Enterprise Miner and SAS Sentiment Analysis Studio

Prediction of Maximal Projection for Semantic Role Labeling

ESSLLI 2010: Resource-light Morpho-syntactic Analysis of Highly

Spoken Language Parsing Using Phrase-Level Grammars and Trainable Classifiers

Bridging Lexical Gaps between Queries and Questions on Large Online Q&A Collections with Compact Translation Models

The Good Judgment Project: A large scale test of different methods of combining expert predictions

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Verbal Behaviors and Persuasiveness in Online Multimedia Content

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Bootstrapping and Evaluating Named Entity Recognition in the Biomedical Domain

Transcription:

Presented by Lasse Soelberg Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 1 / 35 Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying the Polarity of Opinion Sentences Hong Yu Vasileios Hatzivassiloglou Columbia University, New York 24. November, 2008 Proceedings of the 2003 Conference on Empirical Methods in Natural Language Processing (EMNLP-03).

Layout 1 Goals of the Paper 2 Finding Opinion Sentences Identifying the Polarity 3 Data Evaluation Results 4 Related Work Article Evaluation Goals of the Paper Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 2 / 35

Goals of the Paper Towards Answering Opinion Questions Question-answering systems. Easier to use factual statements. Extend to also use subjective opinion statements. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 3 / 35

Goals of the Paper Towards Answering Opinion Questions Question-answering systems. Easier to use factual statements. Extend to also use subjective opinion statements. Simple Question Who was elected as the new US President in 2008? Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 3 / 35

Goals of the Paper Towards Answering Opinion Questions Question-answering systems. Easier to use factual statements. Extend to also use subjective opinion statements. Simple Question Who was elected as the new US President in 2008? Complex Question What has caused the current financial crisis? Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 3 / 35

Goals of the Paper Separating Facts from Opinions and Identifying the Polarity of Opinion Sentences Classifying articles as either subjective or objective Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 4 / 35

Goals of the Paper Separating Facts from Opinions and Identifying the Polarity of Opinion Sentences Classifying articles as either subjective or objective Finding Opinion Sentences In both subjective and objective articles Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 4 / 35

Goals of the Paper Separating Facts from Opinions and Identifying the Polarity of Opinion Sentences Classifying articles as either subjective or objective Finding Opinion Sentences In both subjective and objective articles Identify the Polarity of Opinion Sentences Determine if the opinions are positive or negative Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 4 / 35

Layout 1 Goals of the Paper 2 Finding Opinion Sentences Identifying the Polarity 3 Data Evaluation Results 4 Related Work Article Evaluation Finding Opinion Sentences Identifying the Polarity Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 5 / 35

Document Types Finding Opinion Sentences Identifying the Polarity Training Sets Articles from Wall Street Journal, which is annotated with document types. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 6 / 35

Document Types Finding Opinion Sentences Identifying the Polarity Training Sets Articles from Wall Street Journal, which is annotated with document types. Subjective Articles (Opinion) Editorials Letter to the Editor Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 6 / 35

Document Types Finding Opinion Sentences Identifying the Polarity Training Sets Articles from Wall Street Journal, which is annotated with document types. Subjective Articles (Opinion) Editorials Letter to the Editor Objective Articles (Fact) News Business Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 6 / 35

Classification Finding Opinion Sentences Identifying the Polarity Naive Bayes Calculating the likelihood that the document is either subjective or objective. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 7 / 35

Classification Finding Opinion Sentences Identifying the Polarity Naive Bayes Calculating the likelihood that the document is either subjective or objective. Bayes Rule P(c d) = P(c)P(d c) P(d) where c is a class, d is a document and single words are used as feature. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 7 / 35

Three Different Approaches Finding Opinion Sentences Identifying the Polarity Rely on Expectation Documents classified as opinions tends to have mostly opinion sentences, and documents classified as facts tends to have more factual sentences. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 8 / 35

Three Different Approaches Finding Opinion Sentences Identifying the Polarity Rely on Expectation Documents classified as opinions tends to have mostly opinion sentences, and documents classified as facts tends to have more factual sentences. The Three Approaches Similarity Approach Naive Bayes Classifier Multiple Naive Bayes Classifier Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 8 / 35

Similarity Approach Finding Opinion Sentences Identifying the Polarity Hypothesis Opinion sentences within a given topic will be more similar to other opinion sentences than to factual sentences. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 9 / 35

Similarity Approach Finding Opinion Sentences Identifying the Polarity Hypothesis Opinion sentences within a given topic will be more similar to other opinion sentences than to factual sentences. SimFinder Measures sentence similarity based on shared words, phrases and WordNet synsets. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 9 / 35

Variants Finding Opinion Sentences Identifying the Polarity The score variant Select documents with the same topic as the sentence. Average the similarities with each sentence in the documents. Assign the sentence to the category with the highest average. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 10 / 35

Variants Finding Opinion Sentences Identifying the Polarity The score variant Select documents with the same topic as the sentence. Average the similarities with each sentence in the documents. Assign the sentence to the category with the highest average. The frequency variant Count how many of the sentences, for each category, that exceeds a predetermined threshold (set to 0.65). Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 10 / 35

Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Bayes Rule P(c d) = P(c)P(d c) P(d) Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 11 / 35

Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Bayes Rule P(c d) = P(c)P(d c) P(d) Some of Features Used Words Bigrams Trigrams Parts of Speech Counts of positive and negative words Counts of the polarities of semantically oriented words Average semantic orientation score of the words Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 11 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Problem The designation of all sentences as opinions or facts is an approximation. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 12 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Problem The designation of all sentences as opinions or facts is an approximation. Solution Use multiple Naive Bayes classifiers, each using a different subset of the features. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 12 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Problem The designation of all sentences as opinions or facts is an approximation. Solution Use multiple Naive Bayes classifiers, each using a different subset of the features. The Goal Reduce the training set to the sentences most likely to be correctly labelled. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 12 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Train separate classifiers C 1, C 2,..., C m given separate feature sets F 1, F 2,..., F m. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 13 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Train separate classifiers C 1, C 2,..., C m given separate feature sets F 1, F 2,..., F m. Assume sentences inherit the document classification. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 13 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Train separate classifiers C 1, C 2,..., C m given separate feature sets F 1, F 2,..., F m. Assume sentences inherit the document classification. Train C 1 on the entire training set, and use it to predict labels for the training set. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 13 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Train separate classifiers C 1, C 2,..., C m given separate feature sets F 1, F 2,..., F m. Assume sentences inherit the document classification. Train C 1 on the entire training set, and use it to predict labels for the training set. Remove sentences with labels different from the assumption, and train C 2 on the remaining sentences. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 13 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Train separate classifiers C 1, C 2,..., C m given separate feature sets F 1, F 2,..., F m. Assume sentences inherit the document classification. Train C 1 on the entire training set, and use it to predict labels for the training set. Remove sentences with labels different from the assumption, and train C 2 on the remaining sentences. Continue iteratively until no more sentences can be removed. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 13 / 35

Multiple Naive Bayes Classifier Finding Opinion Sentences Identifying the Polarity Train separate classifiers C 1, C 2,..., C m given separate feature sets F 1, F 2,..., F m. Assume sentences inherit the document classification. Train C 1 on the entire training set, and use it to predict labels for the training set. Remove sentences with labels different from the assumption, and train C 2 on the remaining sentences. Continue iteratively until no more sentences can be removed. Five Feature Sets Starting with only words and adding in bigrams, trigrams, part of speech and polarity. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 13 / 35

Finding Opinion Sentences Identifying the Polarity Identifying the Polarity of Opinion Sentences What We Have Sentences that are distinguished as either opinions or facts. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 14 / 35

Finding Opinion Sentences Identifying the Polarity Identifying the Polarity of Opinion Sentences What We Have Sentences that are distinguished as either opinions or facts. What We Want Separate the opinion sentences into three classes Positive sentences. Negative Sentences. Neutral sentences. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 14 / 35

Finding Opinion Sentences Identifying the Polarity Identifying the Polarity of Opinion Sentences What We Have Sentences that are distinguished as either opinions or facts. What We Want Separate the opinion sentences into three classes Positive sentences. Negative Sentences. Neutral sentences. How We Do It By the number and strength of semantically oriented words (either positive or negative) in the sentence. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 14 / 35

Semantically Oriented Words Finding Opinion Sentences Identifying the Polarity Hypothesis Positive words co-occur more than expected by chance, and so do negative words. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 15 / 35

Semantically Oriented Words Finding Opinion Sentences Identifying the Polarity Hypothesis Positive words co-occur more than expected by chance, and so do negative words. Approach Measure the words co-occurence with words from a known seed set of semantically oriented words. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 15 / 35

Semantically Oriented Words Finding Opinion Sentences Identifying the Polarity Log-likelihood ratio L(W i, POS j ) = log ( Freq(Wi,POS j,adjp)+ɛ Freq(W all,pos j,adjp) Freq(W i,pos j,adjn)+ɛ Freq(W all,pos j,adjn) ) Where W i is a word in the sentence, ADJ p is positive seed word set, ADJ n is negative seed word set, POS j is part of speech collocation frequency ratio with ADJ p and ADJ n and ɛ is a smoothing constant (0.5). Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 16 / 35

Sentence Polarity Tagging Finding Opinion Sentences Identifying the Polarity Determine the orientation of an opinion sentence Specify cutoffs t p and t n. Calculate the sentences average log-likelihood score. Positive sentences have average scores greater than t p. Negative sentences have average scores lower than t n. Neutral sentences have average scores between t p and t n. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 17 / 35

Sentence Polarity Tagging Finding Opinion Sentences Identifying the Polarity Determine the orientation of an opinion sentence Specify cutoffs t p and t n. Calculate the sentences average log-likelihood score. Positive sentences have average scores greater than t p. Negative sentences have average scores lower than t n. Neutral sentences have average scores between t p and t n. Optimal t p and t n values Are obtained from the training data via density estimation, using a small subset of hand-labeled sentences. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 17 / 35

Seed Set Finding Opinion Sentences Identifying the Polarity Seed words used The seed words were subsets of 1.336 adjectives that were manually classified as either positive or negative. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 18 / 35

Seed Set Finding Opinion Sentences Identifying the Polarity Seed words used The seed words were subsets of 1.336 adjectives that were manually classified as either positive or negative. Seed Set Size To see whether seed set sizes would influence the result, seed sets of 1, 20, 100 and over 600 positive and negative pairs of adjectives were used. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 18 / 35

Layout 1 Goals of the Paper 2 Finding Opinion Sentences Identifying the Polarity 3 Data Evaluation Results 4 Related Work Article Evaluation Data Evaluation Results Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 19 / 35

Data Data Evaluation Results Data Used The data is from the TREC 8,9 and 11 collections, which consists of more than 1.7 million newswire articles from six different sources. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 20 / 35

Data Data Evaluation Results Data Used The data is from the TREC 8,9 and 11 collections, which consists of more than 1.7 million newswire articles from six different sources. Wall Street journal Some articles are marked with document type Editorial (2,877) Letter to Editor (1,695) Business (2,009) News (3,714) 2,000 articles from each type is randomly selected. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 20 / 35

Evaluation Metrics Data Evaluation Results Recall The fraction of the relevant documents that are retrieved. recall = {relevant documents} {retrieved documents} {relevant documents} Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 21 / 35

Evaluation Metrics Data Evaluation Results Recall The fraction of the relevant documents that are retrieved. recall = {relevant documents} {retrieved documents} {relevant documents} Precision The fraction of the retrieved documents that are relevant. precision = {relevant documents} {retrieved documents} {retrieved documents} Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 21 / 35

Evaluation Metrics Data Evaluation Results Recall The fraction of the relevant documents that are retrieved. recall = {relevant documents} {retrieved documents} {relevant documents} Precision The fraction of the retrieved documents that are relevant. precision = {relevant documents} {retrieved documents} {retrieved documents} F-measure The weighted harmonic mean of recall and precision. F = 2 precision recall (precision+recall) Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 21 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Precision = 1.00 Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Precision = 1.00, Recall = 0.5 Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Precision = 1.00, Recall = 0.5, F-Measure = 0.67 Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Precision = 1.00, Recall = 0.5, F-Measure = 0.67 Example 2 Retrieves all documents. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Precision = 1.00, Recall = 0.5, F-Measure = 0.67 Example 2 Retrieves all documents. Recall = 1.00 Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Precision = 1.00, Recall = 0.5, F-Measure = 0.67 Example 2 Retrieves all documents. Recall = 1.00, precision = 0.1 Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Examples Data Evaluation Results Common Attributes Body of 1,000 documents. 100 relevant documents. Example 1 50 retrieved documents, all relevant. Precision = 1.00, Recall = 0.5, F-Measure = 0.67 Example 2 Retrieves all documents. Recall = 1.00, precision = 0.1, F-Measure = 0.18 Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 22 / 35

Gold Standards Data Evaluation Results Document-level Standard Already available from Wall Streel Journal. News and Business is mapped to facts. Editorial and Letter to the Editor is mapped to opinions. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 23 / 35

Gold Standards Data Evaluation Results Document-level Standard Already available from Wall Streel Journal. News and Business is mapped to facts. Editorial and Letter to the Editor is mapped to opinions. Sentence-level Standard There is no automated standard that can distinguish between facts and opinions, or between positive and negative opinions. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 23 / 35

Gold Standards Data Evaluation Results Document-level Standard Already available from Wall Streel Journal. News and Business is mapped to facts. Editorial and Letter to the Editor is mapped to opinions. Sentence-level Standard There is no automated standard that can distinguish between facts and opinions, or between positive and negative opinions. Human evaluators classify a set of sentences between facts and opinions and determine the type of opinions. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 23 / 35

Topics and Articles Data Evaluation Results Topics Four topics are chosen for the evaluation Gun control Illegal aliens Social security Welfare reform Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 24 / 35

Topics and Articles Data Evaluation Results Topics Four topics are chosen for the evaluation Gun control Illegal aliens Social security Welfare reform Articles 25 articles were randomly chosen for each topic from the TREC corpus. The articles were found using the Lucene search engine. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 24 / 35

Sentences Data Evaluation Results Selection of Sentences Four sentences chosen from each document. The sentences were grouped into ten 50-sentence blocks. Each block shares ten sentences with the preceding and following block. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 25 / 35

Sentences Data Evaluation Results Selection of Sentences Four sentences chosen from each document. The sentences were grouped into ten 50-sentence blocks. Each block shares ten sentences with the preceding and following block. Standard A The 300 sentences appearing once, and one judgement from the remaining 100 sentences. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 25 / 35

Sentences Data Evaluation Results Selection of Sentences Four sentences chosen from each document. The sentences were grouped into ten 50-sentence blocks. Each block shares ten sentences with the preceding and following block. Standard A The 300 sentences appearing once, and one judgement from the remaining 100 sentences. Standard B The subset of the 100 sentences appearing twice, which were given identical labels. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 25 / 35

Data Evaluation Results Training The classifier was trained on 4,000 articles from WSJ and evaluated on other 4,000 articles. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 26 / 35

Data Evaluation Results Training The classifier was trained on 4,000 articles from WSJ and evaluated on other 4,000 articles. The result Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 26 / 35

Sentence Classification Data Evaluation Results Three Approaches Similarity approach Bayes classifier Multiple Bayes classifiers Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 27 / 35

Sentence Classification Data Evaluation Results Three Approaches Similarity approach Bayes classifier Multiple Bayes classifiers The Similarity Approach {recall, precision} Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 27 / 35

Sentence Classification Data Evaluation Results Bayes classifiers {recall, precision} Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 28 / 35

Sentence Classification Data Evaluation Results Seed Set Size Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 29 / 35

Polarity Classification Data Evaluation Results Accuracy of Sentence Polarity Tagging Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 30 / 35

Layout 1 Goals of the Paper 2 Finding Opinion Sentences Identifying the Polarity 3 Data Evaluation Results 4 Related Work Article Evaluation Related Work Article Evaluation Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 31 / 35

Article Related Work Article Evaluation Document Level A fairly straightforward Bayesian classifier using lexical information can distinguish between mostly factual and opinion documents with very high precision and recall. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 32 / 35

Article Related Work Article Evaluation Document Level A fairly straightforward Bayesian classifier using lexical information can distinguish between mostly factual and opinion documents with very high precision and recall. Sentence Level Three techniques were described for opinion/fact classification achieving up to 91% precision and recall on opinion sentences. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 32 / 35

Article Related Work Article Evaluation Document Level A fairly straightforward Bayesian classifier using lexical information can distinguish between mostly factual and opinion documents with very high precision and recall. Sentence Level Three techniques were described for opinion/fact classification achieving up to 91% precision and recall on opinion sentences. Polarity Examined an automatic method for assigning polarity information (positive, negative or neutral), which assigns the correct polarity in 90% of the cases. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 32 / 35

Related Work Related Work Article Evaluation Other work There is a lot of research in the area of automated opinion detection. Prior works include SimFinder and classification of subjective words. Recent works includes Chinese web opinion mining and german news article. Our Project - Herning Municipality Citizens entering the homecare system gets a function evaluation, in order to establish their needs for help. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 33 / 35

Relation to Our Project Related Work Article Evaluation Function Evaluation Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 34 / 35

Evaluation of the Article Related Work Article Evaluation The Good Good choice of titel. Good written description of the use of their methods. They keep a good flow through the article. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 35 / 35

Evaluation of the Article Related Work Article Evaluation The Good Good choice of titel. Good written description of the use of their methods. They keep a good flow through the article. The Not So Good No definition of recall and precision, not even a reference. SimFinder is presented as state-of-the-art. Made by one of the authors. Hong Yu, Vasileios Hatzivassiloglou Towards Answering Opinion Questions 35 / 35