News Authorship Identification with Deep Learning

Similar documents
System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

Lecture 1: Machine Learning Basics

Python Machine Learning

Assignment 1: Predicting Amazon Review Ratings

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Word Segmentation of Off-line Handwritten Documents

Rule Learning With Negation: Issues Regarding Effectiveness

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Reducing Features to Improve Bug Prediction

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Indian Institute of Technology, Kanpur

arxiv: v1 [cs.lg] 15 Jun 2015

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Linking Task: Identifying authors and book titles in verbose queries

Rule Learning with Negation: Issues Regarding Effectiveness

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

Speech Emotion Recognition Using Support Vector Machine

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Learning From the Past with Experiment Databases

Residual Stacking of RNNs for Neural Machine Translation

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Switchboard Language Model Improvement with Conversational Data from Gigaword

A Case Study: News Classification Based on Term Frequency

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

CS Machine Learning

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Modeling function word errors in DNN-HMM based LVCSR systems

A study of speaker adaptation for DNN-based speech synthesis

Second Exam: Natural Language Parsing with Neural Networks

Probabilistic Latent Semantic Analysis

Model Ensemble for Click Prediction in Bing Search Ads

(Sub)Gradient Descent

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Lecture 1: Basic Concepts of Machine Learning

A Comparison of Two Text Representations for Sentiment Analysis

Modeling function word errors in DNN-HMM based LVCSR systems

ScienceDirect. A Framework for Clustering Cardiac Patient s Records Using Unsupervised Learning Techniques

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

arxiv: v1 [cs.lg] 7 Apr 2015

TRANSFER LEARNING OF WEAKLY LABELLED AUDIO. Aleksandr Diment, Tuomas Virtanen

arxiv: v4 [cs.cl] 28 Mar 2016

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Deep Neural Network Language Models

arxiv: v1 [cs.cv] 10 May 2017

Generative models and adversarial training

Artificial Neural Networks written examination

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Dropout improves Recurrent Neural Networks for Handwriting Recognition

Attributed Social Network Embedding

CSL465/603 - Machine Learning

Semi-Supervised Face Detection

Softprop: Softmax Neural Network Backpropagation Learning

Genre classification on German novels

Knowledge Transfer in Deep Convolutional Neural Nets

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

HIERARCHICAL DEEP LEARNING ARCHITECTURE FOR 10K OBJECTS CLASSIFICATION

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Learning Methods in Multilingual Speech Recognition

INPE São José dos Campos

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Calibration of Confidence Measures in Speech Recognition

SARDNET: A Self-Organizing Feature Map for Sequences

arxiv: v2 [cs.cv] 30 Mar 2017

How to Judge the Quality of an Objective Classroom Test

A Vector Space Approach for Aspect-Based Sentiment Analysis

Universidade do Minho Escola de Engenharia

Online Updating of Word Representations for Part-of-Speech Tagging

Human Emotion Recognition From Speech

Cultivating DNN Diversity for Large Scale Video Labelling

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Dual-Memory Deep Learning Architectures for Lifelong Learning of Everyday Human Behaviors

CS 446: Machine Learning

The Good Judgment Project: A large scale test of different methods of combining expert predictions

arxiv: v3 [cs.cl] 7 Feb 2017

Australian Journal of Basic and Applied Sciences

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Mining Association Rules in Student s Assessment Data

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Axiom 2013 Team Description Paper

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

arxiv: v5 [cs.ai] 18 Aug 2015

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

TRANSFER LEARNING IN MIR: SHARING LEARNED LATENT REPRESENTATIONS FOR MUSIC AUDIO CLASSIFICATION AND SIMILARITY

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

Learning Methods for Fuzzy Systems

Probing for semantic evidence of composition by means of simple classification tasks

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Multi-Lingual Text Leveling

How to read a Paper ISMLL. Dr. Josif Grabocka, Carlotta Schatten

arxiv: v1 [cs.cl] 27 Apr 2016

Transcription:

000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050 051 052 053 News Authorship Identification with Deep Learning Liuyu Zhou Huafei Wang Electrical Engineering Aeronautics & Astronautics Stanford University lyzhou@stanford.edu huafei@stanford.edu Abstract Authorship identification identifies the most possible author from a group of candidate authors for academic articles, news, emails and forum messages. It can be applied to find the original author of an uncited article, to detect plagiarism and to classify spam / non- spam messages. In this project, we tackled this classification task in author level, article level, sentence level and word level with various deep and non-deep classification algorithms and GloVe word vectors are used as the pre-trained word vectors. Among all the algorithms, sentence-level Recurrent Neural Network (RNN) achieves the best performance since it captures the context information as well as word / sentence sequence information from the training dataset. 1 Introduction Authorship identification (authorship attribution) determines the likelihood of a piece of writing to be produced by a particular author by examining other writings by that author. Author Identification study is useful to identify the most plausible authors and to find evidences to support the conclusion. It can be applied in many tasks such as authorship characterization, detecting plagiarism, cybercriminal analysis and classifying spam / non-spam messages. This is a highly interdisciplinary area as it takes advantage of machine learning, information retrieval, and natural language processing. Authorship identification problem has been studied in the last few decades with various of methods and techniques. Most previous studies such as [1,2,3,4] used stylometric analysis techniques for analyzing and attributing authorship of literary texts. Stylistic features are used in the stylometric analysis which include attributes or writing-style markers that are the most effective discriminators of authorship. The vast array of stylistic features includes lexical, syntactic, structural, contentspecific, and idiosyncratic style markers [1]. Over 1,000 different features have been used in previous authorship analysis research, with no consensus on a best set of style markers. This could be attributable to certain feature categories being more effective at capturing style variations in different contexts. This necessitates the use of larger feature sets comprised of several categories of features (e.g., punctuation, word-length distributions, etc.) spanning various feature groups (i.e., lexical, syntactic, etc.) [1]. As a result, the performance of authorship identification tasks depends highly on the chosen features and the quality of these features. Therefore reliable and efficient techniques are needed to extract these features. For learning, we have both supervised learning and unsupervised learning at our disposal. Supervised learning are those that require author-class labels for classification, while unsupervised techniques make classification with no prior knowledge of author classes. In this paper, we only focus on the supervised learning paradigm. 1

054 055 056 057 058 059 060 061 062 063 064 065 066 067 068 069 070 071 072 073 074 075 076 077 078 079 080 081 082 083 084 085 086 087 088 089 090 091 092 093 094 095 096 097 098 099 100 101 102 103 104 105 106 107 Supervised learning methods used in previous studies include regularized least squares, support vector machine (SVM), decision tree, feed-forward neural network and etc. on various datasets (Twitter, RCV1, PAN 2012 and etc.). The accuracy varies significantly when different approaches and datasets are used from low 20s to high 90s. This paper researches the news authorship identification problem by exploring different machine learning algorithms on different levels Article-level linguistic features represented by average word length and etc. with non-deep machine learning algorithms (e.g. support vector machine). Word-level features represented by word vectors with Global Vectors for Word Representation (GloVe) with different classifiers (e.g. nearest neighbor). Recurrent Neural Network (RNN) with pre-trained GloVe word vectors. A sentence vector is generated as the input in each step. The number of steps is the number of sentences in the article. GloVe pre-trained word vectors are used since GloVe captures the word context information (such as word similarity). We mainly focus on Recurrent Neural Network (RNN) on GloVe word vectors since we believe that RNN can also capture the word / sentence sequence information which can help us to better classifiy the authorship of articles. The experiment results confirm this assumption. We focus on news articles since news articles provide not only stylometry features as in other types of texts, but also context information since journalists tend to focus on a narrow range of topics and thus context can also be leveraged for the identification purpose. 2 Approach 2.1 Dataset A subset of the Reuters RCV1 news article dataset is used to develop our multi-level machine learning algorithms. The dataset contains 5000 news articles for 50 different journalism authors (100 texts per author) and pre-splits the dataset 50-50 for training and testing. The dataset is obtained from the Center for Machine Learning and Intelligent Systems hosted by the Unviersity of California, Irvine. 2.2 Baseline High-level linguistic features of articles are used as attributes in our baseline model. The following stylometry features are used: Average Word Length, Average Sentence Length, Hapax Legomenon Ratio (fraction of unique words). Traditional machine learning methods including Support Vector Machines, Naive Bayes, Random Forest and etc. are implemented utilizing the above features. 2.3 GloVe GloVe is an unsupervised learning algorithm for obtaining vector representations for words. Training is performed on aggregated global word-word co-occurrence statistics from a corpus, and the resulting representations showcase interesting linear substructures of the word vector space [5]. In this project, GloVe pre-trained word vectors are used to encode the training articles and find the most representative vector that can represent the author. Then the test articles are encoded using the same GloVe word vectors to obtain the most representative vector that can represent the article. Different GloVe word vector dimensions are used to evaluate their influence on the performance of the algorithm. F1 Score is used for evaluation. 2.4 Recurrent Neural Network Recurrent Neural Network (RNN) is also implemented to tackle this task. GloVe pre-trained word vectors are used to encode each word in the article. The standard model - sigmoid hidden layer and 2

108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 softmax projection layer are used, which are as follows. e (t) = x (t) L (1) ) h (t) = sigmoid (h (t 1) H + e (t) I + b 1 (2) ) ŷ (t) = softmax (h (t) U + b 2 (3) The input vector in each step is a pre-trained constant vector which can represent a word, a sentence, a paragraph or even an article, which will be discussed in detail in the next section. A more complex RNN model LSTM / GRU with forget gate, input gate and output gate is also implemented in our model. This approach is adopted since we believe that some words in the article are not representative to the author so thay may pose no effect or even negative effect on the prediction results while some words are truely indicative of the author and they should be assigned higher weights. Since the size of our dataset is moderate, dropout is applied to the input and output of the RNN model in order to mitigate the undesirable overfitting. 3 Experiment 3.1 Baseline Several machine learning methods including Support Vector Machines, Naive Bayes, Random Forestd using the stylometry features are implemented. The best accuracy is achieved by Gradient Boosting Classifier with an 12.24% accuracy. The low accuracy is expected as only 3 high level summary features are used. As discussed in the Introduction, over 1000 different stylometry features are used in previous studies in order to have a satisfactory classification performance. In addition, word, sentence and paragraph contexts as well as sequences that may contain author writing style and pattern information are discarded during the summarization and feature extraction process. This preliminary result serves as a baseline for this project. 3.2 GloVe The F1 Score and confusion matrix visualizations with the Nearest Neighbor Classifier (L 2 Distance) are shown below. We can observe that the classification with GloVe greatly outperforms the baseline methods because of the different context information associated with each author that is able to be extracted and aggregated from their articles using GloVe. It can also be observed that the higher the word dimension is, the higher the F 1 score the result has. The highest F 1 score of the Nearest Neighbor Classifier comes with GloVe word vectors with dimension 300, which is around 0.46. This is reasonable since higher dimension word vector can generally capture more context information of a word. 3

162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 Figure 1: Confusion Matrix with GloVe D = 50 Figure 2: Confusion Matrix with GloVe D = 100 Figure 3: Confusion Matrix with GloVe D = 200 Figure 4: Confusion Matrix with GloVe D = 300 3.3 RNN The following two approaches of data input are implemented for RNN Word level approach. A single word vector is fed into RNN in one step. So the number of steps of RNN is the number of words in each article. Since there are typically a large number of words in each article, there will be too many steps, which has caused the gradient vanishing problem. The result we have obtained is poor, with the highest F 1 score around 0.2. Besides, since each article has a different length, different models for each article have to be built separately. As a result, the training is undesirably slow. Paddings have been added to the articles to make them have the same length. But unfortunately, since the length of articles covers a wide range - from 200 to 1500, adding paddings does not enhance the performance significantly. Sentence level approach. The mean vector of each sentence is calculated and fed into the RNN model at each step. This approach also captures the word / sentence sequence information since the RNN process is recurrent. Because that the average number of sentences of all articles is around 15-20, this model does not suffer from the gradient valishing problem, so it significantly outperforms the word level approach. The highest F 1 score we have obtained is around 0.6. In addition, the training is relatively fast compared with the word level RNN approach. Confusion matrix visualizations of sentence level RNN with different GloVe word vector dimensions are shown below. 4

216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 Figure 5: Confusion Matrix for RNN (D = 50) Figure 6: Confusion Matrix for RNN (D = 100) Figure 7: Confusion Matrix for RNN (D = 200) Figure 8: Confusion Matrix for RNN (D = 300) We can see from the figures that the sentence level RNN Classifier with GloVe word vectors of dimension 300 has the best performance, with the F 1 score around 0.6. The training loss and validation loss values of sentence level RNN are shown below Figure 9: Loss History for RNN It can be observed from the figure that despite some small spikes in the curve, the training loss value keeps decreasing. The validation loss value drops to a minimum value and then starts to increase. 5

270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 Typically we can get the best model for prediction (test accuracy) when the validation loss reaches the minimum value. Usually the training algorithm will terminate at the first small spike of the validation loss curve. But in order to show the full trend of the training and validation loss, more epoches have been run in our attempt. 3.4 Comparing with Other Machine Learning Algorithms Several other machine learning algorithms that are typically powerful in various of prediction tasks are also implemented. These algorithms include Random Forest Classifier, Extra Tree Classifier, Decision Tree Classifier, AdaBoost Classifier, and Gradient Boosting Classifier. For these methods, GloVe word vectors are used to encode the articles and calculate the mean vector for each article and predictions are based on the calculated mean vector. The comparison results of different machine learning algorithms are shown below. 4 Conclusion Figure 10: Different ML Methods with GloVe We have presented various methods for authorship identification task with a focus on the Recurrent Neural Network approach. Stylometry features lose information and performs poorly in the authorship identification task. GloVe captures context information for different authors and performs relatively better. RNN, in addition to capturing context information, also captures word / sentence sequence information and has the best performance with a 0.6 F 1 score. However, the RNN approach does not show significant advantage over other machine learning methods. For future work, we can attempt to improve the performance of RNN to truely unleash its power. Some strategies are as follows. Use a larger dataset. This is extremely important since our model is already suffering from overfitting problem and a larger dataset can help us remedy the situation. Currently, we are using GloVe pre-trained vectors as our constant word vectors in our model. We can let the GloVe pre-trained vectors be the initial values for the word embedding variables and update them in each RNN step. For large enough dataset, we can use neural network to train our own word vectors instead of using the pre-trained GloVe word vectors. 6

324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 References The more complex RNN models such as stacked RNN, bidirectional RNN, inductive transfer can be applied with a corresponding large dataset. [1] Abbasi A, Chen H. Writeprints: A stylometric approach to identity-level identification and similarity detection in cyberspace[j]. ACM Transactions on Information Systems (TOIS), 2008, 26(2): 7. [2] Castro A, Lindauer B. Author Identification on Twitter[J]. 2012. [3] Narayanan A, Paskov H, Gong N Z, et al. On the feasibility of internet-scale author identification[c]. Security and Privacy (SP), 2012 IEEE Symposium on. IEEE, 2012: 300-314. [4] Nirkhi S, Dharaskar R V. Comparative study of authorship identification techniques for cyber forensics analysis[j]. arxiv preprint arxiv:1401.6118, 2013. [5] Pennington J, Socher R, Manning C D. Glove: Global Vectors for Word Representation[C]. EMNLP. 2014, 14: 1532-1543. 7