Integrating Distributional Lexical Contrast into Word Embeddings for Antonym Synonym Distinction Kim Anh Nguyen, Sabine Schulte im Walde, Ngoc Thang Vu Institute for Natural Language Processing (IMS) University of Stuttgart ACL-2016, Berlin
Overview 1. Introduction 1.1. Word Vector Representations 1.2. Antonym-Synonym Distinction Task 2. Contributions 2.1. Improving Weights of Feature Vectors 2.2. Distributional Lexical Contrast Embeddings Model 3. Experiments 4. Conclusion ACL-2016: Antonym-Synonym Distinction 2 / 20
1.1. Word Vector Representations 1.1.1. Distributional Semantic Model (DSM) A means to represent meaning vectors of words. DSM rely on the distributional hypothesis. (Harris, 1954) Words with similar distributions have related meanings. Each weighted feature can be: Co-occurrence frequency. Association measure: local mutual information (LMI) (Evert, 2005) 1.1.2. Word Embeddings Representing words as low-dimensional dense vectors. Words with similar distributions have similar vectors. ACL-2016: Antonym-Synonym Distinction 3 / 20
1.2. Antonym-Synonym Distinction Task Goal Distinguishing antonyms from synonyms. Problems Causes DSM tend to capture both antonyms (formal-informal) and synonyms (formal-conventional). Word embeddings represent vectors of both antonyms and synonyms as similar vectors. Antonymy and synonymy are paradigmatic relations. Antonyms and synonyms often occur in similar contexts. ACL-2016: Antonym-Synonym Distinction 4 / 20
Outline 1. Introduction 1.1. Word Vector Representations 1.2. Antonym-Synonym Distinction Task 2. Contributions 2.1. Improving Weights of Feature Vectors 2.2. Distributional Lexical Contrast Embeddings Model 3. Experiments 4. Conclusion ACL-2016: Antonym-Synonym Distinction 5 / 20
2.1. Improving Weights of Feature Vectors Goal: Solution: Improving the quality of weighted feature vectors. Strengthening most salient features in the vectors. Using the lexical contrast information of the target words and their contexts. Proposing the new weight for feature vectors. Representing words based on DSM with positive LMI. For each target word w: Determining the sets of antonyms A(w) and synonyms S(w). Determining the set of shared words W(f ) for each feature f. Computing the new weight (called weight SA ) as follows: ACL-2016: Antonym-Synonym Distinction 6 / 20
2.1. Improving Weights of Feature Vectors weight SA 1 (w, f ) = #(w,u) u W(f ) S(w) sim(w, u) 1 #(w,v) Average similarity to synonyms w A(w) v W(f ) S(w ) sim(w, v) Average similarity to antonyms ACL-2016: Antonym-Synonym Distinction 7 / 20
2.1. Improving Weights of Feature Vectors w = formal S(w) = {methodical, precise, conventional,...} f = conception f = issue A(w) = w = informal S(w ) = {irregular, unofficial, unconventional,...} f = rumor
2.1. Improving Weights of Feature Vectors w = formal S(w) = {methodical, precise, conventional,...} f = conception f = issue A(w) = w = informal S(w ) = {irregular, unofficial, unconventional,...} f = rumor weight SA (formal, conception) weight SA (formal, issue) 0 weight SA (formal, rumor) weight SA (w, f ) = 1 1 #(w,u) u W(f ) S(w) sim(w, u) #(w,v) w A(w) v W(f ) S(w ) sim(w, v) ACL-2016: Antonym-Synonym Distinction 8 / 20
2.2. Distributional Lexical Contrast Embeddings Model (dlce) Aims: Solution: Learning word embeddings. Moving synonyms closer to each other in space. Moving antonyms further away from each other in space. Integrating distributional lexical contrast into the transformation of Skip-gram model (Mikolov et al. 2013, Levy et al., 2014). Applying lexical contrast to every single context of the target word. The proposed objective function as follows: ACL-2016: Antonym-Synonym Distinction 9 / 20
w V c V 2.2. Distributional Lexical Contrast Embeddings Model (dlce) Distribution of target word and contexts Distribution of negative contexts { ( #(w, c) log σ(sim(w, c)) + k#(w)p 0 (c) log σ( sim(w, c)) ) 1 +( #(w,u) u W(c) S(w) sim(w, u) 1 #(w,v) v W(c) A(w) sim(w, v) ) } Distribution of synonymous pairs Distribution of antonymous pairs ACL-2016: Antonym-Synonym Distinction 10 / 20
Outline 1. Introduction 1.1. Word Vector Representations 1.2. Antonym-Synonym Distinction Task 2. Contributions 2.1. Improving Weights of Feature Vectors 2.2. Distributional Lexical Contrast Embeddings Model 3. Experiments 4. Conclusion ACL-2016: Antonym-Synonym Distinction 11 / 20
3. Experiments Evaluating weight SA on Antonym-Syntonym distinction task. Evaluating effects of dlce model: Antonym-Syntonym distinction task. Similarity task. ACL-2016: Antonym-Synonym Distinction 12 / 20
3.1. Antonym Synonym Distinction Corpus: ENCOW14A (Schäfer and Bildhauer, 2012) contains 14.5 billion tokens. Dataset: a gold standard resource of paradigmatic relation pairs (Roth and Schulte im Walde, 2014) Word Class Ant-pairs Syn-pairs Total Adjective 300 300 600 Noun 350 350 700 Verb 400 400 800 Using average precision (AP) to evaluate. Using box-plots to compare the cosine medians of antonymous vs. synonymous pairs. ACL-2016: Antonym-Synonym Distinction 13 / 20
3.1. Antonym Synonym Distinction AP evaluation results 1 : Adjectives Nouns Verbs ANT SYN ANT SYN ANT SYN LMI 0.46 0.56 0.42 0.60 0.42 0.62 weight SA 0.36 0.75 0.40 0.66 0.38 0.71 LMI + SVD 0.46 0.55 0.46 0.55 0.44 0.58 weight SA + SVD 0.36 0.76 0.40 0.66 0.38 0.70 1 χ 2, p <.001, p <.005, p <.05 ACL-2016: Antonym-Synonym Distinction 14 / 20
3.1. Antonym Synonym Distinction Results in box-plots: ADJ NOUN VERB 0.5 0.0 0.5 1.0 LMI WeightSA LMI_SVD WeightSA_SVD LMI WeightSA LMI_SVD WeightSA_SVD LMI WeightSA LMI_SVD WeightSA_SVD Cosine ANT SYN ACL-2016: Antonym-Synonym Distinction 15 / 20
3.2. Effects of dlce model 3.2.1. Antonym-Synonym Distinction: Dataset: the gold standard resource of paradigmatic relation pairs (Roth and Schulte im Walde, 2014). Using area under curve (AUC) to identify antonyms. Comparison Models: Skip-gram (SGNS), mlcm (Pham et al., 2015) Results: Adjectives Nouns Verbs SGNS 0.64 0.66 0.65 mlcm 0.85 0.69 0.71 dlce 0.90 0.72 0.81 ACL-2016: Antonym-Synonym Distinction 16 / 20
3.2. Effects of dlce model 3.2.2. Similarity Task: Dataset: SimLex-999 (Hill et al., 2015) Using Spearman correlation coefficient ρ to evaluate. Comparison Models: SGNS, mlcm. Results: SGNS mlcm dlce 0.38 0.51 0.59 ACL-2016: Antonym-Synonym Distinction 17 / 20
Outline 1. Introduction 1.1. Word Vector Representations 1.2. Antonym-Synonym Distinction Task 2. Contributions 2.1. Improving Weights of Feature Vectors 2.2. Distributional Lexical Contrast Embeddings Model 3. Experiments 4. Conclusion ACL-2016: Antonym-Synonym Distinction 18 / 20
4. Conclusion We have presented two methods to address the task of antonym-synonym distinction: Improving the quality of weighted feature vectors. Integrating distributional lexical contrast into word embeddings. The results from the experiments show that our approaches can model semantic similarity and distinguish between antonyms and synonyms. ACL-2016: Antonym-Synonym Distinction 19 / 20
Thank you! ACL-2016: Antonym-Synonym Distinction 20 / 20