NLP and Word Embeddings deeplearning.ai Word representation
Word representation V = [a, aaron,, zulu, <UNK>] 1-hot representation Man (5391) 1 Woman (9853) 1 King (4914) 1 Queen (7157) 1 Apple (456) 1 Orange (6257) 1 I want a glass of orange. I want a glass of apple.
Featurized representation: word embedding Man (5391) Woman (9853) King (4914) Queen (7157) Apple (456) Orange (6257) -.95.97..1.93.95 -.1..7.69.3 -.2.2.1.95.97 I want a glass of orange. I want a glass of apple.
Visualizing word embeddings man woman king queen dog cat fish three four one two apple grape orange [van der Maaten and Hinton., 28. Visualizing data using t-sne]
NLP and Word Embeddings deeplearning.ai Using word embeddings
Named entity recognition example 1 1 Sally Johnson is an orange farmer Robert Lin is an apple farmer
Transfer learning and word embeddings 1. Learn word embeddings from large text corpus. (1-1B words) (Or download pre-trained embedding online.) 2. Transfer embedding to new task with smaller training set. (say, 1k words) 3. Optional: Continue to finetune the word embeddings with new data.
Relation to face encoding $ (&) f($ (&) ) )* $ (() [Taigman et. al., 214. DeepFace: Closing the gap to human level performance] f($ (() )
NLP and Word Embeddings deeplearning.ai Properties of word embeddings
Analogies Man (5391) Woman (9853) King (4914) Queen (7157) Apple (456) Orange (6257) Gender 1 1 -.95.97..1 Royal.1.2.93.95 -.1. Age.3.2.7.69.3 -.2 Food.9.1.2.1.95.97 [Mikolov et. al., 213, Linguistic regularities in continuous space word representations]
Analogies using word vectors man king queen woman cat dog fish three two four one grape orange apple ( )*+ (,-)*+ ( /+1 (?
Cosine similarity 345((,, ( /+1 ( )*+ + (,-)*+ ) Man:Woman as Boy:Girl Ottawa:Canada as Nairobi:Kenya Big:Bigger as Tall:Taller Yen:Japan as Ruble:Russia
NLP and Word Embeddings deeplearning.ai Embedding matrix
Embedding matrix In practice, use specialized function to look up an embedding.
NLP and Word Embeddings deeplearning.ai Learning word embeddings
Neural language model I want a glass of orange. 4343 9665 1 3852 6163 6257 I * +,+, 4 5 +,+, want * -../ 4 5 -../ a * 4 5 glass *,1/2 4 5,1/2 of *.., 4 5.., orange *.2/3 4 5.2/3 [Bengio et. al., 23, A neural probabilistic language model]
Other context/target pairs I want a glass of orange juice to go along with my cereal. Context: Last 4 words. 4 words on left & right Last 1 word Nearby 1 word
NLP and Word Embeddings deeplearning.ai Word2Vec
Skip-grams I want a glass of orange juice to go along with my cereal. [Mikolov et. al., 213. Efficient estimation of word representations in vector space.]
Model Vocab size = 1,k
Problems with softmax classification! " # = % & ' ( ) * -.,... % &, ( ) * 1- How to sample the context #?
NLP and Word Embeddings deeplearning.ai Negative sampling
Defining a new learning problem I want a glass of orange juice to go along with my cereal. [Mikolov et. al., 213. Distributed representation of words and phrases and their compositionality]
Model Softmax:! " # = % & ' ( ) * -.,... % &, ( ) * 1- context orange orange orange orange orange word juice king book the of target? 1
Selecting negative examples context orange orange orange orange orange word juice king book the of target? 1
NLP and Word Embeddings deeplearning.ai GloVe word vectors
GloVe (global vectors for word representation) I want a glass of orange juice to go along with my cereal. [Pennington et. al., 214. GloVe: Global vectors for word representation]
Model
A note on the featurization view of word embeddings Man (5391) Woman (9853) King (4914) Queen (7157) Gender Royal Age Food 1.1.3.9 1.2.2.1 -.95.93.7.2.97.95.69.1 78,888 78,888 minimize ( ) *+, - *. + + * 2 6 + log ) *+ *:7 +:7
NLP and Word Embeddings deeplearning.ai Sentiment classification
Sentiment classification problem! " The dessert is excellent. Service was quite slow. Good for a quick meal, but nothing special. Completely lacking in good taste, good service, and good ambience.
Simple sentiment classification model The dessert is excellent 8928 2468 4694 318 The # $%&$, - $%&$ desert # &'($, - &'($ is # '(%', - '(%' excellent # )*$+, - )*$+ Completely lacking in good taste, good service, and good ambience.
RNN for sentiment classification "8 softmax : ;+< : ;*< : ;&< : ;)< : ;'< : ;*+< - *$6& - '%(( - ''&7 - )$$& - ))+,,,,, Completely lacking in good. ambience
NLP and Word Embeddings deeplearning.ai Debiasing word embeddings
The problem of bias in word embeddings Man:Woman as King:Queen Man:Computer_Programmer as Woman: Homemaker Father:Doctor as Mother: Nurse Word embeddings can reflect gender, ethnicity, age, sexual orientation, and other biases of the text used to train the model. [Bolukbasi et. al., 216. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings]
Addressing bias in word embeddings 1. Identify bias direction. 2. Neutralize: For every word that is not definitional, project to get rid of bias. 3. Equalize pairs. [Bolukbasi et. al., 216. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings]