NLP and Word Embeddings. deeplearning.ai. Word representation

Similar documents
Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

arxiv: v1 [cs.cl] 20 Jul 2015

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

A Vector Space Approach for Aspect-Based Sentiment Analysis

Second Exam: Natural Language Parsing with Neural Networks

LIM-LIG at SemEval-2017 Task1: Enhancing the Semantic Similarity for Arabic Sentences with Vectors Weighting

There are some definitions for what Word

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Probing for semantic evidence of composition by means of simple classification tasks

Build on students informal understanding of sharing and proportionality to develop initial fraction concepts.

Joint Learning of Character and Word Embeddings

Modeling function word errors in DNN-HMM based LVCSR systems

Python Machine Learning

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Autoencoder and selectional preference Aki-Juhani Kyröläinen, Juhani Luotolahti, Filip Ginter

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Psychology and Language

arxiv: v2 [cs.ir] 22 Aug 2016

ON THE USE OF WORD EMBEDDINGS ALONE TO

Semantic and Context-aware Linguistic Model for Bias Detection

Deep Neural Network Language Models

Modeling function word errors in DNN-HMM based LVCSR systems

RATES2015. campingbarcelona. Camping. Normal. Plus. Bus

A Simple VQA Model with a Few Tricks and Image Features from Bottom-up Attention

Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

Attributed Social Network Embedding

NEURAL DIALOG STATE TRACKER FOR LARGE ONTOLOGIES BY ATTENTION MECHANISM. Youngsoo Jang*, Jiyeon Ham*, Byung-Jun Lee, Youngjae Chang, Kee-Eung Kim

been each get other TASK #1 Fry Words TASK #2 Fry Words Write the following words in ABC order: Write the following words in ABC order:

A JOINT MANY-TASK MODEL: GROWING A NEURAL NETWORK FOR MULTIPLE NLP TASKS

A deep architecture for non-projective dependency parsing

Human-like Natural Language Generation Using Monte Carlo Tree Search

Word Embedding Based Correlation Model for Question/Answer Matching

Cogat Sample Questions Grade 2

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

A Latent Semantic Model with Convolutional-Pooling Structure for Information Retrieval

arxiv: v1 [cs.cl] 2 Apr 2017

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Residual Stacking of RNNs for Neural Machine Translation

ASSISTIVE COMMUNICATION

arxiv: v1 [cs.cv] 10 May 2017

Unsupervised Cross-Lingual Scaling of Political Texts

P a g e 1. Grade 5. Grant funded by:

Text-mining the Estonian National Electronic Health Record

arxiv: v5 [cs.ai] 18 Aug 2015

We are going to talk about the meaning of the word weary. Then we will learn how it can be used in different sentences.

The Role of Semantic and Discourse Information in Learning the Structure of Surgical Procedures

THE world surrounding us involves multiple modalities

Name: Class: Date: ID: A

Linking Task: Identifying authors and book titles in verbose queries

arxiv: v4 [cs.cl] 28 Mar 2016

Mock Trial Preparation In-Class Assignment to Prepare Direct and Cross Examination Roles 25 September 2015 DIRECT EXAMINATION

Routledge Library Editions: The English Language: Pronouns And Word Order In Old English: With Particular Reference To The Indefinite Pronoun Man

FBK-HLT-NLP at SemEval-2016 Task 2: A Multitask, Deep Learning Approach for Interpretable Semantic Textual Similarity

Counting To 120 First Grade

arxiv: v3 [cs.cl] 24 Apr 2017

An Empirical and Computational Test of Linguistic Relativity

Context Free Grammars. Many slides from Michael Collins

Using Proportions to Solve Percentage Problems I

Conteúdos de inglês para o primeiro bimestre. Turma 21. Turma 31. Turma 41

Student Handbook. This handbook was written for the students and participants of the MPI Training Site.

Objective: Add decimals using place value strategies, and relate those strategies to a written method.

Boosting Named Entity Recognition with Neural Character Embeddings

L1 and L2 acquisition. Holger Diessel

Assignment 1: Predicting Amazon Review Ratings

Leveraging Sentiment to Compute Word Similarity

Construction Grammar. University of Jena.

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

Lecture 1: Machine Learning Basics

Middle School Newsletter

Dragonwings by Laurence Yep

arxiv: v2 [cs.cl] 26 Mar 2015

Truth Inference in Crowdsourcing: Is the Problem Solved?

Function Tables With The Magic Function Machine

Lip Reading in Profile

Performance Analysis of Optimized Content Extraction for Cyrillic Mongolian Learning Text Materials in the Database

Topic Modelling with Word Embeddings

Georgetown University at TREC 2017 Dynamic Domain Track

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

OUCH! That Stereotype Hurts Cultural Competence & Linguistic Training Summary of Evaluation Results June 30, 2014

English-German Medical Dictionary And Phrasebook By A.H. Zemback

1. READING ENGAGEMENT 2. ORAL READING FLUENCY

Films for ESOL training. Section 2 - Language Experience

How to Use Vocabulary Maps to Deliver Explicit Vocabulary Instruction: A Guide for Teachers

Answer each question by placing an X over the appropriate answer. Select only one answer for each question.

Agricultural Production, Business, and Trade in Spain and France ECON 496

Prewriting: Drafting: Revising: Editing: Publishing:

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

What Can Neural Networks Teach us about Language? Graham Neubig a2-dlearn 11/18/2017

BASIC ENGLISH. Book GRAMMAR

Theme 10. THEME 10: We Can Do It!

Lesson Overview: This lesson will introduce what a possessive pronoun is by reviewing

German (65 marks) Coimisiún na Scrúduithe Stáit State Examinations Commission. Leaving Certificate Applied General Directions

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Switchboard Language Model Improvement with Conversational Data from Gigaword

Applications of memory-based natural language processing

arxiv: v3 [cs.cl] 7 Feb 2017

Transcription:

NLP and Word Embeddings deeplearning.ai Word representation

Word representation V = [a, aaron,, zulu, <UNK>] 1-hot representation Man (5391) 1 Woman (9853) 1 King (4914) 1 Queen (7157) 1 Apple (456) 1 Orange (6257) 1 I want a glass of orange. I want a glass of apple.

Featurized representation: word embedding Man (5391) Woman (9853) King (4914) Queen (7157) Apple (456) Orange (6257) -.95.97..1.93.95 -.1..7.69.3 -.2.2.1.95.97 I want a glass of orange. I want a glass of apple.

Visualizing word embeddings man woman king queen dog cat fish three four one two apple grape orange [van der Maaten and Hinton., 28. Visualizing data using t-sne]

NLP and Word Embeddings deeplearning.ai Using word embeddings

Named entity recognition example 1 1 Sally Johnson is an orange farmer Robert Lin is an apple farmer

Transfer learning and word embeddings 1. Learn word embeddings from large text corpus. (1-1B words) (Or download pre-trained embedding online.) 2. Transfer embedding to new task with smaller training set. (say, 1k words) 3. Optional: Continue to finetune the word embeddings with new data.

Relation to face encoding $ (&) f($ (&) ) )* $ (() [Taigman et. al., 214. DeepFace: Closing the gap to human level performance] f($ (() )

NLP and Word Embeddings deeplearning.ai Properties of word embeddings

Analogies Man (5391) Woman (9853) King (4914) Queen (7157) Apple (456) Orange (6257) Gender 1 1 -.95.97..1 Royal.1.2.93.95 -.1. Age.3.2.7.69.3 -.2 Food.9.1.2.1.95.97 [Mikolov et. al., 213, Linguistic regularities in continuous space word representations]

Analogies using word vectors man king queen woman cat dog fish three two four one grape orange apple ( )*+ (,-)*+ ( /+1 (?

Cosine similarity 345((,, ( /+1 ( )*+ + (,-)*+ ) Man:Woman as Boy:Girl Ottawa:Canada as Nairobi:Kenya Big:Bigger as Tall:Taller Yen:Japan as Ruble:Russia

NLP and Word Embeddings deeplearning.ai Embedding matrix

Embedding matrix In practice, use specialized function to look up an embedding.

NLP and Word Embeddings deeplearning.ai Learning word embeddings

Neural language model I want a glass of orange. 4343 9665 1 3852 6163 6257 I * +,+, 4 5 +,+, want * -../ 4 5 -../ a * 4 5 glass *,1/2 4 5,1/2 of *.., 4 5.., orange *.2/3 4 5.2/3 [Bengio et. al., 23, A neural probabilistic language model]

Other context/target pairs I want a glass of orange juice to go along with my cereal. Context: Last 4 words. 4 words on left & right Last 1 word Nearby 1 word

NLP and Word Embeddings deeplearning.ai Word2Vec

Skip-grams I want a glass of orange juice to go along with my cereal. [Mikolov et. al., 213. Efficient estimation of word representations in vector space.]

Model Vocab size = 1,k

Problems with softmax classification! " # = % & ' ( ) * -.,... % &, ( ) * 1- How to sample the context #?

NLP and Word Embeddings deeplearning.ai Negative sampling

Defining a new learning problem I want a glass of orange juice to go along with my cereal. [Mikolov et. al., 213. Distributed representation of words and phrases and their compositionality]

Model Softmax:! " # = % & ' ( ) * -.,... % &, ( ) * 1- context orange orange orange orange orange word juice king book the of target? 1

Selecting negative examples context orange orange orange orange orange word juice king book the of target? 1

NLP and Word Embeddings deeplearning.ai GloVe word vectors

GloVe (global vectors for word representation) I want a glass of orange juice to go along with my cereal. [Pennington et. al., 214. GloVe: Global vectors for word representation]

Model

A note on the featurization view of word embeddings Man (5391) Woman (9853) King (4914) Queen (7157) Gender Royal Age Food 1.1.3.9 1.2.2.1 -.95.93.7.2.97.95.69.1 78,888 78,888 minimize ( ) *+, - *. + + * 2 6 + log ) *+ *:7 +:7

NLP and Word Embeddings deeplearning.ai Sentiment classification

Sentiment classification problem! " The dessert is excellent. Service was quite slow. Good for a quick meal, but nothing special. Completely lacking in good taste, good service, and good ambience.

Simple sentiment classification model The dessert is excellent 8928 2468 4694 318 The # $%&$, - $%&$ desert # &'($, - &'($ is # '(%', - '(%' excellent # )*$+, - )*$+ Completely lacking in good taste, good service, and good ambience.

RNN for sentiment classification "8 softmax : ;+< : ;*< : ;&< : ;)< : ;'< : ;*+< - *$6& - '%(( - ''&7 - )$$& - ))+,,,,, Completely lacking in good. ambience

NLP and Word Embeddings deeplearning.ai Debiasing word embeddings

The problem of bias in word embeddings Man:Woman as King:Queen Man:Computer_Programmer as Woman: Homemaker Father:Doctor as Mother: Nurse Word embeddings can reflect gender, ethnicity, age, sexual orientation, and other biases of the text used to train the model. [Bolukbasi et. al., 216. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings]

Addressing bias in word embeddings 1. Identify bias direction. 2. Neutralize: For every word that is not definitional, project to get rid of bias. 3. Equalize pairs. [Bolukbasi et. al., 216. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings]