NEURAL DIALOG STATE TRACKER FOR LARGE ONTOLOGIES BY ATTENTION MECHANISM. Youngsoo Jang*, Jiyeon Ham*, Byung-Jun Lee, Youngjae Chang, Kee-Eung Kim

Similar documents
System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

arxiv: v4 [cs.cl] 28 Mar 2016

Probabilistic Latent Semantic Analysis

Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models

The stages of event extraction

POS tagging of Chinese Buddhist texts using Recurrent Neural Networks

Context Free Grammars. Many slides from Michael Collins

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Modeling function word errors in DNN-HMM based LVCSR systems

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Python Machine Learning

arxiv: v3 [cs.cl] 24 Apr 2017

Heuristic Sample Selection to Minimize Reference Standard Training Set for a Part-Of-Speech Tagger

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Second Exam: Natural Language Parsing with Neural Networks

Modeling function word errors in DNN-HMM based LVCSR systems

Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY DOWNLOAD EBOOK : ADVANCED MACHINE LEARNING WITH PYTHON BY JOHN HEARTY PDF

First Grade Curriculum Highlights: In alignment with the Common Core Standards

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

arxiv: v1 [cs.cl] 2 Apr 2017

Calibration of Confidence Measures in Speech Recognition

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Loughton School s curriculum evening. 28 th February 2017

Rule Learning With Negation: Issues Regarding Effectiveness

A New Perspective on Combining GMM and DNN Frameworks for Speaker Adaptation

BANGLA TO ENGLISH TEXT CONVERSION USING OPENNLP TOOLS

Lecture 1: Machine Learning Basics

Dialog-based Language Learning

Using dialogue context to improve parsing performance in dialogue systems

Speech Recognition at ICSI: Broadcast News and beyond

Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

Houghton Mifflin Reading Correlation to the Common Core Standards for English Language Arts (Grade1)

Word Segmentation of Off-line Handwritten Documents

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Developing Grammar in Context

California Department of Education English Language Development Standards for Grade 8

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Глубокие рекуррентные нейронные сети для аспектно-ориентированного анализа тональности отзывов пользователей на различных языках

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Taught Throughout the Year Foundational Skills Reading Writing Language RF.1.2 Demonstrate understanding of spoken words,

Attributed Social Network Embedding

AQUA: An Ontology-Driven Question Answering System

Rule Learning with Negation: Issues Regarding Effectiveness

A study of speaker adaptation for DNN-based speech synthesis

arxiv: v5 [cs.ai] 18 Aug 2015

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

Prediction of Maximal Projection for Semantic Role Labeling

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

arxiv: v1 [cs.lg] 7 Apr 2015

WHEN THERE IS A mismatch between the acoustic

CS 598 Natural Language Processing

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

Residual Stacking of RNNs for Neural Machine Translation

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Learning Methods in Multilingual Speech Recognition

BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

Sample Goals and Benchmarks

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Cross Language Information Retrieval

Memory-based grammatical error correction

Rule-based Expert Systems

Deep Neural Network Language Models

A Case Study: News Classification Based on Term Frequency

cmp-lg/ Jan 1998

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Assignment 1: Predicting Amazon Review Ratings

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

arxiv: v1 [cs.cv] 10 May 2017

Matching Similarity for Keyword-Based Clustering

Grammars & Parsing, Part 1:

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

This scope and sequence assumes 160 days for instruction, divided among 15 units.

IN THIS UNIT YOU LEARN HOW TO: SPEAKING 1 Work in pairs. Discuss the questions. 2 Work with a new partner. Discuss the questions.

Advanced Grammar in Use

Model Ensemble for Click Prediction in Bing Search Ads

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

BASIC ENGLISH. Book GRAMMAR

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

SEMAFOR: Frame Argument Resolution with Log-Linear Models

The College Board Redesigned SAT Grade 12

Online Updating of Word Representations for Part-of-Speech Tagging

Writing a composition

Applications of memory-based natural language processing

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Mining Association Rules in Student s Assessment Data

Knowledge Transfer in Deep Convolutional Neural Nets

Linking Task: Identifying authors and book titles in verbose queries

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Georgetown University at TREC 2017 Dynamic Domain Track

Author: Justyna Kowalczys Stowarzyszenie Angielski w Medycynie (PL) Feb 2015

Transcription:

NEURAL DIALOG STATE TRACKER FOR LARGE ONTOLOGIES BY ATTENTION MECHANISM Youngsoo Jang*, Jiyeon Ham*, Byung-Jun Lee, Youngjae Chang, Kee-Eung Kim School of Computing KAIST Daejeon, South Korea ABSTRACT This paper presents a dialog state tracker submitted to Dialog State Tracking Challenge 5 (DSTC 5) with details. To tackle the challenging cross-language human-human dialog state tracking task with limited training data, we propose a tracker that focuses on words with meaningful context based on attention mechanism and bi-directional long short term memory (LSTM). The vocabulary including a plenty of proper nouns is vectorized with a sufficient amount of related texts crawled from web to learn a good embedding for words not existent in training dialogs. Despite its simplicity, our proposed tracker succeeded to achieve high accuracy without sophisticated pre- and post-processing. Index Terms Recurrent Neural Network, Dialog state tracking, DSTC5, Attention mechanism, Word embedding 1. INTRODUCTION The dialog state tracking challenge (DSTC) is a research challenge of recognizing the intentions of the users from noisy utterances in task-oriented dialogs. Unlike the previous challenges (DSTC 2&3) where the human-system dialogs were given, provided dialogs are replaced with human-human dialogs in DSTC 4&5 widening the diversity of expressions. The cross-language feature is further added in DSTC 5, where the source language and the target language is now different, becoming the most challenging task so far [1]. Among numerous methodologies, recurrent neural network (RNN) and its variants are now common in handling sequential data with their promise of performance and ease of use. Long-short term memory (LSTM) [2], one of the most popular RNN variants that is able to capture long-range dependencies, is commonly adopted. Their performance has shown great potential in previous challenges (DSTC 2,3&4) ([3], [4], [5]). Here, we adopted the bi-directional LSTM, *:These authors contributed equally. This material is based upon work supported by the Ministry of Trade, Industry & Energy (MOTIE, Korea) under Industrial Technology Innovation Program. No.10063424, Development of Distant Speech Recognition and Multi-Task Dialog Processing Technologies for In-Door Conversational Robots. Topic Utterance Label ATTRACTION Uh what about East Coast Beach? East Coast Beach is also a nice place. East Coast? Yah. Uh East Coast. INFO: Preference PLACE: East Coast Park TYPE OF PLACE: Beach Table 1. Example of training set segment in DSTC 5. Given Topic and Utterance, we should predict the slot-value pairs in Label. INFO, PLACE and TYPE OF PLACE are the examples of slots, and Preference, East Coast Park and Beach are corresponding values of such slots. which has been found to outperform other models in sequential data processing that requires capturing local context ([6], [7]). The main difficulties in using such models are, however, the large size of the ontology and sparsity of the training data. Since only 56% of keywords in the ontology are present in the training dialogs, it is not possible to completely train typical models and the need of special treatment arises (e.g. value independent network in [3]). In this paper, we present a tracker based on bi-directional LSTM with attention mechanism, which is the recent trend in deep learning. After eliminating obvious stop words and reducing lexical variations, pre-trained word embedding model converts linguistic words into vectors. Bi-directional LSTM works on word embedding vectors to find notable keywords in each utterance. Then the tracker finds proper values for each slot from weighted sum of word vectors with attention weight which is determined from the bi-directional LSTM. Finally, the tracker decides whether the found values are reliable based on entropy or cosine similarity. This paper is organized as follows. In Section 2, we give a description of the DSTC 5 dataset and the main task, with the brief review of previous works of LSTM and attention mechanism. In Section 3, we explain our dialog state tracker 978-1-5090-4903-5/16/$31.00 2016 IEEE 531 GlobalSIP 2016

maintained neglected POS Description POS Description VB base form verb: ask, assume, build NN singular noun: child, dog, car VBD past tense verb: adopted, used, asked NNS plural noun: children, dogs, cars VBG present participle verb: using, focusing NNP proper noun: Singapore, Merlion VBZ 3rd person singular verb: marks, bases CD numeral, cardinal: nine-thirty, 1987, one TO to as preposition or infinitive marker IN preposition: among, upon, on, by RB adverb: occasionally, very, only CC conjunction: and, or, but, either DT, EX, FW, LS, MD, PDT, PRP, RBR, RBS, SYM, UH, WDT, WP, WRB Table 2. List of POS tags to be maintained or neglected. We maintain nouns, verbs, numerals and prepositions and neglect the others. 2 Given utterance It will be on the trail as you are walking and so you can see their relatives here %uh in the Night Safari POS tagging POS tagged utterance It will (MD) be (VB) on (IN) the (DT) trail (NN) as (IN) you are (VBP) walking (VBG) and (CC) so (RB) you can (MD) see (VB) their relatives (NNS) here (RB) in (IN) the (DT) night (NN) (NN) Maintained words Lexicalized words trail trail walking walk see see relatives relative in in night night Stemming Lemmatizing Preprocessed utterance trail walk see relative in night Fig. 1. The preprocessing example in our tracker. part-by-part, and in Section 4, we discuss the detailed results of DSTC 5 main task. 2.1. Task description 2. BACKGROUND In DSTC 5, the dialog state trackers compete each other over TourSG dataset. TourSG dataset consists of dialog sessions on touristic information for Singapore collected from Skype calls between a tour guide and a tourist. Dialog states are defined for each sub-dialog level, denoted as the segment. A full dialog session is divided into segments considering their topical coherence. Each segment has one topic and dialog state, defined by the collection of slot-value pairs. The main task is to fill the slot-value pairs in each segment where the corresponding topics and utterances are given (see Table 1). Possible slot-value pairs are provided in the form of ontology. Although there exist some cases that more than one value are assigned in single slot, we only predicted one value per slot for simplicity. 2 see http://www.nltk.org/book/ch05.html for detailed description of POS Slots are categorized into two types: regular slots and INFO slot. While regular slots are filled if some specific values have directly discussed in the segment, INFO slot takes the place if such specific values are not discussed, and is filled with corresponding subject. DSTC 5 differs from previous challenges in that it is a Chinese-English cross-language task. The goal of the task is to construct the dialog state tracker on Chinese dialog while the training set is given in English. The dataset consists of 35 English training set, 2 Chinese validation set and 10 Chinese test set. The top 5 results of machine translations of all dialogs are provided. In case of Chinese dialogs in validation and test set, we use the topmost machine translation result. 2.2. RNN and LSTM RNN and LSTM have been generally applied to natural language processing (NLP) problems. The basic idea of RNN is to make use of sequential information. The fundamental neural network does not consider the dependency of all inputs and outputs. However, in various tasks, there exist dependency in inputs and outputs, such as sentence analysis. RNN re- 532

100 50 Tanjong Pagar MRT Station Pasir Panjang MRT Station Hotel 81 - Hollywood Fr gr nce H 1el - Oasis Budget hotel H 1el C rl1 n H 1el New Cape Inn 0 Orchard Road Victoria Street West Mall Tampines Mall Night market Hougang Pasir Ris Fernvale Bukit Merah Universal Studios Singapore National Orchid Garden Amusement ride Farm Movie Check-out B king 650 6100 INFO CUISINE TYPE_OF_PLACE DRINK PLACE DISH NEIGHBOURHOOD ACTIVITY STATION C r 20el Rumah Makan Minang B ng B2rger0 Radin Mas Kebab Pork Duck rice F20i n Th i c2i0ine Teh 1 rik Nasi lemak N r1h Indi n c2i0ine 80 60 640 620 0 20 40 60 Fig. 2. Two-dimensional t-sne [8] embeddings of key phrases of ontology learned by Word2Vec. The figure shows clear clusters of words of similar concepts. The phrases shown in the figure are chosen randomly. currently use the previous computation result to compute the current output. LSTM is RNN with gates, which is proposed to prevent the vanishing gradient problem, becoming more effective in dealing with long sequences. The basic structure of LSTM unit consists of a cell state with three essential gates: input gate, forget gate and output gate. The cell controls the information storing for a long period via gates. Given an input vector x t at time step t, the formal equation for updating gates, output and cell state are defined as follows: i t = σ ( x t U i + h t 1 W i) f t = σ ( x t U f + h t 1 W f ) o t = σ (x t U o + h t 1 W o ) c t = c t 1 f t + i t tanh (x t U c + h t 1 W c ) h t = tanh (c t ) o t where W i, W f, W o, W c R N N,U i, U f, U o, U c R N N are weight matrices, h t is output vector and i, f and o represent input (i), forget (f) and output (o) gates. 2.3. Related works It is now very popular to use neural networks in NLP tasks. In the last challenge (DSTC 4) that had very similar dataset to this challenge, two teams have proposed neural network based model. [9] reduced the task into multi-domain text classification problem by focusing on INFO slot filling. Using convolutional neural networks (CNN), they combined topic-shared and topic-specific structures. [5] suggested the tracker, which integrates with the baseline tracker and the unidirectional LSTM to convey the information of previous utterances. The aforementioned trackers, however, predict by outputting one-hot-encoding on the ontology, which is not easily trainable for the data with much larger ontology than the training data. More recently, Neural Belief Tracker (NBT, [10]) used deep neural network and convolutional neural network with pre-trained word embedding model. To overcome the same problem, we take a slightly different approach the attention mechanism. Attention mechanism is a method to focus on the meaningful information in the utterance. In dialog example, not all words in each sentence are related with state of dialog. Attention weight represents the contribution of words towards slot-specific information and it is computed using a softmax. Attention mechanism has been broadly applied to other NLP tasks such as sentence summarization [11], recognizing textual entailment(rte) [12], and machine translation [13]. CopyNet [14] is an interesting variant of attention-based RNN Encoder-Decoder model adopting the copying mecha- 533

Given utterance Preprocessed utterance Word vector It will be on the trail as you are walking and so you can see their relatives here %uh in the Night Safari. Preprocessing Ontology hint vector trail walk see relative in night Word2vec Bidirectional LSTM Attention weight Attention mechanism Output vector Label (cosine similarity) + Value vector (Regular slots) PLACE: Night Safari (0.87676161065) + Value vector (INFO slot) INFO: Exhibit (0.99221589149) Fig. 3. The overall architecture of our tracker. We consider regular slots and INFO slot separately. Our tracker is inputted with the utterance, and outputs the value vectors (see Section 3 for more details). nism to deal with large ontology. CopyNet has the ability to handle out-of-vocabulary words as proper nouns by copying consecutive sequence. For our task, target values are not long enough to be predicted sequentially. We overcome proper noun issue by improving embedding model. 3. TRACKER ARCHITECTURE 3.1. Preprocessing utterances 3.1.1. POS tagging, Stemming and Lemmatization POS tagging is the process of classifying words and labeling them with lexical categories. TourSG corpus contains insignificant words such as articles (a, an, the), pronouns (they, we, he, she), auxiliary verbs (can, will, could, would) and onomatopoeia (uh, umm, ha). We identify and eliminate the word that does not affect the meaning of utterance by POS tagging so that the clarified utterance is easier to be understood by LSTM. We maintain nouns, verbs, adjectives, propositions and numerals, and rule the others out with the pre-trained POS tagging module in natural language toolkit (NLTK) (see Table 2). Since the variation of words such as tense and plurality adds extra complication, there is also a need of converting these as a lexicalized form for the efficient embedding of words. We standardize the words by using porter-stemmer and word-net-lemmatizer in NLTK. 3.1.2. Word embedding Now each utterance contains only meaningful lemmatized words. We project those words into a high-dimensional space maintaining relationship between them using Word2Vec. We convert each word into 100 dimensional embedding vector. More than 13 million sentences crawled from TripAdvisor 3 are used as training set with given TourSG corpus. 3.1.3. Ontology hint vector Ontology also contains a valuable information about the topic to which a word is related. We construct the ontology hint vector that denotes whether a word is in the ontology or not for each topic to convey such information, resulting 30 dimensional one-hot vector. We concatenate an ontology hint vector to a word embedding vector and use it as an input for value prediction network. 3.2. Predicting values using attention mechanism Our value prediction model consists of a bi-directional LSTM and a network of attention mechanism. Regard an utterance u that consists of N words (u 1, u 2,, u N ). Each word has word embedding vector (w 1, w 2,, w N ) and ontology hint vector (c 1, c 2,, c N ). The input of the LSTM is a concatenated vector of a word embedding vector and an ontology hint 3 https://www.tripadvisor.com 534

5.5 5.5 5.0 5.0 4.5 4.0 3.5 3.0 2.5 2.0 4.5 4.0 3.5 3.0 2.5 1.5 2.0 1.0 (202) (387) (1568) 1.5 (226) (756) (1175) (a) Entropy of NEIGHBOURHOOD slot (b) Entropy of PLACE slot 5.5 1.0 5.0 0.8 0.6 4.5 0.4 4.0 0.2 3.5 (573) (860) (724) 0.0 (573) (882) (734) (c) Entropy of INFO slot (d) Cosine similarity of INFO slot Fig. 4. The entropy of attention weights and the cosine similarity between prediction and closest word in ontology. denotes that current utterance has a relevant slot and output value is correct. denotes that output value is not correct. denotes that current utterance does not have a relevant slot. The numbers in parenthesis are the counts of sample cases. vector of each word: [w 1 c 1 ; w 2 c 2 ; ; w N c N ] (1) where denotes vector concatenation. Each word has the output of the bi-directional LSTM: H = [h 1 ; h 2 ; ; h N ] (2) The cell values are passed to the time-distributed dense networks for each slot to formulate a single scalar per word that works as importance of corresponding word. These scalars are normalized with softmax function over words in the utterance, and further denoted as attention weights: a s = σ(w s H) (3) where a s R N is attention weight vector, W s is weight matrix of the dense network, σ is a softmax function, and superscript s denotes a specific slot. As an attention mechanism, we calculated the output value vector by a weighted sum of word vectors with the weights from dense network of each slot: v s = N a s i w i (4) i=1 where a s i is scalar value at index i of as. While attention weights are calculated for all slots, the model obtains values for all slots. This vector is now can be said to contain an essential information of the utterance. In contrast to conventional attention mechanism that utilizes a weighted sum of hidden cell vectors of LSTM, the model proposed here outputs a weighted sum of input words to ensure clear focus on key phrases. The tracker chooses the word among the available slot values listed in ontology with the closest embedding by cosine similarity: S = a b a b. 3.3. Excluding unreliable predictions It is equally important to decide what unreliable predictions to exclude as it is to predict correct values in dialog state tracking challenge since to predict nothing is also a choice. 535

Schedule 1 Schedule 2 Team Accuracy Precision Recall F-measure Accuracy Precision Recall F-measure Baseline 1 0.0250 0.1148 0.1102 0.1124 0.0321 0.1425 0.1500 0.1462 Baseline 2 0.0161 0.1743 0.1279 0.1475 0.0222 0.1979 0.1774 0.1871 Team 1 0.0417 0.3650 0.2795 0.3166 0.0612 0.3811 0.3548 0.3675 Team 2 0.0788 0.5195 0.3315 0.4047 0.0956 0.5643 0.3769 0.4519 Team 3 0.0351 0.3216 0.1515 0.2060 0.0505 0.3350 0.2045 0.2539 Team 4 0.0583 0.4008 0.2776 0.3280 0.0765 0.4127 0.3284 0.3658 Team 5 0.0330 0.3377 0.2318 0.2749 0.0520 0.3637 0.3044 0.3314 Ours 0.0491 0.4684 0.2193 0.2988 0.0643 0.4758 0.2623 0.3381 Team 7 0.0092 0.4287 0.0431 0.0783 0.0107 0.4000 0.0441 0.0794 Team 8 0.0192 0.3130 0.1048 0.1570 0.0214 0.3021 0.1046 0.1554 Team 9 0.0231 0.1139 0.1090 0.1114 0.0314 0.1412 0.1487 0.1449 Table 3. The DSTC 5 main task results of the best trackers from each team, chosen based on schedule 2 accuracy. 4 Since uncertainty estimates are not available in neural networks, other criteria had to be chosen. 3.3.1. Entropy of attention weights For the regular slots, some keywords or phrases usually lead to the correct answers, and it is more likely to have concentrated attention weights on specific part of the utterances. The entropy of attention weights H = N i=1 as i log as i is an attractive choice in this sense: predictions with concentrated weights result in low entropy so that predictions with high entropy can be seen as uncertain, therefore discarded. 3.3.2. Cosine similarity On the other hand, INFO slot predicts overall topic of the utterance and the whole parts may worth attention, making entropy an inappropriate choice. We thus threshold on the cosine similarity measure between the prediction and the closest word in ontology instead of entropy, so that the predictions that are not close enough to any subject are discarded. Fig. 4 exhibits how the correct/wrong/absent values differ by the criteria above. As can be seen in (a) and (b), the entropy of correct values in regular slots are clearly distinguishable from the entropy of wrong or absent values. While (c) shows that the INFO slot can not be treated in the same way, (d) shows that the cosine similarity can be used instead for INFO slot. We also tried to construct a neural network that gives the decision, namely the slot activation network, but it turned out to be perform poorly than two criteria above in practice. It will be dealt in further research. 4. RESULTS AND DISCUSSION Table 3 summarizes the DSTC 5 final result. There are two kinds of evaluation method: Schedule 1 scores prediction at every utterance (utterance level evaluation) and Schedule 2 only scores prediction at the last utterance of every segment (segment level evaluation). Baseline tracker is based on fuzzy matching algorithm. Among 9 teams participated the main task, We are team 6, and denoted as ours in Table 3. Our best entry accuracy scored 0.0491 in schedule 1 and 0.0643 in schedule 2, taking third place among all trackers submitted. Although we are behind the performance of the tracker of team 2&4, we place more emphasis on the fact that our tracker is fairly simple and its performance is not heavily affected by the size of ontology unlike neural trackers in previous challenges. Sophisticated pre- and post-processing are not included as well, implying that the tracker proposed can be directly applied to other domains. There still remains some room for improvement in our model; in training and predicting with our tracker, we treated the segments as independent instance disregarding the context among the segments from the same dialog. Since the information of dialog states appears a prior to current segment quite often, such context is crucial in determining dialog states. Adopting hierarchical structures such as HRED introduced in [15] would have helped for further improvements and remained as future work. 5. CONCLUSION This paper presented our proposed tracker for DSTC 5. The proposed tracker is bi-directional LSTM with noble attention mechanism that can capture key phrases from utterances, enabling good performance in challenging task where the ontology is much larger than training vocabulary without sophisticated pre- and post-processing. Our concise model proved its efficiency by taking third place in the challenge. The code can be downloaded from our repository. 5 4 see https://github.com/seokhwankim/dstc5/tree/master/results for detailed results 5 https://github.com/jys5609/dstc5 536

6. REFERENCES [1] Seokhwan Kim, Luis Fernando D Haro, Rafael E. Banchs, Jason Williams, Matthew Henderson, and Koichiro Yoshino, The Fifth Dialog State Tracking Challenge, in Proceedings of the 2016 IEEE Workshop on Spoken Language Technology (SLT), 2016. [2] Sepp Hochreiter and Jürgen Schmidhuber, Long shortterm memory, Neural computation, vol. 9, no. 8, pp. 1735 1780, 1997. [3] Matthew Henderson, Blaise Thomson, and Steve Young, Word-based dialog state tracking with recurrent neural networks, in Proceedings of the 15th Annual Meeting of the Special Interest Group on Discourse and Dialogue (SIGDIAL), 2014, pp. 292 299. [4] Matthew Henderson, Blaise Thomson, and Steve Young, Robust dialog state tracking using delexicalised recurrent neural networks and unsupervised adaptation, in Spoken Language Technology Workshop (SLT), 2014 IEEE. IEEE, 2014, pp. 360 365. [5] Koichiro Yoshino, Takuya Hiraoka, Graham Neubig, and Satoshi Nakamura, Dialog state tracking using long short-term memory neural networks, in Proceedings of the 7th International Workshop on Spoken Dialogue Systems (IWSDS), 2016. [6] Alex Graves and Jrgen Schmidhuber, Framewise phoneme classification with bidirectional lstm and other neural network architectures, Neural Networks, pp. 5 6, 2005. [7] Hasim Sak, Andrew W. Senior, and Françoise Beaufays, Long short-term memory based recurrent neural network architectures for large vocabulary speech recognition, CoRR, vol. abs/1402.1128, 2014. [8] Laurens van der Maaten and Geoffrey E. Hinton, Visualizing high-dimensional data using t-sne, Journal of Machine Learning Research, vol. 9, pp. 2579 2605, 2008. [9] Hongjie Shi, Takashi Ushio, Mitsuru Endo, Katsuyoshi Yamagami, and Noriaki Horii, Convolutional neural networks for multi-topic dialog state tracking, in Proceedings of the 7th International Workshop on Spoken Dialogue Systems (IWSDS), 2016. [10] Nikola Mrksic, Diarmuid Ó Séaghdha, Tsung-Hsien Wen, Blaise Thomson, and Steve J. Young, Neural belief tracker: Data-driven dialogue state tracking, CoRR, vol. abs/1606.03777, 2016. [11] Alexander M. Rush, Sumit Chopra, and Jason Weston, A neural attention model for abstractive sentence summarization, CoRR, vol. abs/1509.00685, 2015. [12] Tim Rocktäschel, Edward Grefenstette, Karl Moritz Hermann, Tomás Kociský, and Phil Blunsom, Reasoning about entailment with neural attention, CoRR, vol. abs/1509.06664, 2015. [13] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio, Neural machine translation by jointly learning to align and translate, CoRR, vol. abs/1409.0473, 2014. [14] Jiatao Gu, Zhengdong Lu, Hang Li, and Victor O. K. Li, Incorporating copying mechanism in sequence-tosequence learning, CoRR, vol. abs/1603.06393, 2016. [15] Alessandro Sordoni, Yoshua Bengio, Hossein Vahabi, Christina Lioma, Jakob Grue Simonsen, and Jian-Yun Nie, A hierarchical recurrent encoder-decoder for generative context-aware query suggestion, in Proceedings of the 24th ACM International on Conference on Information and Knowledge Management. ACM, 2015, pp. 553 562. 537