Command Palette
Search for a command to run...
Victor Zhong; Caiming Xiong; Richard Socher

Abstract
Dialogue state tracking, which estimates user goals and requests given the dialogue context, is an essential part of task-oriented dialogue systems. In this paper, we propose the Global-Locally Self-Attentive Dialogue State Tracker (GLAD), which learns representations of the user utterance and previous system actions with global-local modules. Our model uses global modules to share parameters between estimators for different types (called slots) of dialogue states, and uses local modules to learn slot-specific features. We show that this significantly improves tracking of rare states and achieves state-of-the-art performance on the WoZ and DSTC2 state tracking tasks. GLAD obtains 88.1% joint goal accuracy and 97.1% request accuracy on WoZ, outperforming prior work by 3.7% and 5.5%. On DSTC2, our model obtains 74.5% joint goal accuracy and 97.5% request accuracy, outperforming prior work by 1.1% and 1.0%.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| dialogue-state-tracking-on-second-dialogue | Zhong et al. | Area: - Food: - Joint: 74.5 Price: - Request: 97.5 |
| dialogue-state-tracking-on-wizard-of-oz | Zhong et al. | Joint: 88.1 Request: 97.1 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.