Dialogue Learning with Human Teaching and Feedback in End-to-End Trainable Task-Oriented Dialogue Systems

In this work, we present a hybrid learning method for training task-oriented dialogue systems through online user interactions. Popular methods for learning task-oriented dialogues include applying reinforcement learning with user feedback on supervised pre-training models. Efficiency of such learning method may suffer from the mismatch of dialogue state distribution between offline training and online interactive learning stages. To address this challenge, we propose a hybrid imitation and reinforcement learning method, with which a dialogue agent can effectively learn from its interaction with users by learning from human teaching and feedback. We design a neural network based task-oriented dialogue agent that can be optimized end-to-end with the proposed learning method. Experimental results show that our end-to-end dialogue agent can learn effectively from the mistake it makes via imitation learning from user teaching. Applying reinforcement learning with user feedback after the imitation learning stage further improves the agent's capability in successfully completing a task.

PDF Abstract NAACL 2018 PDF NAACL 2018 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Dialogue State Tracking Second dialogue state tracking challenge Liu et al. Request - # 4
Area 90 # 2
Food 84 # 2
Price 92 # 2
Joint 72 # 6

Methods


No methods listed for this paper. Add relevant methods here