During dialogue policy learning, the world model is constantly updated with real user experience to approach real user behavior, and in turn, the dialogue agent is optimized using both real experience and simulated experience.
This paper presents a Discriminative Deep Dyna-Q (D3Q) approach to improving the effectiveness and robustness of Deep Dyna-Q (DDQ), a recently proposed framework that extends the Dyna-Q algorithm to integrate planning for task-completion dialogue policy learning.
Training task-completion dialogue agents with reinforcement learning usually requires a large number of real user experiences.
ACTIVE LEARNING Q-LEARNING TASK-COMPLETION DIALOGUE POLICY LEARNING