Search Results for author: Kaixiang Mo

Found 6 papers, 0 papers with code

Transferring SLU Models in Novel Domains

no code implementations ICLR 2019 Yaohua Tang, Kaixiang Mo, Qian Xu, Chao Zhang, Qiang Yang

When building models for novel natural language domains, a major challenge is the lack of data in the new domains, no matter whether the data is annotated or not.

Intent Recognition Meta-Learning +4

Cross-domain Dialogue Policy Transfer via Simultaneous Speech-act and Slot Alignment

no code implementations20 Apr 2018 Kaixiang Mo, Yu Zhang, Qiang Yang, Pascale Fung

Also, they depend on either common slots or slot entropy, which are not available when the source and target slots are totally disjoint and no database is available to calculate the slot entropy.

Fine Grained Knowledge Transfer for Personalized Task-oriented Dialogue Systems

no code implementations11 Nov 2017 Kaixiang Mo, Yu Zhang, Qiang Yang, Pascale Fung

Training a personalized dialogue system requires a lot of data, and the data collected for a single user is usually insufficient.

Sentence Task-Oriented Dialogue Systems +1

Integrating User and Agent Models: A Deep Task-Oriented Dialogue System

no code implementations10 Nov 2017 Weiyan Wang, Yuxiang Wu, Yu Zhang, Zhongqi Lu, Kaixiang Mo, Qiang Yang

Then the built user model is used as a leverage to train the agent model by deep reinforcement learning.

Task-Oriented Dialogue Systems

Flexible End-to-End Dialogue System for Knowledge Grounded Conversation

no code implementations13 Sep 2017 Wenya Zhu, Kaixiang Mo, Yu Zhang, Zhangbin Zhu, Xuezheng Peng, Qiang Yang

Although existing generative question answering (QA) systems can be applied to knowledge grounded conversation, they either have at most one entity in a response or cannot deal with out-of-vocabulary entities.

Generative Question Answering

Personalizing a Dialogue System with Transfer Reinforcement Learning

no code implementations10 Oct 2016 Kaixiang Mo, Shuangyin Li, Yu Zhang, Jiajun Li, Qiang Yang

One way to solve this problem is to consider a collection of multiple users' data as a source domain and an individual user's data as a target domain, and to perform a transfer learning from the source to the target domain.

reinforcement-learning Reinforcement Learning (RL) +1

Cannot find the paper you are looking for? You can Submit a new open access paper.