Fine-grained Post-training for Improving Retrieval-based Dialogue Systems

Retrieval-based dialogue systems display an outstanding performance when pre-trained language models are used, which includes bidirectional encoder representations from transformers (BERT). During the multi-turn response selection, BERT focuses on training the relationship between the context with multiple utterances and the response. However, this method of training is insufficient when considering the relations between each utterance in the context. This leads to a problem of not completely understanding the context flow that is required to select a response. To address this issue, we propose a new fine-grained post-training method that reflects the characteristics of the multi-turn dialogue. Specifically, the model learns the utterance level interactions by training every short context-response pair in a dialogue session. Furthermore, by using a new training objective, the utterance relevance classification, the model understands the semantic relevance and coherence between the dialogue utterances. Experimental results show that our model achieves new state-of-the-art with significant margins on three benchmark datasets. This suggests that the fine-grained post-training method is highly effective for the response selection task.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Conversational Response Selection Douban BERT-FP MAP 0.644 # 1
MRR 0.680 # 2
P@1 0.512 # 2
R10@1 0.324 # 2
R10@2 0.542 # 1
R10@5 0.870 # 1
Conversational Response Selection E-commerce BRET-FP R10@1 0.870 # 1
R10@2 0.956 # 1
R10@5 0.993 # 1
Conversational Response Selection RRS BERT-FP R10@1 0.488 # 2
MAP 0.702 # 1
MRR 0.712 # 2
P@1 0.543 # 2
R10@2 0.708 # 1
R10@5 0.927 # 2
Conversational Response Selection RRS Ranking Test BERT-FP NDCG@3 0.609 # 4
NDCG@5 0.709 # 4
Conversational Response Selection Ubuntu Dialogue (v1, Ranking) BERT-FP R10@1 0.911 # 1
R10@2 0.962 # 1
R10@5 0.994 # 1

Methods