About

Benchmarks

You can find evaluation results in the subtasks. You can also submitting evaluation metrics for this task.

Subtasks

Datasets

Greatest papers with code

A Repository of Conversational Datasets

WS 2019 PolyAI-LDN/conversational-datasets

Progress in Machine Learning is often driven by the availability of large datasets, and consistent evaluation metrics for comparing modeling approaches.

CONVERSATIONAL RESPONSE SELECTION DIALOGUE UNDERSTANDING

DREAM: A Challenge Dataset and Models for Dialogue-Based Reading Comprehension

1 Feb 2019nlpdata/dream

DREAM is likely to present significant challenges for existing reading comprehension systems: 84% of answers are non-extractive, 85% of questions require reasoning beyond a single sentence, and 34% of questions also involve commonsense knowledge.

DIALOGUE UNDERSTANDING READING COMPREHENSION

A Natural Language Corpus of Common Grounding under Continuous and Partially-Observable Context

8 Jul 2019Alab-NII/onecommon

Finally, we evaluate and analyze baseline neural models on a simple subtask that requires recognition of the created common ground.

DIALOGUE UNDERSTANDING GOAL-ORIENTED DIALOG LANGUAGE ACQUISITION

Masking Orchestration: Multi-task Pretraining for Multi-role Dialogue Representation Learning

27 Feb 2020wangtianyiftd/dialogue_pretrain

Multi-role dialogue understanding comprises a wide range of diverse tasks such as question answering, act classification, dialogue summarization etc.

DIALOGUE UNDERSTANDING QUESTION ANSWERING REPRESENTATION LEARNING