Multitask Semi-Supervised Learning for Class-Imbalanced Discourse Classification

As labeling schemas evolve over time, small differences can render datasets following older schemas unusable. This prevents researchers from building on top of previous annotation work and results in the existence, in discourse learning in particular, of many small class-imbalanced datasets. In this work, we show that a multitask learning approach can combine discourse datasets from similar and diverse domains to improve discourse classification. We show an improvement of 4.9% Micro F1-score over current state-of-the-art benchmarks on the NewsDiscourse dataset, one of the largest discourse datasets recently published, due in part to label correlations across tasks, which improve performance for underrepresented classes. We also offer an extensive review of additional techniques proposed to address resource-poor problems in NLP, and show that none of these approaches can improve classification accuracy in our setting.

PDF Abstract

Datasets


  Add Datasets introduced or used in this paper

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Text Classification NewsDiscourse Human (Post-Rec.) (Spangher et al., 2021) macro F1 73.69 # 1
Text Classification NewsDiscourse Human (Blind) (Spangher et al., 2021) macro F1 46.18 # 7
Text Classification NewsDiscourse MT-Mic (Spangher et al., 2021) macro F1 61.89 # 3
Text Classification NewsDiscourse MT-Mac (Spangher et al., 2021) macro F1 63.46 # 2

Methods


No methods listed for this paper. Add relevant methods here