SCROLLS: Standardized CompaRison Over Long Language Sequences

NLP benchmarks have largely focused on short texts, such as sentences and paragraphs, even though long texts comprise a considerable amount of natural language in the wild. We introduce SCROLLS, a suite of tasks that require reasoning over long texts. We examine existing long-text datasets, and handpick ones where the text is naturally long, while prioritizing tasks that involve synthesizing information across the input. SCROLLS contains summarization, question answering, and natural language inference tasks, covering multiple domains, including literature, science, business, and entertainment. Initial baselines, including Longformer Encoder-Decoder, indicate that there is ample room for improvement on SCROLLS. We make all datasets available in a unified text-to-text format and host a live leaderboard to facilitate research on model architecture and pretraining methods.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Long-range modeling SCROLLS LED Base GovRep 56.2 / 26.6 / 28.8 # 7
SumScr 24.2 / 4.5 / 15.4 # 10
QMSum 25.1 / 6.7 / 18.8 # 10
Qspr 26.6 # 8
Nrtv 18.5 # 8
QALT EM-T/H 25.8 / 25.4 # 8
CNLI 71.5 # 9
Avg. 29.16 # 8
Long-range modeling SCROLLS BART Base GovRep 47.9 / 18.6 / 22.7 # 9
SumScr 27.2 / 4.9 / 16.7 # 9
QMSum 30.2 / 8.7 / 20.7 # 9
Qspr 26.3 # 9
Nrtv 15.4 # 9
QALT EM-T/H 26.0 / 25.9 # 7
CNLI 77.4 # 8
Avg. 29.01 # 9
Long-range modeling SCROLLS Naive GovRep 45.3 / 17.9 / 20.8 # 10
SumScr 19.6 / 1.8 / 11.0 # 11
QMSum 14.2 / 2.0 / 9.3 # 11
Qspr 3.4 # 10
Nrtv 1.5 # 10
QALT EM-T/H 25.2 / 26.1 # 9
CNLI 66 # 10
Avg. 19.35 # 10

Methods