GLUE (General Language Understanding Evaluation benchmark)

Introduced by Wang et al. in GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding

General Language Understanding Evaluation (GLUE) benchmark is a collection of nine natural language understanding tasks, including single-sentence tasks CoLA and SST-2, similarity and paraphrasing tasks MRPC, STS-B and QQP, and natural language inference tasks MNLI, QNLI, RTE and WNLI.

Source: Align, Mask and Select: A Simple Method for Incorporating Commonsense Knowledge into Language Representation Models

Papers


Paper Code Results Date Stars

Tasks


Similar Datasets


License


Modalities


Languages