Some tasks are inferred based on the benchmarks list.
The benchmarks section lists all benchmarks using a given dataset or any of
its variants. We use variants to distinguish between results evaluated on
slightly different versions of the same dataset. For example, ImageNet 32⨉32
and ImageNet 64⨉64 are variants of the ImageNet dataset.
Semantic Textual Similarity (2012 - 2016) involves a set of semantic textual similarity datasets that were part of previous shared tasks (2012-2016):
STS12 - Semeval-2012 task 6: A pilot on semantic textual similarity
STS13 - SEM 2013 shared task: Semantic Textual Similarity
STS14 - SemEval-2014 task 10: Multilingual semantic textual similarity
STS15 - SemEval-2015 task 2: Semantic textual similarity, English, Spanish and pilot on interpretability
STS16 - SemEval-2016 task 1: Semantic textual similarity, monolingual and cross-lingual evaluation