DABS: A Domain-Agnostic Benchmark for Self-Supervised Learning

23 Nov 2021  ·  Alex Tamkin, Vincent Liu, Rongfei Lu, Daniel Fein, Colin Schultz, Noah Goodman ·

Self-supervised learning algorithms, including BERT and SimCLR, have enabled significant strides in fields like natural language processing, computer vision, and speech processing. However, these algorithms are domain-specific, meaning that new self-supervised learning algorithms must be developed for each new setting, including myriad healthcare, scientific, and multimodal domains. To catalyze progress toward domain-agnostic methods, we introduce DABS: a Domain-Agnostic Benchmark for Self-supervised learning. To perform well on DABS, an algorithm is evaluated on seven diverse domains: natural images, multichannel sensor data, English text, speech recordings, multilingual text, chest x-rays, and images with text descriptions. Each domain contains an unlabeled dataset for pretraining; the model is then is scored based on its downstream performance on a set of labeled tasks in the domain. We also present e-Mix and ShED: two baseline domain-agnostic algorithms; their relatively modest performance demonstrates that significant progress is needed before self-supervised learning is an out-of-the-box solution for arbitrary domains. Code for benchmark datasets and baseline algorithms is available at https://github.com/alextamkin/dabs.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Self-Supervised Learning DABS Pretraining: None Natural Images 10.1 # 3
Text 42.3 # 3
Speech 24.9 # 3
Sensors 69.8 # 3
Med. Imaging 68.1 # 3
Images & Text 57.5 # 1
Self-Supervised Learning DABS Pretraining: ShED Natural Images 20.9 # 2
Text 48.4 # 1
Speech 36.5 # 2
Sensors 88.7 # 1
Med. Imaging 74.5 # 1
Images & Text 54.3 # 2
Self-Supervised Learning DABS Pretraining: e-Mix Natural Images 27.9 # 1
Text 44.1 # 2
Speech 41.8 # 1
Sensors 79.5 # 2
Med. Imaging 72.4 # 2
Images & Text 48.9 # 3

Methods