The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike License.
825 PAPERS • 3 BENCHMARKS
Dataset Details Total Labeled: 100%
1 PAPER • NO BENCHMARKS YET