ELECTRA is a transformer with a new pre-training approach which trains two transformer models: the generator and the discriminator. The generator replaces tokens in the sequence - trained as a masked language model - and the discriminator (the ELECTRA contribution) attempts to identify which tokens are replaced by the generator in the sequence. This pre-training task is called replaced token detection, and is a replacement for masking the input.
Source: ELECTRA: Pre-training Text Encoders as Discriminators Rather Than GeneratorsPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Language Modelling | 31 | 14.35% |
Sentence | 16 | 7.41% |
Question Answering | 16 | 7.41% |
Reading Comprehension | 10 | 4.63% |
Named Entity Recognition (NER) | 7 | 3.24% |
Natural Language Inference | 7 | 3.24% |
NER | 6 | 2.78% |
Natural Language Understanding | 6 | 2.78% |
Text Classification | 5 | 2.31% |