Linear Warmup With Linear Decay is a learning rate schedule in which we increase the learning rate linearly for $n$ updates and then linearly decay afterwards.
Paper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Language Modelling | 96 | 11.96% |
Sentiment Analysis | 42 | 5.23% |
Text Classification | 39 | 4.86% |
Retrieval | 34 | 4.23% |
Question Answering | 31 | 3.86% |
Classification | 25 | 3.11% |
Large Language Model | 21 | 2.62% |
NER | 21 | 2.62% |
Named Entity Recognition (NER) | 15 | 1.87% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |