The Universal Transformer is a generalization of the Transformer architecture. Universal Transformers combine the parallelizability and global receptive field of feed-forward sequence models like the Transformer with the recurrent inductive bias of RNNs. They also utilise a dynamic per-position halting mechanism.
Source: Universal TransformersPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Sentence | 4 | 10.26% |
Language Modelling | 3 | 7.69% |
Semantic Communication | 3 | 7.69% |
Text Generation | 2 | 5.13% |
Instance Segmentation | 1 | 2.56% |
Semantic Segmentation | 1 | 2.56% |
Reinforcement Learning (RL) | 1 | 2.56% |
Semantic Similarity | 1 | 2.56% |
Semantic Textual Similarity | 1 | 2.56% |