Paper

Transformer++

Recent advancements in attention mechanisms have replaced recurrent neural networks and its variants for machine translation tasks. Transformer using attention mechanism solely achieved state-of-the-art results in sequence modeling... (read more)

Results in Papers With Code
(↓ scroll down to see all results)