Self-Attention with Relative Position Representations

Peter ShawJakob UszkoreitAshish Vaswani

   Papers with code   Abstract  PDF

Relying entirely on an attention mechanism, the Transformer introduced by Vaswani et al. (2017) achieves state-of-the-art results for machine translation... (read more)

Benchmarked Models

No benchmarked models yet. Click here to submit a model.