2 Matching Annotations
- Oct 2023
-
typeshare.co typeshare.co
-
- Nov 2022
-
arxiv.org arxiv.org
-
we propose the Transformer, a model architecture eschewing recurrence and insteadrelying entirely on an attention mechanism to draw global dependencies between input and output.The Transformer allows for significantly more parallelization a
Using the attention mechanism to determine global dependencies between input and output instead of using recurrent links to past states. This is the essence of their new idea.
Tags
Annotators
URL
-