top | item 45698346 (no title) Mithriil | 4 months ago My opinion on the "Attention is all you need" paper is that its most important idea is the Positional Encoding. The transformer head itself... is just another NN block among many. discuss order hn newest No comments yet.
No comments yet.