top | item 45698346

(no title)

Mithriil | 4 months ago

My opinion on the "Attention is all you need" paper is that its most important idea is the Positional Encoding. The transformer head itself... is just another NN block among many.

discuss

order

No comments yet.