top | item 19356100

(no title)

rasmi | 7 years ago

Here are two more great articles about Transformers:

The Illustrated Transformer (referenced in the parent): http://jalammar.github.io/illustrated-transformer/

The Annotated Transformer: http://nlp.seas.harvard.edu/2018/04/03/attention.html

discuss

order

giacaglia|7 years ago

I reference one of the articles but I hadn’t look at the other one! Very interesting. Thanks for sharing

panic|7 years ago

You do more than reference it -- you've copied a bunch of text and figures from it as well. Search for "The encoder’s inputs first flow through a self-attention layer" and read on from there. Most of the article is a word-for-word copy.