top | item 45862338

(no title)

reyqn | 3 months ago

NMT doesn't "contain" tranformers and deep RNNs, it can use them. LLMs use a transformer architecture, not everything using a transformer architecture is an LLM. NMT can actually use an LLM, but that's not the case according to the documentation you linked, they use a parallel dataset to train their models.

discuss

order

lostmsu|3 months ago

> they use a parallel dataset to train their models

If you want to be pedantic you should look up the LLM definition.

reyqn|3 months ago

Care to explain why?