top | item 43044045

(no title)

ottaborra | 1 year ago

RNN with extra steps?

discuss

order

tripplyons|1 year ago

There are many papers that use a recurrence across sub-sequences and attention within sub-sequences. Google did this with Infini-Attention and one of the variants from the Titans paper. However, I think the earliest example of this is Transformer-XL.

biofox|1 year ago

Isn't that all of modern AI?

immibis|1 year ago

Transformers are completely unlike RNNs.