Well it's cool that they released a paper, but at this point it's been 11 months and you can't download a Titans-architecture model code or weights anywhere. That would put a lot of companies up ahead of them (Meta's Llama, Qwen, DeepSeek).
Closest you can get is an unofficial implementation of the paper https://github.com/lucidrains/titans-pytorch
alyxya|2 months ago
tyre|2 months ago
You don't necessarily have to prove it out on large foundation models first. Can it beat out a 32b parameter model, for example?
p1esk|2 months ago
If Google is not willing to scale it up, then why would anyone else?
nickpsecurity|2 months ago
So, I think they could default on doing it for small demonstrators.
m101|2 months ago
UltraSane|2 months ago
root_axis|2 months ago
innagadadavida|2 months ago
SilverSlash|2 months ago
mupuff1234|2 months ago
Is that supposed to be a long time? Seems fair that companies don't rush to open up their models.
informal007|2 months ago
jstummbillig|2 months ago
mapmeld|2 months ago
AugSun|2 months ago
FpUser|2 months ago