(no title)
alekandreev | 11 months ago
We wanted the long context recipe to be friendly for finetuning, and training at 128k is a bit of a pain we don't do it. For inference, we see inference at 128k with the 5/1 is close to RAM usage for a fully-global-layer model at 32k.
Individual attention layers are always dense.
sidkshatriya|11 months ago
[Edit: You answered the question when you said that individual attention layers are always dense.]