top | item 44555669

(no title)

orderone_ai | 7 months ago

Man, that is truly fascinating. Do you have ideas on how to expand the study to capture broader analysis like that...?

discuss

order

johnsmith1840|7 months ago

I was trying to solve AGI at the time this was just a side study I did to better understand how models forget the effect was not what I was looking for.

It could be expanded to better understand alignment.

But the resolution makes that cost prohibitive.

I did ~100 runs on different sizes but inferencing 100s of thousands of times made it computationally prohibitive. The key random statement is what allowed accurate measurements of the model.

The equivalent would be for every fine tuning data you train on run the entire evaluation dataset through it.

victor22|7 months ago

Yeah I didnt understand shit either