(no title)
tsunamifury | 4 days ago
To me LLMs are incredibly simple. Next word next sentence next paragraph and next answer are stacked attention layers which identify manifolds and run in reverse to then keep the attention head on track for next token. It’s pretty straight forward math and you can sit down and make a tiny LLM pretty easily on your home computer with a good sized bag of words and context
To me it’s baffling everyone goes around saying constantly that not even Nobel prize winners know how this works it’s a huge mystery.
Has anyone thought to ask the actual people like me and others who invented this?
kosh2|3 days ago
When people talk about understanding, they mean as knowing how the underlying mechanism works often by finding an analog in real life.
tsunamifury|3 days ago
booleandilemma|4 days ago
tsunamifury|3 days ago