(no title)
disgruntledphd2 | 4 days ago
I'm honestly not sure how this issue could be solved. Like, fundamentally LLMs are next (or N-forward) token predictors. They don't have any way (in and of themselves) to ground their token generations, and given that token N is dependent on all of tokens (1...n-1) then small discrepancies can easily spiral out of control.
simianwords|4 days ago
disgruntledphd2|4 days ago