(no title)
briga
|
5 months ago
I have a theory: all these people reporting degrading model quality over time aren't actually seeing model quality deteriorate. What they are actually doing is discovering that these models aren't as powerful as they initially thought (ie. expanding their sample size for judging how good the model is). The probabilistic nature of LLM produces a lot of confused thinking about how good a model is, just because a model produces nine excellent responses doesn't mean the tenth response won't be garbage.
vintermann|5 months ago
Also, from what I understand from the article, it's not a difficult task but an easily machine checkable one, i.e. whether the output conforms to a specific format.
Spivak|5 months ago
lostmsu|5 months ago
nothrabannosir|5 months ago
What’s missing is the actual evidence. Which I would love of course. But assuming they’re not actively lying, this is not as subjective as you suggest.
chaos_emergent|5 months ago
zzzeek|5 months ago
gtsop|5 months ago
yieldcrv|5 months ago
I think they have receipts, but did not post them there
Aurornis|5 months ago
Even a simple graph of the output would be better than nothing, but instead it’s just an empty claim.
colordrops|5 months ago
Aurornis|5 months ago