(no title)
aszen | 5 months ago
We already know that larger models hallucinate less since they can store more information, are there any smaller models which hallucinate less
aszen | 5 months ago
We already know that larger models hallucinate less since they can store more information, are there any smaller models which hallucinate less
gobdovan|5 months ago
excerpt: Claim: Avoiding hallucinations requires a degree of intelligence which is exclusively achievable with larger models. Finding: It can be easier for a small model to know its limits. For example, when asked to answer a Māori question, a small model which knows no Māori can simply say “I don’t know” whereas a model that knows some Māori has to determine its confidence. As discussed in the paper, being “calibrated” requires much less computation than being accurate.
K0balt|5 months ago
They are merely an output that we find unuseful, but in all other ways is optimal based on the training data, context, and model precision and parameters being used.
euroderf|5 months ago