(no title)
swid | 2 months ago
And if an LLM is consistent, even with a high temp, it could give the same PR the same grade while choosing different words to say.
The tokens are still chosen from the distribution, so a higher probability of the same grade will result in the same grade being chosen regardless of the temp set.
smt88|2 months ago
The model could "assess" the code qualitatively the same and still give slightly different letter grades.