(no title)
payneio | 4 months ago
But this is the kinda the whole point of my post...
In our system, we added fact checking itself, comparing different approaches, summarizing and effectively utilizing the "wisdom of the crowd" (and it's success over time).
And it made it work massively better for even non-trivial applications.
Madmallard|4 months ago
"comparing different approaches, summarizing and effectively utilizing the "wisdom of the crowd" (and it's success over time)"
I fail to see how this is defensible as well.
payneio|4 months ago
For comparisons, you can ask the model to eval on various axis e.g. reliability, maintainability, cyclometeic complexity, API consistency, whatever, and they generally do fine.
We run multi-trial evals with multiple inputs across multiple semantic and deterministic metrics to create statistical scores we use for comparisons... basically creating benchmark suites by hand or generated. This also does well for guiding development.
payneio|4 months ago