top | item 39697085

(no title)

vecinu | 1 year ago

This might be a naive question but how does one determine what "best" is for multiple subjects?

Even in your example, physics and mathematics could be curated for "best" when dealing with equations and foundational knowledge that has been hardened over decades. For history, climate change of invasion of Ukraine, isn't that sensitive to bias, manipulation and interpretation? These are not exact sciences.

discuss

order

mncharity|1 year ago

> how does one determine what "best" is for multiple subjects?

Perhaps invert the question - how to recognize "not-best"? If it's on a consensus list of common misconceptions, it's not-best. Science textbooks, web and outreach content, are thus often not-best. If the topic isn't the author's direct research or professional focus, it's likely non-best. People badly underestimate how rapidly expertise degrades as you blur from focus to subfield, let alone to broader field. Journalism is pervasively not-best. If the author won't be embarrassed by serving not-best, it likely is. Beware communities where avoiding not-best embarrassment isn't a dominating incentive.

> not exact sciences.

Most content fails even the newspaper test, that any professional familiar with the topic will recognize that it's wrong. This applies as much to science and engineering as to whatever. Not-best.

"Soft" fields do have challenges. Subcultures with incompatible "this work is great/trash" evaluations. Integration of diverse perspectives in general.

But note that agreement and uncertainty is often poorly characterized. A description of "A, B, and C" rather than "A. And also B and C, orders-of-magnitude down.". "B vs C!" rather than "A. And A.B vs A.C." Leaving out the important insight, the foundational context, is common. And sloppy argumentation. Not-best. Basically, there's opportunity for very atypically extensive pruning of not-best before becoming constrained by uncertainty rather than by effort.

Once you eliminate the not-best, whatever remains, however imperfect, is... far less wretched than usual.

bigger_cheese|1 year ago

Perhaps you can limit to training only on peer reviewed sources. The peer review process is imperfect but it is perhaps the closest thing we have to flagging something as the "best" answer for a particular topic.

History (maybe exception for scientific history), politics and current affairs I would say falls outside the scope of "scientific knowledge". I do not think it is possible to avoid bias in those topics.

A significant question is what the cutoff point would be for a model based on "scientific knowledge"? Should subjects like economics, philosophy etc be included as Scientific knowledge or Should it be limited to "hard" sciences only?

hotdogscout|1 year ago

Peer review isn't all that and a lot of subjects are censored and even disinformation is accepted if it flatters the ideological inclinations of the publication. See: Proximal Origins from Nature

colechristensen|1 year ago

You have to spend quite a lot of time thinking about quality and values. It becomes impossible as the size of the “best slice” you’re seeking gets smaller (top half is much easier than top ten percent, etc)

If your values are “everyone should agree with my opinions” you’ll have a garbage biased data set. There are other values though. Bias free is also impossible because having a definition of a perfectly neutral bias is itself a very strong bias.

animal_spirits|1 year ago

"Best" will be chosen by the creators of software for specific application uses. Medical software will use the "best" medical LLM under the hood. Programming software (Copilot et. all) will use the "best" programming LLM. General purpose language models will probably still be used by the public when doing internet searches. Or, an idea that just popped into my head, use a classifier to determine which model can most accurately answer the user's query, and send the query off to that model for a response.

edgyquant|1 year ago

Reread the comment you responded to. We aren’t discussing the best LLM here but how to determine the best source.

Terr_|1 year ago

> how does one determine what "best" is for multiple subjects?

Also, "best" depends on audience and use-case.

Imagine a horribly tone-deaf LLM-powered Sesame Street episode about the importance of recycling, illustrated by supply-demand graphs and Kekulé structures of plastic polymers.

wolverine876|1 year ago

What do you think of how that was addressed in the GP?