top | item 35796499

(no title)

SmooL | 2 years ago

The proposed solution is to feed relevant data from a database of "ground truth facts" into the query (I'm assuming using the usual method of similarity search leveraging embedding vectors).

This solution... doesn't prohibit hallucinations? As far as I can tell it only makes them less likely. The AI is still totally capable of hallucinating, it's just less likely to hallucinate an answer to _question X_ if the query includes data that has the answer.

I've been thinking that it might be useful if you could actually _remove_ all the stored facts that the LLM has inside of it. I believe that an LLM that didn't natively know a whole bunch of random trivia facts, didn't know basic math, didn't know much about anything _except_ what was put into the initial query would be valuable. The AI can't hallucinate anything if it doesn't know anything to hallucinate.

How you achieve this practically I have no clue. I'm not sure it's even possible to remove the knowledge that 1+1=2 without removing the knowledge of how to write a python script one could execute to figure it out.

discuss

order

pjc50|2 years ago

Interesting this was the "old" version of AI, as done by people like Cycorp: https://en.wikipedia.org/wiki/Cyc

They've got a big database of logical reasoning propositions that they have been trying to do a much more formal-logic process with.