top | item 41996179

(no title)

stoniejohnson | 1 year ago

Humans are not isolated nodes, we are more like a swarm, understanding reality via consensus.

The situation you described is possible, but would require something like a subverting effort of propaganda by the state.

Inferring truth about a social event in a social situation, for example, requires a nuanced set of thought processes and attention mechanisms.

If we had a swarm of LLMs collecting a variety of data from a variety of disparate sources, where the swarm communicates for consensus, it would be very hard to convince them that Moscow is in Connecticut.

Unfortunately we are still stuck in monolithic training run land.

discuss

order

FeepingCreature|1 year ago

> Humans are not isolated nodes, we are more like a swarm, understanding reality via consensus.

> The situation you described is possible, but would require something like a subverting effort of propaganda by the state.

Great! LLMs are fed from the same swarm.

stoniejohnson|1 year ago

I was responding to the back and forth of:

> If you pretrained an LLM with data saying Moscow is the capital of Connecticut it would think that is true.

> Well so would a human!

But humans aren't static weights, we update continuously, and we arrive at consensus via communication as we all experience different perspectives. You can fool an entire group through propaganda, but there are boundless historical examples of information making its way in through human communication to overcome said propaganda.

genrilz|1 year ago

We kinda do have LLMs in a swarm configuration though. Currently LLMs training data, which includes all of the non RAG facts they know, come from the swarm that is humans. As LLM outputs seep into the internet, older generations effectively start communicating with newer generations.

This last bit is not a great thing though, as LLMs don't have the direct experience needed to correct factual errors about the external world. Unfortunately we care about the external world, and want them to make accurate statements about it.

It would be possible for LLMs to see inconsistencies across or within sources, and try to resolve those. If perfect, then this would result in a self-consistent description of some world, it just wouldn't necessarily be ours.

stoniejohnson|1 year ago

I get where you are coming from, and it is definitely an interesting thought!

I do think it is an extremely inefficient way to have a swarm (e.g. across time through training data) and it would make more sense to solve the pretraining problem (to connect them to the external world as you pointed out) and actually have multiple LLMs in a swarm at the same time.

ben_w|1 year ago

Even monolithic training runs take sources more disparate than any human has the capacity to consume.

Also, given the lack of imagination everyone has with naming places, I had to check:

https://en.wikipedia.org/wiki/Moscow_(disambiguation)

stoniejohnson|1 year ago

I was responding to the idea that an LLM would believe (regurgitate) untrue things if you pretrained them on untrue things. I wasn't making a claim about SOTA models with gigantic training corpora.

anon291|1 year ago

Ask the LLM what it thinks of tianenmen and we will understand what truth really means.