Is anybody shocked that when prompted to be a psychotherapy client models display neurotic tendencies? None of the authors seem to have any papers in psychology either.
There is nothing shocking about this, precisely, and yes, it is clear by how the authors are using the word "psychometric" that they don't really know much about psychology research either.
The broader point is that if you start from the premise that LLMs cannot ever be considered the way sentient beings are then all abuse becomes reasonable.
We are in danger of unscientifically returning to the point at which newborn babies weren’t considered to feel pain.
Given our lack of a definition for sentience it’s unscientific to presume no system has any sentient trait at any level.
I'm not shocked at all. This is how the tech works at all, word prediction until grokking occurs. Thus like any good stochastic parrot, if it's smart when you tell it it's a doctor, it should be neurotic when you tell it it's crazy. it's just mapping to different latent spaces on the manifold
I think popular but definitely-fictional characters are a good illustration: If the prompt describes a conversation with Count Dracula living in Transylvania, we'll percieve a character in the extended document that "thirsts" for blood and is "pained" by sunlight.
Switching things around so that the fictional character is "HelperBot, AI tool running in a datacenter" will alter things, but it doesn't make those qualities any less-illusory than CountDraculatBot's.
D-Machine|25 days ago
MrSkelter|25 days ago
We are in danger of unscientifically returning to the point at which newborn babies weren’t considered to feel pain.
Given our lack of a definition for sentience it’s unscientific to presume no system has any sentient trait at any level.
fkdk|24 days ago
agarwaen163|25 days ago
Terr_|25 days ago
Switching things around so that the fictional character is "HelperBot, AI tool running in a datacenter" will alter things, but it doesn't make those qualities any less-illusory than CountDraculatBot's.