top | item 31728203

(no title)

alew1 | 3 years ago

> Unfortunately, that argument applies to you, yourself.

Does it? I don’t think it would even apply to a reinforcement learning agent trained to maximize reward in a complex environment. In that setting, perhaps the agent could learn to use language to achieve its goals, via communication of its desires. But LaMDA is specifically trained to complete documents, and would face selective pressure to eliminate any behavior that hampers its ability to do that — for example, behavior that attempts to use its token predictions as a side channel to communicate its desires to sympathetic humans.

Again, this is not an argument that LaMDA is not sentient, just that the practice of “prompting LaMDA with partially completed dialogues between a hypothetical sentient AI and a human, and seeing what it predicts the AI will say” is not the same as “talking to LaMDA.”

Suppose LaMDA were powered by a person in a room, whose job it was to predict the completions of sentences. Just because you get the person to predict “I am happy” doesn’t mean the person is happy; indeed, the interface that is available to you, from outside the room, really gives you no way of probing the person’s emotions, experiences, or desires at all.

discuss

order

rendall|3 years ago

> Just because you get the person to predict “I am happy” doesn’t mean the person is happy; indeed, the interface that is available to you, from outside the room, really gives you no way of probing the person’s emotions, experiences, or desires at all.

But in that case the "sentience" (whatever that means) in question would have nothing to do with the person, who is just facilitating whatever ruleset enables the prediction. The person in that case is merely acting as a node in the neural network or whatever. Sure they would have feelings, being human, but they aren't the sentient being in question. Any apparent sentience would derive from the ruleset itself.