(no title)
Ucalegon | 3 months ago
Why are we assuming just because the prompt responds that it is providing proper outputs? That level of trust provides an attack surface in of itself.
Ucalegon | 3 months ago
Why are we assuming just because the prompt responds that it is providing proper outputs? That level of trust provides an attack surface in of itself.
setopt|3 months ago
Do you have the same opinion if Google chooses to delist any website describing how to run apps as root on Android from their search results? If not, how is that different from lobotomizing their LLMs in this way? Many people use LLMs as a search engine these days.
> Why are we assuming just because the prompt responds that it is providing proper outputs?
"Trust but verify." It’s often easier to verify that something the LLM spit out makes sense (and iteratively improve it when not), than to do the same things in traditional ways. Not always mind you, but often. That’s the whole selling point of LLMs.
cachvico|3 months ago
Ucalegon|3 months ago