(no title)
agosta | 27 days ago
There is without a doubt a variation of this prompt you can pre-test to successfully bait the LLM into exfiltrating almost any data on the user's machine/connected accounts.
That explains why you would want to go out and buy a mac mini... To isolate the dang thing. But the mini would ostensibly still be connected to your home network. Opening you up to a breach/spill over onto other connected devices. And even in isolation, a prompt could include code that you wanted the agent to run which could open a back door for anyone to get into the device.
Am I crazy? What protections are there against this?
fwip|27 days ago
Nothing that will work. This thing relies on having access to all three parts of the "lethal trifecta" - access to your data, access to untrusted text, and the ability to communicate on the network. What's more, it's set up for unattended usage, so you don't even get a chance to review what it's doing before the damage is done.
toomuchtodo|27 days ago
“Exploit vulnerabilities while the sun is shining.” As long as generative AI is hot, attack surface will remain enormous and full of opportunities.
uxhacker|27 days ago
For example I would love for an agent to do my grocery shopping for me, but then I have to give it access to my credit card.
It is the same issue with travel.
What other useful tasks can one offload to the agents without risk?
johnsmith1840|27 days ago
Control all input out of it with proper security controls on it.
While not perfect it aleast gives you a fighting chance when your AI decides to send a random your SSN and a credit card to block it.
bretpiatt|27 days ago
sebmellen|27 days ago
xXSLAYERXx|27 days ago
mmooss|27 days ago
LLMs obviously can be controlled - their developers do it somehow or we'd see much different output.
zbentley|27 days ago
Such a supervisor layer for a system as broad and arbitrary as an internet-connected assistant (clawdbot/openclaw) is also not an easy thing to create. We're talking tons of events to classify, rapidly-moving API targets for things that are integrated with externally, and the omnipresent risk that the LLMs sending the events could be tricked into obfuscating/concealing what they're actually trying to do just like a human attacker would.
unknown|27 days ago
[deleted]
hazeii|27 days ago
BrouteMinou|27 days ago
Social, err... Clanker engineering!
jfyi|27 days ago
This is something computers in general have struggled with. We have 40 years of countermeasures and still have buffer overflow exploits happening.