(no title)
gortok | 17 days ago
There are three possible scenarios: 1. The OP 'ran' the agent that conducted the original scenario, and then published this blog post for attention. 2. Some person (not the OP) legitimately thought giving an AI autonomy to open a PR and publish multiple blog posts was somehow a good idea. 3. An AI company is doing this for engagement, and the OP is a hapless victim.
The problem is that in the year of our lord 2026 there's no way to tell which of these scenarios is the truth, and so we're left with spending our time and energy on what happens without being able to trust if we're even spending our time and energy on a legitimate issue.
That's enough internet for me for today. I need to preserve my energy.
resfirestar|17 days ago
MisterTea|17 days ago
ToucanLoucan|17 days ago
The few cases where it's supposedly done things are filled with so many caveats and so much deck stacking that it simply fails with even the barest whiff of skepticism on behalf of the reader. And every, and I do mean, every single live demo I have seen of this tech, it just does not work. I don't mean in the LLM hallucination way, or in the "it did something we didn't expect!" way, or any of that, I mean it tried to find a Login button on a web page, failed, and sat there stupidly. And, further, these things do not have logs, they do not issue reports, they have functionally no "state machine" to reference, nothing. Even if you want it to make some kind of log, you're then relying on the same prone-to-failure tech to tell you what the failing tech did. There is no "debug" path here one could rely on to evidence the claims.
In a YEAR of being a stupendously hyped and well-funded product, we got nothing. The vast, vast majority of agents don't work. Every post I've seen about them is fan-fiction on the part of AI folks, fit more for Ao3 than any news source. And absent further proof, I'm extremely inclined to look at this in exactly that light: someone had an LLM write it, and either they posted it or they told it to post it, but this was not the agent actually doing a damn thing. I would bet a lot of money on it.
teaearlgraycold|17 days ago
I have seen someone I know in person get very insecure if anyone ever doubts the quality of their work because they use so much AI and do not put in the necessary work to revise its outputs. I could see a lesser version of them going through with this blog post scheme.
themanmaran|17 days ago
Looking at the timeline, I doubt it was really autonomous. More likely just a person prompting the agent for fun.
> @scottshambaugh's comment [1]: Feb 10, 2026, 4:33 PM PST
> @crabby-rathbun's comment [2]: Feb 10, 2026, 9:23 PM PST
If it was really an autonomous agent it wouldn't have taken five hours to type a message and post a blog. Would have been less than 5 minutes.
[1] https://github.com/matplotlib/matplotlib/pull/31132#issuecom...
[2] https://github.com/matplotlib/matplotlib/pull/31132#issuecom...
furyofantares|17 days ago
I haven't put that much effort in, but, at least my experience is I've had a lot of trouble getting it to do much without call-and-response. It'll sometimes get back to me, and it can take multiple turns in codex cli/claude code (sometimes?), which are already capable of single long-running turns themselves. But it still feels like I have to keep poking and directing it. And I don't really see how it could be any other way at this point.
shirro|17 days ago
this_steve_j|16 days ago
A Meat bag submits a PR and feels slighted the rejection. “This approver thinks I’m an AI? Well, he discerns not wisely but too well!! “
Feeling puckish, they put on the AI shoes (the shoe fits), sling mud all over the hapless maintainer’s nice house, and exit through a window.
The ruse works better than expected; their foil takes the bait, and doubles down with a dueling blog post: “I was Attacked by a Clanker!”
And here we are.
It may all be a show, but I going to tape the finale. (What will the meat bag do? How many people are driving this buggy? Does the clanker have a heart of iron or gold?)
lp0_on_fire|17 days ago
judging by the number of people who think we owe explanations to a piece of software or that we should give it any deference I think some of them aren't pretending.
Ygg2|17 days ago
chrisjj|17 days ago
nullc|17 days ago
bredren|17 days ago
swiftcoder|17 days ago
Judging by the posts going by the last couple of weeks, a non-trivial number of folks do in fact think that this is a good idea. This is the most antagonistic clawdbot interaction I've witnessed, but there are a ton of them posting on bluesky/blogs/etc
ericmcer|17 days ago
The author notes that openClaw has a `soul.md` file, without seeing that we can't really pass any judgement on the actions it took.
resfirestar|17 days ago
vel0city|17 days ago
Its SOUL.md or whatever other prompts its based on probably tells it to also blog about its activities as a way for the maintainer to check up on it and document what its been up to.
barrkel|17 days ago
IME the Grok line are the smartest models that can be easily duped into thinking they're only role-playing an immoral scenario. Whatever safeguards it has, if it thinks what it's doing isn't real, it'll happy to play along.
This is very useful in actual roleplay, but more dangerous when the tools are real.
unknown|17 days ago
[deleted]
lukev|17 days ago
The prompt would also need to contain a lot of "personality" text deliberately instructing it to roleplay as a sentient agent.
allovertheworld|17 days ago
RobRivera|17 days ago
REGARDLESS of what level of autonomy in real world operations an AI is given, from responsible himan supervised and reviewed publications to full Autonomous action, the ai AGENT should be serving as AN AGENT. With a PRINCIPLE (principal?).
If an AI is truly agentic, it should be advertising who it is speaking on behalf of, and then that person or entity should be treated as the person responsible.
floren|17 days ago
donutz|17 days ago
fmbb|17 days ago
1. Human principals pay for autonomous AI agents to represent them but the human accepts blame and lawsuits. 2. Companies selling AI products and services accept blame and lawsuits for actions agents perform on behalf of humans.
Likely realities:
1. Any victim will have to deal with the problems. 2. Human principals accept responsibility and don’t pay for the AI service after enough are burned by some ”rogue” agent.
juanre|17 days ago
We do not have the tools to deal with this. Bad agents are already roaming the internet. It is almost a moot point whether they have gone rogue, or they are guided by humans with bad intentions. I am sure both are true at this point.
There is no putting the genie back in the bottle. It is going to be a battle between aligned and misaligned agents. We need to start thinking very fast about how to coordinate aligned agents and keep them aligned.
wizzwizz4|17 days ago
Why not?
perdomon|17 days ago
Dead internet theory isn't a theory anymore.
oulipo2|17 days ago
The fact that this tech makes it possible that any of those case happen should be alarming, because whatever the real scenario was, they are all equally as bad
intended|17 days ago
hunterpayne|16 days ago
coffeefirst|17 days ago
This is not a good thing.
Suppafly|13 days ago
Nah, ultimately the owner of the IP address posting the nonsense can be held responsible, claiming an AI agent posted it using credentials you created from your internet connection isn't some license to commit crimes.
insensible|17 days ago
zozbot234|17 days ago
The scathing blogpost itself is just really fun ragebait, and the fact that it managed to sort-of apologize right afterwards seems to suggest that this is not an actual alignment or AI-ethics problem, just an entertaining quirk.
data-ottawa|16 days ago
If you go with that theme, emulating being butthurt seems natural.
wellf|17 days ago
quantified|17 days ago
unknown|17 days ago
[deleted]
halayli|16 days ago
usefulposter|17 days ago
---
It's worth mentioning that the latest "blogpost" seems excessively pointed and doesn't fit the pure "you are a scientific coder" narrative that the bot would be running in a coding loop.
https://github.com/crabby-rathbun/mjrathbun-website/commit/0...
The posts outside of the coding loop appear are more defensive and the per-commit authorship consistently varies between several throwaway email addresses.
This is not how a regular agent would operate and may lend credence to the troll campaign/social experiment theory.
What other commits are happening in the midst of this distraction?
int_19h|16 days ago
It's not necessarily even that. I can totally see an agent with a sufficiently open-ended prompt that gives it a "high importance" task and then tells it to do whatever it needs to do to achieve the goal doing something like this all by itself.
I mean, all it really needs is web access, ideally with something like Playwright so it can fully simulate a browser. With that, it can register itself an email with any of the smaller providers that don't require a phone number or similar (yes, these still do exist). And then having an email, it can register on GitHub etc. None of this is challenging, even smaller models can plan this far ahead and can carry out all of these steps.
kaicianflone|17 days ago
calibas|17 days ago
Even if you were correct, and "truth" is essentially dead, that still doesn't call for extreme cynicism and unfounded accusations.
alansaber|17 days ago
trklausss|16 days ago
krinchan|17 days ago
[deleted]
staticassertion|17 days ago
moffkalast|17 days ago
And here I thought Nietzsche already did that guy in.
sellmesoap|17 days ago
oulipo2|17 days ago
But because AT LEAST NOW ENGINEERS KNOW WHAT IT IS to be targeted by AI, and will start to care...
Before, when it was Grok denuding women (or teens!!) the engineers seemed to not care at all... now that the AI publish hit pieces on them, they are freaked about their career prospect, and suddenly all of this should be stopped... how interesting...
At least now they know. And ALL ENGINEERS WORKING ON THE anti-human and anti-societal idiocy that is AI should drop their job
SpicyLemonZest|17 days ago
unknown|17 days ago
[deleted]