Ask HN: Has anyone achieved recursive self-improvement with agentic tools?
10 points| nycdatasci | 18 days ago
Theoretically, we can now task tools like Claude Code or OpenClaw to monitor a git repo, analyze the abstractions in completed work, and then autonomously generate new agents or skills capable of handling similar tasks in the future.
Is anyone successfully running a loop like this? I’m curious if anyone here has shifted the majority of their time from writing code to crafting these systems—essentially bootstrapping agents that learn from the repo history to build better agents. I'd love to hear from those pushing the boundaries on this.
drsalt|18 days ago
ra0x3|17 days ago
Things are time bound by instruction creation - at some point you still need a human to dictate the instructions that the orchestrated agents use. From there I've found that -- (1) derive a goal from the instructions (2) break that goal into tasks (3) order those tasks into a DAG (5) spawn the agents to work via the DAG -- seems to be doing everything I want it to do.
sdrinf|18 days ago
The purely code part you described is a bit of an "extra steps" -you can just... vscode open target repo, "claude what does this do, how does it do it, spec it out for me" then paste into claude code for your repo "okay claude implement this". This sidesteps the security issue, the deadly trifecta, and the accumulation of unused cruft.
acoyfellow|17 days ago
Self healing, I try two ways:
1) use a memory tool to store learnings for next iteration (Deja.coey.dev) and have the loop system instructions tell how to use it. One orchestrator, and sequential worker agents who run til their context is full and then hand off to the next run with learnings
2) the agent Shelley on exe can search past convos when promoted too for continuation.
I’ve been doing this with great success just “falling” into the implementation after proper guardrails are placed
haute_cuisine|17 days ago
dormento|17 days ago
You could technically instruct the agent to pilot local ollama and launch minions for "dumb" tasks in parallel, but i don't know if it could break out and modify the file system this way... but then, if it resides say in its own VPS, the damage will be contained.
allinonetools_|17 days ago
nycdatasci|18 days ago
jvalencia|18 days ago
0xbadcafebee|16 days ago
Your agent will basically never get good at learning. The only ways to get closer to that are 1) fine-tuning (expensive, slow and inaccurate), and 2) reinforcement learning (slow and inaccurate). So you can't just build an agent that automatically, incrementally gets better, without waiting for 10+ years for the process to iterate sufficiently. (ask AI researchers, this has been the case for a long time)
However, you can build an agent that can iterate on one specific problem so much that it becomes amazingly good at it, and then do that on another specific problem, and another, until you have a whole bunch of mini-experts. Then you can use those together. To get better than that... use a new model, new prompting techniques, etc.