HACKER Q&A
📣 nycdatasci

Has anyone achieved recursive self-improvement with agentic tools?


It feels like all the necessary components are finally available to build a self-reinforcing development loop.

Theoretically, we can now task tools like Claude Code or OpenClaw to monitor a git repo, analyze the abstractions in completed work, and then autonomously generate new agents or skills capable of handling similar tasks in the future.

Is anyone successfully running a loop like this? I’m curious if anyone here has shifted the majority of their time from writing code to crafting these systems—essentially bootstrapping agents that learn from the repo history to build better agents. I'd love to hear from those pushing the boundaries on this.


  👤 nycdatasci Accepted Answer ✓
To head off the semantics debate: I don't mean a model rewriting its own source code. I'm asking about 'process recursion'—systems that analyze completed work to autonomously generate new agents or heuristics for future tasks.

👤 drsalt
I've tried to replicate the real world, so I give my agents backstories, triabl loyalties, and deep-seated character flaws. my agents try to dominate and manipulate each other. they make sure to take credit for every line code. I have manager agents that promote based on shared hobbies. so far it's going well.

👤 sdrinf
I'm working on something like this. Specifically, I'm doing recursive self-improvement via autocatalysis -but predominantly in writing/research / search tasks. It's very early, but shows some very interesting signs.

The purely code part you described is a bit of an "extra steps" -you can just... vscode open target repo, "claude what does this do, how does it do it, spec it out for me" then paste into claude code for your repo "okay claude implement this". This sidesteps the security issue, the deadly trifecta, and the accumulation of unused cruft.


👤 acoyfellow
I do a fun orchestration system for long running loops on exe.dev (small write up docs.coey.dev) and I feel like I have super powers.

Self healing, I try two ways:

1) use a memory tool to store learnings for next iteration (Deja.coey.dev) and have the loop system instructions tell how to use it. One orchestrator, and sequential worker agents who run til their context is full and then hand off to the next run with learnings

2) the agent Shelley on exe can search past convos when promoted too for continuation.

I’ve been doing this with great success just “falling” into the implementation after proper guardrails are placed


👤 allinonetools_
I’ve experimented with semi-recursive loops where agents review outputs and refine prompts or workflows, but fully autonomous self-improvement still feels fragile. Most gains come from structured feedback and constraints rather than open-ended recursion. Stability becomes the real challenge.

👤 dormento
I think the key to really "unlock" these things is to separate as much as possible from where it can do harm (no important credentials, no shared identify, etc) then just give it its own home folder, its own credentials and let it rip.

You could technically instruct the agent to pilot local ollama and launch minions for "dumb" tasks in parallel, but i don't know if it could break out and modify the file system this way... but then, if it resides say in its own VPS, the damage will be contained.