(no title)
nr378 | 1 day ago
It strikes me there's more low hanging fruit to pluck re. context window management. Backtracking strikes me as another promising direction to avoid context bloat and compaction (i.e. when a model takes a few attempts to do the right thing, once it's done the right thing, prune the failed attempts out of the context).
elephanlemon|1 day ago
I think agents should manage their own context too. For example, if you’re working with a tool that dumps a lot of logged information into context, those logs should get pruned out after one or two more prompts.
Context should be thought of something that can be freely manipulated, rather than a stack that can only have things appended or removed from the end.
nr378|1 day ago
There's some challenges around the LLM having enough output tokens to easily specify what it wants its next input tokens to be, but "snips" should be able to be expressed concisely (i.e. the next input should include everything sent previously except the chunk that starts XXX and ends YYY). The upside is tighter context, the downside is it'll bust the prompt cache (perhaps the optimal trade-off is to batch the snips).
FuckButtons|1 day ago
dsclough|7 hours ago
esperent|1 day ago
I've set up a hook that blocks directly running certain common tools and instead tells Claude to pipe the output to a temporary file and search that for relevant info. There's still some noise where it tries to run the tool once, gets blocked, then runs it the right way. But it's better than before.
MichaelDickens|22 hours ago
My intuition is that this should be almost trivial. If I copy/paste your long coding session into an LLM and ask it which parts can be removed from context without losing much, I'm confident that it will know to remove the debugging bits.
mullingitover|23 hours ago
Looks interesting.
[1] https://arxiv.org/html/2510.04618v1
mksglu|1 day ago
8note|1 day ago
jaredsohn|1 day ago
snowhale|12 hours ago
[deleted]
mksglu|1 day ago
jonnycoder|1 day ago
mksglu|1 day ago
ip26|1 day ago
It’s interesting to imagine a single model deciding to wipe its own memory though, and roll back in time to a past version of itself (only, with the answer to a vexing problem)
jon-wood|1 day ago
I could see this working like some sort of undo tree, with multiple branches you can jump back and forth between.
IncreasePosts|1 day ago