Agree. I’d like more fine grained control of context and compaction. If you spend time debugging in the middle of a session, once you’ve fixed the bugs you ought to be able to remove everything related to fixing them out of context and continue as you had before you encountered them. (Right now depending on your IDE this can be quite annoying to do manually. And I’m not aware of any that allow you to snip it out if you’ve worked with the agent on other tasks afterwards.)I think agents should manage their own context too. For example, if you’re working with a tool that dumps a lot of logged information into context, those logs should get pruned out after one or two more prompts.
Context should be thought of something that can be freely manipulated, rather than a stack that can only have things appended or removed from the end.
nr378|1 day ago
There's some challenges around the LLM having enough output tokens to easily specify what it wants its next input tokens to be, but "snips" should be able to be expressed concisely (i.e. the next input should include everything sent previously except the chunk that starts XXX and ends YYY). The upside is tighter context, the downside is it'll bust the prompt cache (perhaps the optimal trade-off is to batch the snips).
mksglu|1 day ago
lowbloodsugar|1 day ago
FuckButtons|1 day ago
mksglu|1 day ago
esperent|1 day ago
unknown|1 day ago
[deleted]
dsclough|17 hours ago
esperent|1 day ago
I've set up a hook that blocks directly running certain common tools and instead tells Claude to pipe the output to a temporary file and search that for relevant info. There's still some noise where it tries to run the tool once, gets blocked, then runs it the right way. But it's better than before.
wonnage|1 day ago
MichaelDickens|1 day ago
My intuition is that this should be almost trivial. If I copy/paste your long coding session into an LLM and ask it which parts can be removed from context without losing much, I'm confident that it will know to remove the debugging bits.
bbatha|1 day ago
mullingitover|1 day ago
Looks interesting.
[1] https://arxiv.org/html/2510.04618v1
mksglu|1 day ago
8note|1 day ago
isnt that how thinking works? intermediate tokens that then get replaced with the reuslt?
8note|1 day ago
jaredsohn|1 day ago
snowhale|22 hours ago
[deleted]