Discussion about this post

User's avatar
JP's avatar

Your golden rules for context management are solid. The "don't fill it up" one is the hardest for people to internalise because it feels counterintuitive. You'd think more context = better results, but it's the opposite past a certain point.

I've been taking the "regularly clear the context" rule to its logical extreme with multi-agent setups. Rather than clearing and restarting within one session, you delegate chunks of work to sub-agents that each start with a clean window. They do their thing in isolation and report back a summary. The main thread never sees the intermediate mess. Covered the mechanics of how Codex does this natively https://reading.sh/codex-has-a-multi-agent-mode-and-almost-nobody-is-using-it-088e44f774ef and it builds on exactly what you're describing here. Your SPEC.md approach would pair well with it too since each sub-agent could reference the spec independently without polluting the parent context.

Pawel Jozefiak's avatar

Your point about working in small, well-defined chunks resonates deeply. I've been building Wiz - a personal AI agent using Claude Code - and the biggest lesson was exactly this: the AI is brilliant at executing clear specs, but struggles with ambiguous scope.

The feedback loop through testing is crucial. I found that building verification layers into my agent workflows made the difference between "cool demo" and "reliable system." Now my agents run overnight with confidence because they have guardrails.

One thing I'd add: persistent context matters as much as chunk size. I maintain a CLAUDE.md file that grows over time with preferences, past decisions, and learned mistakes. It's what turns Claude Code from a conversation partner into a team member with memory.

For anyone looking at AI-assisted coding honestly (what works vs. the hype), I wrote about my experience here: https://thoughts.jock.pl/p/claude-code-review-real-testing-vs-zapier-make-2026

No posts

Ready for more?