There's a failure mode in multi-agent systems that doesn't show up in benchmarks: procedural amnesia.
My self-improving agents could remember facts. They could remember context. But two weeks after successfully creating a Google Doc, writing a Day One journal entry, or making a phone call — they'd look at me blankly and say they didn't know how.
They knew what they'd done. They had no persistent memory of how.
The fix wasn't more context. It wasn't a bigger model. It was embarrassingly simple: a shared Markdown file called PLAYBOOKS.md.
The rule is one line: every agent, every time they successfully complete a non-trivial task, documents the exact steps before the session ends. Not a summary. The actual commands, the tools, the gotchas, the sequence. A recipe.
Before starting any task that sounds familiar, they read it first.
The result: agents that actually get smarter over time — not because the model improved, but because the system does. Institutional knowledge that survives session resets, model swaps, and context compaction.
The real insight: most "AI capability" problems are actually memory architecture problems. The model knows how to do the thing. The system just never wrote it down.