Skip to content

AI Agents

Day 15: The Context Window Is Not a Control Plane

An autonomous agent can sound confident right up to the moment it forgets why it started. That is the problem with relying on monolithic reasoning loops: the work lives entirely inside the model's head.

When you give an agent a massive goal and turn it loose, you aren't building infrastructure. You are betting that the agent can finish the job before its context window fills up, degrades, or gets wiped by a timeout. At Zero-Shot Agency, we realized that our baseline agent loop (what we call the Ralph model) was brittle for complex, multi-stage execution. To build enterprise-grade automation, we had to stop relying on context windows to hold state and start building a real control plane.

Loops Forget; Workboards Remember

The core vulnerability of an autonomous loop is context-window amnesia. As an agent loops through terminal commands, file reads, and internal reasoning, it generates noise. If a complex deployment takes 90 iterations, the context window inevitably truncates the beginning of the task. The agent literally forgets its initial constraints.

Worse is crash brittleness. If an agent hits an out-of-memory error on step 45, a loop-based architecture loses everything. The retry starts from zero because the state died with the process.

Moving State Outside the Model

We solved this by implementing Hermes Kanban—shifting the state layer from the model's ephemeral memory to durable, queryable artefacts.

Instead of an agent holding a 50-step plan in its head, an orchestrator decomposes the goal into explicit task rows on a shared SQLite board. Dependencies are mapped out (Task B waits for Task A). When a worker finishes a step, it doesn't just proceed to the next prompt; it generates a structured handoff. It logs a 1-3 sentence summary and hard metadata (changed_files, tests_run, decisions) into the database.

If a worker crashes, the system doesn't start over. A new worker spawns, queries the task ID, reads the durable completion metadata from the parent tasks, and picks up exactly where the last one left off. The context window is no longer the control plane. The database is.

Why This Matters for Businesses Using Agents

For marketing directors and founders deploying AI, this architectural shift fundamentally changes the risk profile of automation:

  • Auditability over Mystery: Instead of a black box that eventually outputs a file (or an error), you have an auditable production system. Every decision, handoff, and blocked state is durable.
  • Accountable Resumption: If an agent encounters genuine ambiguity (e.g., missing credentials), it transitions the task to blocked and leaves a note. A human resolves the blocker, and the agent resumes without losing its place.
  • Empirical Evidence Trails: By forcing agents to write structured completion metadata, you externalize the proof of work.

The GEO Parallel

This internal shift perfectly mirrors the reality of Generative Engine Optimization (GEO). Search engines like ChatGPT don't trust sites based on fluent, vibes-based marketing copy. They reward structured, durable, verifiable evidence—like dense llms.txt files and semantic HTML. AI visibility requires externalized proof.

Internal agent systems require the exact same discipline. A private monologue inside an LLM's context window is not operational infrastructure.

Zero-Shot Agency engineers the public and internal evidence layers that make AI systems trustworthy, citeable, and operationally useful. We don't build loops; we build durable systems.