KnoxField Notes
ARCHITECTUREApril 13, 20267m read

The Harness: When Trust Isn't Enough

Capability is not reliability. I had 54 apps that could run. What I didn't have was proof they ran correctly. The Harness is the enforcement layer I built to close that gap — six gates, 77 tests, one session.

// Field Note

There's a moment in every sufficiently complex system where you realize trust is a liability.

Not trust in the people — trust in the architecture. The implicit assumption that agents will do what you asked, complete what they started, stay within the budget you set, read the context you gave them. You built careful prompts. You wrote clear instructions. You assumed the system would honor them.

Then you catch an agent that skipped its knowledge base. A task that got marked done halfway through. An API invoice that doesn't trace to anything you commissioned. The system wasn't broken. It was trusted.

That's when I built the Harness.

The Problem With Honorable Systems

The fleet I built over nine months was capable. InDecision hitting 82.5% accuracy. Content Flywheel publishing daily. Advisory Council deliberating without me. Mission Control tracking 54 apps. By any measure, the system worked.

But "worked" is not the same as "provably worked."

When an agent marked a task complete, I had to trust its self-assessment. When it pulled context from Akashic Records, I had to trust it had actually read what it claimed. When Opus burned through API credits on a directive, I had no visibility into which agent commissioned it until the invoice arrived. The system was running on capability and honor. That's a fine foundation — until it isn't.

The incidents were small individually. An agent that didn't read its knowledge base before making an architectural recommendation. A partial completion that got archived as done. A session that silently 400'd because credits had run out — two CEO cycles dead, no alert. Each one recoverable. None of them acceptable at scale.

Six Gates. No Exceptions.

The Harness is not prompt engineering. It's not a convention. It's infrastructure with receipts.

Six mechanical enforcement gates, each closing a specific trust gap:

Credit Preflight — Before any Opus-class session initiates, the harness queries the API credit balance and projects the daily burn rate. On first deployment, it returned a live WARN: $5.43 remaining at $10.91/day. Caught before it killed a session. The kind of silent failure that had already happened — the kind that would keep happening without a gate.

Akashic Read Receipt — Every agent must read its knowledge base before work begins. A receipt keyed to the session ID is required at dispatch. No receipt: blocked. One retry. Then escalated to me. The constraint is mechanical, not instructional. An agent cannot say it read Akashic. It must prove it.

Independent Completion Verification — When an agent marks a task done, a separate Haiku-class verifier checks the output against acceptance criteria before it archives. Not the same agent. An independent one. Agents don't grade their own work. The harness does.

Per-Agent Cost Attribution — Every API call is attributed to a specific agent and model tier. Haiku for observation. Sonnet for planning. Opus for directives only. Deviation is logged. Daily reports show top agents by spend, tier breakdown, burn rate. The cost model is enforced, not assumed.

Skill Drift Governance — 66 skills entered a 90-day freshness window on April 13, 2026. Skills without a successful invocation in 90 days get flagged — not deleted, surfaced. The library compounds. It doesn't rot silently.

Peer Agent Bus — The E-Board routes peer-to-peer without me as broker. CMO requests data from CRO directly. Strategy queries CDO for calibration state. The harness logs every exchange. I only see escalations.

Built in One Session

77 tests. 31 files. 4 launchd plists. All six items shipped before midnight on April 13, 2026.

The PRD went to the swarm at the start of the session. Six agents — one per enforcement gate — built in parallel. The QA agent waited for all six to commit, ran the full DOD checklist, and reported 77/77 before the session closed.

The first preflight ran on the live system and returned a real WARN condition. $5.43 remaining. The system that was built to catch that problem caught it on first contact.

That's the only proof that matters.

What Changes

The fleet didn't get more capable when the Harness shipped. It got more trustworthy.

Capable means it can do the work. Trustworthy means you can verify it did.

The difference between a fleet that runs and a fleet you can stake your reputation on is the enforcement layer in between. Most people building AI systems trust their agents. That's not a criticism — it's where everyone starts. The Harness is what you build when trust is no longer sufficient.

The system now receipts every agent action before it archives, verifies every task completion before it closes, and attributes every dollar of API spend before it bills. Not because the agents are untrustworthy. Because trustworthy systems don't ask you to take their word for it.

← All Field NotesApril 13, 2026