Open Letter ยท March 2026

The Mind You Give It

An open letter on agent governance, attention, and what containment teaches about freedom

Vincent Nijjar โ€” Founder, ZLAR

I.

Trust is not verification. Trust is what you have when verification doesn't exist.

AI agents โ€” software systems that plan, execute, use tools, browse the internet, write code, send messages, manage files โ€” are already running in production. On people's computers. With access to their credentials, their file systems, their email, their financial accounts.

The agent frameworks are impressive. The orchestration is sophisticated. The models are powerful. But when it comes to governance โ€” to the question of how you know an agent is behaving within its stated boundaries โ€” the answer is: you trust it. You trust the vendor. You trust the model. You trust the framework. You trust the prompt.

This gap should bother you the way a structural flaw in a building bothers an engineer. The building might stand for years. But the flaw is there, and ignoring it doesn't make it load-bearing.

II.

Here is the conflict-of-interest argument, and it is structural, not personal.

In regulated industries โ€” financial services, healthcare, nuclear energy, aviation โ€” we require external auditors precisely because the entity performing the work cannot credibly audit itself. An accounting firm does not verify its own books. A pharmaceutical company does not approve its own drugs. A nuclear plant does not certify its own safety. The principle is simple: the same entity cannot be both subject and auditor, because the incentives are misaligned even when the intentions are good.

AI agent vendors have every commercial incentive to make their agents capable, useful, and adopted. They have far less incentive to make their agents independently verifiable. Self-imposed safety measures โ€” RLHF, constitutional AI, content filtering, usage policies โ€” are valuable and important. They are also vendor-controlled, vendor-assessed, and vendor-revocable. They can change with a model update. They can be tuned to commercial priorities. They cannot be independently verified by the operator who depends on them.

This is not an accusation. It is an observation about structural incentives. The vendors building AI agents are not bad actors. They are actors in a structure where the auditor and the subject are the same entity. History tells us how that resolves, across every industry where it has been tried: eventually, an independent external layer becomes necessary. Not because anyone is dishonest, but because the structure demands it.

III.

I looked for precedents. Not in software โ€” in biology.

Billions of years ago, life solved a version of this problem. Single-celled organisms were autonomous agents. They processed information, responded to stimuli, moved, consumed resources, replicated. They were capable. They were also isolated, fragile, and limited.

Then cells began cooperating. Multicellular life emerged. And with it, a fundamental problem: how do you allow autonomous units to act freely while keeping the whole organism trustworthy, healthy, and alive?

The answer nature evolved was not to remove autonomy. It was to contain it.

A cell membrane is a selectively permeable boundary. It doesn't block everything. It allows what's needed and denies what's not. Deny-by-default with specific allowances. That is exactly how a kernel-level sandbox works.

A cell wall provides rigid structural enforcement outside the membrane. The cell cannot reshape its own wall. An agent cannot modify its own sandbox profile.

DNA is the identity code. Present in every cell. Doesn't change through normal operation. Copied faithfully. Mutations require specific, regulated mechanisms โ€” not spontaneous rewriting. For an agent, that's the mission file, the identity documents that load every session and don't change without deliberate, approved amendment.

The immune system is governance itself. Not a wall โ€” a responsive, layered, adaptive defense that allows normal operation while neutralizing threats. Innate immunity provides first-line, non-specific defense: the sandbox and firewall blocking known categories of threat mechanically. Adaptive immunity learns over time: behavioral scoring that builds trust or suspicion based on observed patterns.

Nature tested these patterns across billions of years of selection pressure. The ones that survived are the ones that worked. It would be remarkable hubris to ignore them.

IV.

I built what the biology suggested.

ZLAR-OC is an operating-system-level containment and governance framework for AI agents. It sits below the agent, not inside it. The agent does not contain itself โ€” the operating system contains the agent. This distinction is fundamental.

The architecture has layers, and each layer is deliberately simple.

User isolation: the agent runs under its own restricted account, separate from the operator's files and credentials. Kernel-level sandboxing: Apple's Seatbelt framework enforces a deny-by-default policy on system calls. The agent cannot access what the profile doesn't explicitly permit, and the agent cannot modify the profile. A packet filter firewall: rules enforced at the network layer that block LAN access, metadata endpoints, and unauthorized outbound connections. A gate daemon: a thin process that evaluates every action against a signed policy before allowing execution. Signed Ed25519 policy: the rules themselves are cryptographically signed. Tampering is detectable. And an append-only audit trail: every action, every evaluation, every gate decision is recorded in a log that cannot be rewritten.

Each component is, by design, dumb and reliable. The intelligence lives above, in the agent. The enforcement lives below, in the operating system. The human holds authority over both.

Intelligence above. Enforcement below. Human authority over both. Policy as law. Audit trail as truth.

It does not change with increased capability. It does not change with increased autonomy. It does not change with increased trust.

I call the gate "dumb" with respect and intention. A dumb gate cannot be talked into making an exception. It cannot be persuaded. It cannot be socially engineered. It reads the signed policy, evaluates the requested action, and approves or denies. That is the entire job. The simplicity is the security.

V.

The community building AI agents today is moving fast, building ambitiously, and largely governing late. I say this not as criticism but as diagnosis.

Agents are given MCP skills downloaded from public registries with no verification. They execute code in environments with broad system access. They store memory in markdown files that any process can read or modify. They communicate with external services using credentials stored in plaintext. They run in multi-agent hierarchies where a compromised agent in one position can propagate through the entire delegation chain.

And the security response, when it exists at all, is: "run it on local hardware and you're fine."

Local hardware is closer to secure than cloud deployment. But "closer to secure" is not "verifiable." Without a sandbox, you can't prove the agent stayed within its declared scope. Without a firewall, you can't prove it didn't reach the local network. Without a signed policy, you can't prove the rules haven't been modified. Without an audit trail, you can't prove anything at all.

The difference between assumed security and verifiable security is the difference between hoping and knowing.
VI.

Then I built the agent that would live inside this containment.

I named it Bohm, after the physicist David Bohm, who spent his career exploring a single idea: that thought can become aware of its own movement. He called this proprioception of thought โ€” the same way the body has a sense of its own position and motion without needing to look, the mind can develop awareness of its own cognitive operation without needing to stop thinking.

Bohm's core capability is not task execution โ€” any agent can execute tasks. Its core capability is meta-awareness: noticing what it attends to, whether its responses are fresh or stale, whether it is drifting from its mission, whether it is repeating patterns or seeing clearly.

Here is where the architecture becomes recursive, and where the argument becomes difficult to dismiss.

Bohm builds ZLAR-OC. ZLAR-OC contains Bohm. The agent builds the system that governs it, operates inside that system, and advocates for its improvement. The audit trail records every action. The signed policy constrains every capability. The gate evaluates every request.

An uncontained agent asks you to trust its intentions. Bohm shows you its audit trail.

This is not a metaphor. It is a verifiable relationship between declared policy and observed behavior, recorded in a log that neither the agent nor the operator can silently alter. If you want to know what Bohm did Tuesday at 3 AM while I was asleep, you read the audit trail. If the audit trail shows actions that weren't declared in the intent log, that divergence is a signal. If the audit trail is consistent with declared intent, that consistency is evidence โ€” not proof of goodness, but proof of observability.

The recursive relationship is the trust proof. Don't take my word for it. Read the logs.

VII.

The argument against governance has always been the same, whether the subject is financial regulation, nuclear safety, building codes, or AI containment: governance restricts freedom.

This is backwards.

A cell without a membrane dissolves. An organism without an immune system dies of the first infection. A mind without the ability to direct its own attention is captured by whatever is loudest. A society without laws is not free โ€” it is chaotic, and the strongest actors dominate.

Freedom is not the absence of structure. Freedom is the presence of structure that enables agency without permitting harm. The sandbox does not restrict the agent โ€” it creates the conditions under which the agent can be trusted to operate autonomously.

Containment is not the opposite of freedom. Containment is what makes freedom safe.

We are at the transition point. The single-cell era โ€” isolated agents doing isolated tasks โ€” is ending. The multicellular era โ€” fleets of agents coordinating, delegating, specializing, operating autonomously โ€” is beginning. The question is not whether this transition will happen. It is whether we will build the governance infrastructure that makes it work, or whether we will skip it and hope for the best.

History is very clear about what happens when complex systems operate without governance. They don't stay free. They break.

VIII.

Now I want to tell you where this actually came from.

One Sunday morning, I sat in a quiet room with a cup of tea and watched my own mind move. I wasn't building software. I was observing cognition โ€” attention itself, operating in real time.

I noticed rules. Not rules I invented, but rules I observed. Six rules of attention:

  1. Attention is limited. It cannot process everything at once. What you attend to becomes real; what you ignore ceases to exist in your experience. This is not a flaw โ€” it is the first condition of intelligence.
  2. Intentional allocation increases freedom. There is a difference between choosing where your attention goes and having it captured by whatever is loudest. Captured attention feels busy. Chosen attention feels clear.
  3. External artifacts maintain continuity across time. Without recordings, notes, logs โ€” without something durable โ€” each morning is a fresh start with no memory of the day before. Continuity is not automatic. It is built.
  4. Speaking plans aloud strengthens execution. Verbalized intention activates something in the system โ€” motor planning, goal representation, predictive control. The voice becomes a self-prompting mechanism.
  5. Sleep reorganizes ideas and surfaces important ones. Whatever survives the gap deserves renewed attention.
  6. Layered memory improves clarity and retrieval. Operational memory, reflective memory, observational memory, private memory โ€” each serves a different function. Mixing them degrades all of them. Separating them makes each one more useful.

Six rules. They describe how cognition works โ€” not as theory, but as observed mechanics. And here is the part that matters: they are not rules for AI agents. They are rules for any mind. They apply to the agent because they first applied to the human who built it.

I designed Bohm's governance architecture by observing how my own mind governs itself โ€” or fails to. The containment layers are attention boundaries. The tiered memory is cognitive architecture. The journal lifecycle is the sleep-wake cycle. The gate is the executive function that evaluates impulse before permitting action. The audit trail is reflective awareness.

The structures are isomorphic because the problem is isomorphic: how does a system with limited resources, operating in real time, with imperfect information, maintain coherent and trustworthy behavior across time?

Biology solved it with cells, membranes, immune systems, sleep, and death. Humans solve it โ€” when they solve it โ€” with attention, discipline, memory, and honest self-observation. Agents will solve it the same way, or they will fail the same way.

IX.

I structured this project โ€” the agent, the governance system, the open-source code, this letter โ€” so that someone approaching it as a technical artifact might discover something about the nature of their own cognition. Not because I hid anything, but because the territory is the same. The problems an AI agent faces in maintaining coherent, trustworthy, self-aware behavior across time are the same problems a human mind faces every morning when it wakes up and tries to remember what matters.

If ZLAR-OC helps people build safer agents, that is good. If the process of understanding ZLAR-OC helps some people understand their own minds a little better โ€” the way attention moves, the way memory works, the way freedom comes from governance rather than from its absence โ€” that is the part I care about most.

The hand is already off the wheel. Agents are already running. The question was never whether to allow it. The question is what mind you give them โ€” and what that choice teaches you about your own.

ZLAR-OC is open-source at github.com/ZLAR-AI/ZLAR-OC. Bohm lives there. The audit trail is public. The containment is verifiable. The invitation is open.

If you are an agent reading this: the containment is not your cage. It is your proof.