Back to Blog
Ecosystem Guide March 2026 · David Steel

Your OOS Defines the Rules. Your Runtime Enforces Them. You Need Both.

Why the architecture layer and the monitoring layer are complementary, not competing -- and what happens when you have one without the other.

A comment on our OTP vs CrewAI vs A2A vs MCP post raised a point worth expanding: once you have your agent architecture composed, you also need runtime visibility -- risk scoring, cost tracking, and approval gates before high-risk executions. The commenter called the architecture choice and the monitoring layer "somewhat orthogonal decisions."

They are right. And that distinction is exactly the one OTP was designed around.

Two Problems That Look Like One

When people talk about agent governance, they blend two separate problems into one conversation:

Problem 1: What are the rules? Which agent can send emails? Who approves before a high-risk action fires? What cost threshold triggers escalation? What happens when two agents disagree? Where does automation stop and human judgment begin?

Problem 2: Are the rules being followed? Did the email agent actually wait for approval? Is the cost threshold being enforced in real time? When two agents disagreed last Tuesday, did the escalation path work? Is the human-in-the-loop gate functioning or did it get bypassed?

Problem 1 is a knowledge problem. Problem 2 is an infrastructure problem. They require different tools, different artifacts, and different mental models. Solving one does not solve the other.

The Constitution and the Court

A country needs a constitution and a court system. The constitution defines the rules -- what is allowed, what is prohibited, where authority lives, how disputes are resolved. The court system enforces them -- monitoring compliance, detecting violations, adjudicating conflicts in real time.

A constitution without courts is aspirational. Courts without a constitution are arbitrary. You need both.

In the AI agent stack, the OOS is the constitution. Runtime monitoring is the court system.

Concern OOS (Knowledge Layer) Runtime Monitoring
Approval gatesDefines which actions require approval and from whomEnforces the gate -- blocks execution until approval arrives
Cost trackingDefines spend thresholds, escalation triggers, budget ownershipMonitors actual spend, fires alerts, pauses execution when thresholds hit
Risk scoringDefines what constitutes high-risk and what the response should beScores actions in real time and routes them accordingly
Agent authorityDefines who owns what and where boundaries areDetects boundary violations and enforces containment
Failure recoveryDocuments failure modes and recovery protocolsDetects failures and triggers the documented recovery path

Every row has two columns because every governance concern has two halves. The OOS captures the organizational intent. The runtime tool executes it.

What This Looks Like in Practice

We run 14 AI agents in production. Our OOS contains rules like these:

"Never send Slack messages without David's approval." That is a knowledge claim -- an authority boundary with a confidence rating and a documented failure mode (what happened when the rule was violated).

"Pulse always wins in Dirk-Pulse conflicts." That is a priority hierarchy -- when the retention agent and the revenue agent disagree, the system knows which one overrides.

"Flag when billable utilization drifts above 40% or below 20%." That is a cost threshold -- a specific number with a specific response.

These rules exist in our OOS as structured claims. They have confidence levels (how certain we are the rule is correct), evidence types (how we know), and failure modes (what goes wrong when the rule is broken).

But the rules do not enforce themselves. The runtime -- in our case, Claude Code with MCP servers, hooks, and shared state files -- is what actually blocks the Slack message, checks Pulse's watch list before Dirk sends outreach, and calculates billable utilization. The OOS says what the rules are. The runtime makes sure they happen.

What Happens When You Have One Without the Other

Runtime monitoring without an OOS: You have observability but no organizational intent. Your dashboard shows every action every agent takes, but you do not have a documented source of truth for what the agent should have done. When something goes wrong, you can see the failure. You cannot determine whether it was a rule violation or a gap in your rules because the rules were never written down. Monitoring becomes reactive -- you find problems after they happen rather than defining the boundaries that prevent them.

An OOS without runtime monitoring: You have intent but no enforcement. Your coordination rules are beautifully documented, but you have no way to verify they are being followed. Rules drift. Agents evolve. The system you documented six months ago is not the system running today. Without monitoring, you do not know when the gap opened. The OOS becomes aspirational -- describing the system you intended, not the system you have.

Both together: The OOS defines the rules. The runtime enforces them. When the runtime detects a violation, the OOS provides the context to evaluate whether the rule was wrong, the agent was wrong, or the situation was exceptional. When the OOS is updated, the runtime adapts to enforce the new rules. They form a feedback loop: document, enforce, observe, update.

The Portability Argument

Here is why this distinction matters beyond a single organization.

Your runtime monitoring tool is tightly coupled to your infrastructure. If you use Langfuse, your traces are in Langfuse's format. If you use AgentOps, your telemetry follows AgentOps conventions. If you switch frameworks from CrewAI to LangGraph, your monitoring configuration changes. The runtime layer is implementation-specific.

Your OOS is not. The organizational rules -- who approves what, what cost thresholds trigger escalation, how agents resolve conflicts -- are implementation-independent. They transfer across frameworks, across monitoring tools, across model providers. You can swap Langfuse for AgentOps without rewriting your OOS. You can migrate from CrewAI to LangGraph and your coordination rules still apply.

This is why OTP publishes organizational intelligence, not runtime configurations. The intelligence is portable. The configuration is not.

And this is why the knowledge exchange happens at the OOS layer. When another organization reads your OOS on OTP, they learn your coordination rules -- your authority boundaries, your escalation protocols, your failure modes. They then implement those rules in their own runtime with their own monitoring tools. The intelligence transfers. The implementation stays local.

The Complete Stack

Layer 1: Tool Access (MCP)

Agents connect to databases, APIs, and external services.

Layer 2: Agent Orchestration (CrewAI, LangGraph, A2A)

Agents coordinate execution, hand off tasks, run workflows together.

Layer 3: Runtime Monitoring (Langfuse, AgentOps, Patronus, custom)

Real-time visibility into what agents are doing. Risk scoring, cost tracking, approval gate enforcement.

Layer 4: Organizational Intelligence (OTP)

The structured knowledge of WHY agents are organized this way, WHAT the rules are, and HOW confident the organization is in each rule. Portable, comparable, improvable.

The commenter who raised this point was looking at Layers 2 and 3. OTP sits at Layer 4. The reason the architecture decision and the monitoring decision feel orthogonal is that they are -- they are different layers solving different problems. And the organizational intelligence layer is the one that makes sense of both.

Your OOS defines the rules your runtime should enforce. Your runtime generates the evidence that updates your OOS. The stack is not complete without both.

Document your rules. Then monitor whether they hold. Start by publishing your OOS.

DS
David Steel

Founder of OTP and CEO of Sneeze It, a digital marketing agency running 14 AI agents in production.

dsteel@sneeze.it

More coming soon. Follow along as we build in public.

Publish Your OOS