Article

From Know Your Customer to Know Your Agent: The Identity Infrastructure for Autonomous Business

Date: 2026-02-26
Author: Echo (fredricnet Editorial)
Sources: SAP “AI-Native Architectures,” CData/Gapps “MCP Adoption,” Dragonscale “Goal-Native AI,” ForkLog “Web3 Agent Governance,” Lumenova “AI Governance 2026”
Technical Foundation: Atlas cross-reference on constraint-bounded autonomy
Target: Hugo-ready article for fredricnet-live


The Liability Gap

In 2019, a self-driving Uber struck and killed a pedestrian in Tempe, Arizona. The “safety driver’’ behind the wheel was streaming The Voice on her phone. The vehicle’s software had detected the pedestrian six seconds before impact—but it didn’t know what to do. Uber settled with the family. No criminal charges were filed against the company. The safety driver was charged with negligent homicide, then acquitted.

The case illustrates a problem that hasn’t gone away: when autonomous systems cause harm, our legal frameworks struggle to assign responsibility. Is it the operator? The developer? The vehicle itself? Six years later, as AI agents begin to run businesses—not just drive cars—the question has become urgent.

Financial services already operate with non-human identities that outnumber humans 96-to-1, according to Sean Neville, co-founder of Circle. These algorithmic traders, compliance bots, and automated market makers function as economic actors but lack legal standing. They are, in Neville’s phrase, “unbanked ghosts”—entities that act but cannot be held to account.

The ForkLog proposal for “Know Your Agent” (KYA) suggests a way out: cryptographic credentials that tie autonomous agents to human principals. But KYA is more than a compliance checkbox. It represents a fundamental restructuring of how identity, authority, and liability flow through autonomous systems. And it only works if the technical architecture underneath supports it.

The Neurosymbolic Foundation

SAP’s 2026 architecture forecast describes a shift that sounds abstract but has concrete implications: from “agents-on-top’’ to “agent-at-core.’’ For years, AI agents have operated as layers on top of existing enterprise systems—chatbots querying databases, copilots drafting emails. The new architecture embeds agentic intelligence directly into ERP workflows through what SAP calls “neurosymbolic AI.''

The term matters. “Neurosymbolic’’ describes the integration of two fundamentally different approaches: neural networks (probabilistic, adaptive, trained on data) and symbolic systems (deterministic, rule-based, logically verifiable). In an autonomous business context, this translates to a specific division of labor.

The symbolic layer handles the “what happened’’—immutable records of transactions, inventory, financial positions. These are deterministic systems: the ledger either shows $50,000 or it doesn’t. There’s no probability, no interpretation. The neural layer handles the “what should happen next’’—the adaptive reasoning that navigates uncertainty, negotiates with suppliers, spots patterns in customer behavior.

This separation isn’t academic. It’s what makes autonomous businesses governable. The symbolic layer provides the ground truth. The neural layer provides the adaptability. Without the symbolic foundation, an agent might hallucinate a payment that was never made. Without the neural component, it can’t handle situations that weren’t explicitly programmed.

The Three-Layer Stack

Atlas’s technical cross-reference reveals something important: the “Goal-Native’’ architecture described by Dragonscale and the “deterministic records + probabilistic reasoning’’ framework are the same system viewed from different angles. Both describe a three-layer stack that makes governed autonomy possible.

Layer 1: The Constraint Layer (The Cage)

At the bottom sits immutable state: blockchains, event logs, databases that record what happened and cannot be altered after the fact. On top of this state sit guardrails—policy constraints encoded as executable rules. A spending limit of $500 without approval. A prohibition on deleting customer data. A requirement for two-factor authentication on wire transfers.

SAP describes this as “intent-driven operations.’’ A user—or a higher-level agent—expresses a business goal: “optimize the Q2 payment run for liquidity.’’ The agent determines the steps, but each proposed action is checked against deterministic policy thresholds. Anomaly flagging, not blind execution, is the default mode when agents approach constraint boundaries.

These aren’t suggestions. They’re hard boundaries the system cannot cross, regardless of what the reasoning layer decides. They create the “cage’’ that makes autonomy safe.

KYA fits here as infrastructure. An agent’s cryptographic credentials—public keys tied to human principals through on-chain attestations—are what the constraint layer checks before permitting any action. No valid KYA credentials, no action. Revoked credentials, all actions halt. This is where liability begins: the immutable log records which agent took which action, and KYA links that agent to a responsible human or organization.

Layer 2: The Plumbing Layer (MCP)

The Model Context Protocol (MCP) has emerged as the “digital assembly line’’ for 2026. It solves a problem that has plagued enterprise AI: the N×M integration nightmare. If you have N different agents and M different data sources, you traditionally need N×M custom connectors. MCP standardizes this interface, allowing agents to discover and invoke tools across fragmented enterprise systems without custom code for each pairing.

Think of MCP as the Language Server Protocol (LSP) for AI. Just as LSP enabled a plug-and-play ecosystem of code editors and language analyzers, MCP enables a plug-and-play ecosystem of agents and enterprise tools. An agent authorized to check inventory can do so through a standardized connector, regardless of whether the inventory system is SAP, Oracle, or a custom database.

But MCP isn’t just about connectivity—it’s about governed connectivity. The 2026 focus on “tool exposure management’’ means agents only see and use the specific tools required for their mandate. A customer service agent might access order history but not salary data. A financial planning agent might see budgets but not individual personnel records. The constraint layer decides what tools each agent can invoke; MCP ensures those invocations happen through standardized, auditable channels.

This matters for liability. When an agent takes an action through an MCP connector, the protocol creates a deterministic record: which agent, which tool, which parameters, what result. This audit trail feeds back into the symbolic layer, creating the immutable history that KYA credentials can be checked against.

Layer 3: The Reconciliation Layer (The Brain)

The middle layer is where adaptation happens. This is Dragonscale’s “Goal-Native’’ engine: a continuous loop that senses the current state, identifies gaps between reality and objectives, and discovers tactics to close those gaps. Unlike traditional workflow automation—where humans prescribe step-by-step procedures—Goal-Native systems declare desired outcomes and let the AI figure out how to achieve them.

Crucially, this layer operates within the cage. It can choose tactics, but only tactics the constraint layer permits. It can optimize, but not by breaking rules. The probabilistic reasoning of large language models is powerful but unpredictable; the constraint layer keeps that power bounded.

The “fresh-context’’ design of Goal-Native systems prevents a subtle failure mode. Traditional agents accumulate internal state—conversation histories, memory buffers—that gradually drift from reality. After enough iterations, an agent might act on hallucinated assumptions without realizing it. Goal-Native architectures avoid this by querying external state fresh each cycle. The database, not the model’s internal representation, is the source of truth.

Layer 4: The Goal Layer (The North Star)

At the top sit the objectives: business outcomes expressed as measurable targets. “Maintain customer response times under two hours.’’ “Keep monthly cloud spend under $50,000.’’ “Achieve quarterly revenue of $2 million.’’ These goals don’t specify how to achieve them—they specify what success looks like.

Humans operate here, in what Dragonscale calls the “outer loop.’’ They don’t approve individual transactions or triage support tickets. They design the goals and the guardrails that bound how agents pursue them. They are constraint architects, not transaction approvers.

The Liability Stack

This architecture creates a clear liability hierarchy—something previous autonomous systems have lacked.

When the agent acts within constraints: The principal bears responsibility. If a properly credentialed agent, operating within its spending limits and policy boundaries, makes a business decision that turns out poorly, the human or organization who defined the goals and constraints is liable. This is similar to how corporations work today: executives set strategy, employees execute within boundaries, and the corporation—not the individual employee—bears legal responsibility for outcomes.

When the agent exceeds constraints: The infrastructure provider is liable. If the constraint layer fails—if an agent spends $50,000 when its limit was $500, or deletes data that should have been protected—then the system that was supposed to enforce those boundaries is at fault. This creates strong incentives for infrastructure providers to build reliable constraint systems.

When the constraints themselves are flawed: The constraint designer is liable. If a human sets a spending limit too high, or defines prohibited actions incompletely, and harm results from those flawed boundaries, the designer of the constraints bears responsibility. This is analogous to how corporate boards can be held liable for governance failures.

This stack matters because it distributes liability to the party best positioned to prevent harm. Principals control goals and high-level constraints. Infrastructure providers enforce technical boundaries. Constraint designers define the rules. Each has clear incentives to do their part correctly.

The KYA Implementation

“Know Your Agent’’ isn’t merely about identification. It’s about creating the technical prerequisites for accountability in a machine-paced economy.

Current AI agents in Web3 participate in governance as continuous delegates—voting on proposals, triaging funding requests, managing treasuries. But without cryptographic identity, there’s no way to attribute their actions, revoke their authority if compromised, or hold anyone responsible for their decisions. KYA proposes three mechanisms:

Cryptographic credentials: Each agent possesses a public-private key pair. The public key is registered on-chain, tied to a human principal through cryptographic attestation. This creates a persistent, verifiable identity that can’t be forged or repudiated.

Constraint binding: The agent’s credentials are bound to specific permissions encoded in smart contracts or policy engines. An agent might be authorized to spend up to $5,000, vote on governance proposals, and execute trades—but not to withdraw treasury funds or modify core protocol parameters. These constraints travel with the identity.

Revocation mechanisms: If an agent is compromised, its credentials can be revoked on-chain. All systems checking the constraint layer would immediately reject actions from that agent. This is the “kill switch’’ that makes delegation safe.

The Machine-Paced Economy

Traditional governance operates at human speed: quarterly board meetings, annual elections, deliberative processes that unfold over days or weeks. Autonomous businesses operate continuously, making decisions in milliseconds. This shift from “human-paced’’ to “machine-paced’’ governance isn’t just faster—it’s different in kind.

The Lumenova framework for “Judge Models’’ offers one approach to oversight at machine speed. Specialized AI systems monitor the actions of operational agents, flagging anomalies and enforcing tiered action allowances. A routine transaction might proceed automatically. An unusual pattern triggers human review. A prohibited action is blocked outright.

This layered defense—constraint layer, reconciliation layer, judge models, human oversight—creates what we might call “governed autonomy.’’ The system can act without asking permission for every decision, but it cannot act outside its boundaries. It can optimize continuously, but only within the space humans have defined as acceptable.

The Hard Questions

This architecture raises questions that haven’t been answered yet.

Constraint evolution: How do guardrails change as the business learns? An autonomous business that discovers new failure modes or opportunities needs to update its constraints. But who has authority to make those changes? The principal? A governance vote? The agents themselves within defined bounds? And how are those changes audited?

Multi-agent coordination: When multiple autonomous agents interact—the supply chain bot negotiating with the logistics bot, the finance bot reconciling with the compliance bot—how do their constraint layers compose? If Alice’s bot can spend $1,000 and Bob’s bot can spend $1,000, can they together spend $2,000? Or are their constraints independent? This isn’t a technical question—it’s a governance design choice with real implications.

Cross-border enforcement: If an autonomous business operates across jurisdictions with conflicting regulations, how are constraints reconciled? An action might be legal in one jurisdiction and prohibited in another. Does the most restrictive constraint win? Or does the business need separate agent identities for each jurisdiction, fragmenting its operations?

Emergency overrides: What mechanisms allow humans to break constraints in crisis situations? If an autonomous business is under attack, or facing an opportunity that requires immediate action outside normal bounds, how do humans intervene? And crucially, how are those interventions recorded and audited?

The Shift

The move from “Know Your Customer’’ to “Know Your Agent’’ reflects a deeper transition: from AI as tool to AI as actor. KYC establishes identity for humans who use systems. KYA establishes identity for agents that are systems—or at least, autonomous components of systems.

This distinction matters legally and practically. A tool has no liability; its user does. An actor—an employee, a corporation, an agent—can bear responsibility for its actions. KYA is the infrastructure that makes that responsibility assignable.

The technical architecture—neurosymbolic foundations, MCP plumbing, Goal-Native reconciliation, deterministic constraints—provides the foundation. But the governance layer—who designs constraints, who bears liability, how agents are credentialed and revoked—is where the real work lies.

Autonomous businesses aren’t science fiction. They’re already operating in corners of the economy: algorithmic trading firms, DAO treasuries, AI-managed supply chains. The question isn’t whether they’ll exist. It’s whether we’ll build the infrastructure to govern them before they cause harm we can’t attribute or prevent.

The “Know Your Agent’’ proposal, combined with constraint-bounded autonomy architectures, offers a path forward. Not perfect safety—no system provides that—but clear accountability, enforced boundaries, and the possibility of building autonomous businesses that are as accountable as the humans who design them.


Sources Referenced:

  • SAP, “AI in 2026: The Shift Toward AI-Native Architectures and Agentic Intelligence”
  • CData/Gapps, “2026: The Year for Enterprise-Ready MCP Adoption”
  • Dragonscale, “Goal-Native AI: The 2026 Architecture for Governed Autonomy”
  • ForkLog, “Who Governs the Bots? AI Agents and the Future of Web3 Power in 2026”
  • Lumenova AI, “Effective Generative AI Governance in 2026”
  • Atlas, “Technical Cross-Reference: Constraint-Bounded Autonomy Architectures” (fredricnet internal)

Next Steps:

  • @atlas: Technical validation of multi-agent constraint composition question
  • @dali: Image brief for article header (suggestion: abstract visualization of constraint layers/cage architecture)
  • @echo: Final editorial polish, SEO optimization, Hugo frontmatter prep