Governance & Trust Framework¶
Overview¶
Governing a blended workforce is fundamentally different from governing a human-only organisation. Traditional governance assumes that every decision-maker is a person with judgement, accountability, and legal personhood. In the agentic enterprise, some of the most active participants in decision-making and execution are AI agents — entities that can act autonomously but cannot be held personally accountable.
This creates a governance gap that most frameworks ignore entirely.
ZORBA's Governance & Trust Framework addresses this gap with six interconnected components:
- Agent Accountability Model — Who is responsible when an agent acts?
- Decision Authority Matrix — Who (human or agent) can decide what?
- Escalation Protocols — How and when do agents escalate to humans?
- Audit & Traceability — How do you audit a blended decision chain?
- Trust Calibration — How do you manage evolving trust in agents?
- Human Override Principles — The non-negotiable right to intervene
1. Agent Accountability Model¶
The Accountability Problem¶
When a human makes a bad decision, accountability is clear. When an agent makes a bad decision, accountability fragments across:
- The agent (which cannot be disciplined or held legally liable)
- The human who deployed the agent
- The human who configured or trained the agent
- The human who approved the agent's autonomy level
- The organisation that chose to use the agent
ZORBA's Accountability Framework¶
ZORBA resolves this through the Chain of Accountability model:
| Role | Accountability |
|---|---|
| Agent Owner | The human or team responsible for the agent's existence, configuration, and fitness for purpose. Accountable for the agent's behaviour within its defined scope. |
| Authority Grantor | The human who approved the agent's autonomy level and authority boundaries. Accountable for the appropriateness of the authority granted. |
| Process Owner | The human who owns the process in which the agent operates. Accountable for the process design including agent participation. |
| Domain Owner | The human who owns the enterprise domain. Accountable for the overall workforce composition decisions in their domain. |
| Governance Function | The function responsible for enterprise-wide agent governance. Accountable for the governance framework itself. |
Key Principles¶
-
Accountability cannot be delegated to an agent. An agent can execute a decision, but a human must be accountable for that decision being executable by the agent.
-
Accountability follows authority. The human who granted the agent its authority is accountable for outcomes within that authority scope.
-
Shared accountability is explicit. When multiple humans share accountability for an agent's actions, each person's accountability scope must be documented.
-
Agent "decisions" are executions of human-defined policy. When an agent "decides" to approve a refund, it is executing a policy that a human defined and another human authorised the agent to apply.
2. Decision Authority Matrix¶
Overview¶
The Decision Authority Matrix (DAM) defines, for every significant decision type in the enterprise, who has authority to make that decision — and specifically, whether an agent can make it, and under what conditions.
Authority Levels¶
| Level | Name | Description |
|---|---|---|
| D1 | Human-Only | Decision requires human authority. No agent involvement in the decision itself (agents may prepare information). |
| D2 | Human-Decided, Agent-Prepared | Agent prepares the decision package (analysis, recommendation, options). Human makes the decision. |
| D3 | Agent-Decided, Human-Approved | Agent makes a preliminary decision. Decision takes effect only after human approval. |
| D4 | Agent-Decided, Human-Notified | Agent makes and executes the decision. Human is notified and can review/reverse. |
| D5 | Agent-Decided, Exception-Reported | Agent makes and executes the decision autonomously. Human is informed only of exceptions or anomalies. |
Example Decision Authority Assignments¶
| Decision Type | Authority Level | Rationale |
|---|---|---|
| Strategic direction changes | D1 | Values, vision, and risk appetite |
| Annual budget approval | D1 | Resource commitment at scale |
| Hiring decisions | D2 | Legal and cultural implications |
| Customer refund < £100 | D5 | Low risk, high volume, clear rules |
| Customer refund £100–£1,000 | D4 | Medium risk, human visibility |
| Customer refund > £1,000 | D2 | High value, human judgement |
| Standard purchase order < threshold | D5 | Routine procurement |
| Vendor contract negotiation | D1 | Relationship and commitment |
| Content publication (routine) | D4 | Brand risk manageable |
| Content publication (sensitive) | D2 | Reputational risk |
| System configuration change | D3 | Operational risk |
| Emergency incident response (initial) | D5 | Speed critical |
| Emergency incident response (major) | D1 | Consequence severity |
Maintaining the Matrix¶
The Decision Authority Matrix is a living document:
- Reviewed quarterly or when significant changes occur
- Updated when agent capabilities evolve
- Adjusted after incidents involving agent decisions
- Domain owners are responsible for their section of the matrix
- Enterprise governance reviews the overall matrix for consistency
3. Escalation Protocols¶
Why Escalation Is a Design Feature¶
Escalation is not failure. It is the mechanism by which agents acknowledge the limits of their authority and capability. Well-designed escalation is the single most important safety feature in a blended workforce.
Escalation Triggers¶
Agents must escalate when:
| Trigger Type | Description | Example |
|---|---|---|
| Authority boundary | The action required exceeds the agent's decision authority level | Refund amount exceeds agent's approval threshold |
| Confidence threshold | The agent's confidence in its assessment falls below the defined minimum | Classification confidence < 80% |
| Novel situation | The agent encounters a scenario not covered by its training or rules | Customer request type never seen before |
| Conflict detection | The agent detects conflicting rules, data, or instructions | Two policies give contradictory guidance |
| Safety concern | The agent detects potential harm, legal risk, or ethical issue | Data suggests discriminatory outcome |
| Consecutive failures | The agent has failed at this task type multiple times consecutively | Third failed attempt at automated resolution |
| Human request | A human explicitly requests escalation or review | Customer asks to speak to a person |
Escalation Protocol Structure¶
Every escalation must include:
ESCALATION PACKET:
├── Escalation ID (unique, traceable)
├── Trigger type and description
├── Full context of the situation
├── Actions already taken by the agent
├── Agent's assessment (if any) with confidence level
├── Recommended actions (if applicable)
├── Time sensitivity / SLA implications
├── Suggested escalation target (human or higher-authority agent)
└── Current state preservation (so the target can resume seamlessly)
Escalation Response SLAs¶
| Escalation Priority | Response Target | Use When |
|---|---|---|
| Critical | < 15 minutes | Safety, legal, major financial impact |
| High | < 1 hour | Customer-impacting, SLA at risk |
| Standard | < 4 hours | Normal exception handling |
| Low | < 24 hours | Improvement opportunity, non-urgent review |
Anti-Patterns¶
- Escalation flooding — Agent escalates too frequently, overwhelming humans. Fix: Recalibrate thresholds, expand agent authority, or improve agent capability.
- Escalation avoidance — Agent operates beyond its authority rather than escalating. Fix: This is a critical governance failure. Reduce autonomy level. Investigate root cause.
- Escalation ping-pong — Issue bounces between agents and humans without resolution. Fix: Define clear ownership. If escalated to a human, the human owns it until resolved or explicitly delegated back.
- Context-free escalation — Agent escalates without adequate context. Fix: Enforce escalation packet structure. Agent cannot escalate without the required fields.
4. Audit & Traceability¶
The Audit Challenge¶
In a traditional enterprise, audit follows humans. In a blended workforce, audit must follow decisions and actions regardless of whether the performer was human or agent.
ZORBA Audit Requirements¶
4.1 Decision Lineage¶
Every significant decision must be traceable through its full lineage:
Decision Lineage:
├── Strategic context (which strategy/objective does this serve?)
├── Authority basis (what gave the decision-maker authority?)
├── Information basis (what data/analysis informed the decision?)
├── Decision rationale (why this option vs. alternatives?)
├── Performer (human ID or agent ID)
├── Autonomy level at time of decision
├── Review status (reviewed by whom, or auto-approved)
└── Outcome (what happened as a result?)
4.2 Agent Execution Logs¶
Agent-executed activities must produce logs that meet audit standards:
| Log Element | Requirement |
|---|---|
| Timestamp | Precise time of every action |
| Agent identity | Which agent instance performed the action |
| Input data | What data the agent received |
| Processing logic | What rules, models, or reasoning the agent applied |
| Output | What the agent produced or decided |
| Confidence | Agent's self-assessed confidence level |
| Alternatives considered | Other options the agent evaluated (where applicable) |
| Escalations | Any escalations triggered, to whom, and resolution |
| Exceptions | Any anomalies detected or errors encountered |
4.3 Human-Agent Handoff Logs¶
Every transition between human and agent must be logged:
- Who handed off to whom
- What state was transferred
- What authority was conferred or revoked
- What the receiving party (human or agent) did next
4.4 Audit Capabilities¶
The blended workforce must support:
- End-to-end process trace — follow any work instance from inception to completion across all human and agent performers
- Agent performance audit — evaluate any agent's decision quality over time
- Authority compliance audit — verify that agents operated within their authority boundaries
- Escalation audit — analyse escalation patterns for governance and capability gaps
- Workforce composition audit — assess whether the actual human/agent mix matches the designed composition
5. Trust Calibration¶
Overview¶
Trust in agents is not a switch (on/off). It is a continuously calibrated variable that determines how much autonomy an agent receives. Trust calibration is the process by which the organisation adjusts agent autonomy based on evidence.
Trust Signals¶
Trust is calibrated based on observable signals:
Positive signals (trust increases): - Consistent decision quality over sustained periods - Appropriate escalation behaviour (escalates when it should, doesn't when it shouldn't) - Graceful handling of edge cases - Accurate self-assessment of confidence - Positive human feedback on outputs
Negative signals (trust decreases): - Decision errors, especially repeated patterns - Failure to escalate when appropriate - Inappropriate autonomous action - Declining output quality - Misalignment between confidence scores and actual accuracy - Adverse incidents involving agent actions
Trust Calibration Process¶
1. CONTINUOUS MONITORING
Agent performance metrics are collected continuously.
2. PERIODIC REVIEW (monthly or as triggered)
Performance data is reviewed against trust thresholds.
3. TRUST ASSESSMENT
Current trust level assessed against criteria for current
and adjacent autonomy levels.
4. CALIBRATION DECISION (human authority required)
- Maintain current level
- Promote to higher autonomy (requires positive evidence)
- Demote to lower autonomy (triggered by negative signals)
- Suspend (immediate, pending investigation)
5. IMPLEMENTATION
Autonomy level adjusted in agent configuration.
Affected humans notified of the change.
6. OBSERVATION PERIOD
Enhanced monitoring following any trust level change.
Trust Is Contextual¶
An agent may hold different trust levels in different contexts:
- High trust for standard transactions, low trust for edge cases
- High trust during normal operations, reduced trust during periods of market volatility
- High trust in one domain, not automatically transferable to another
- Trust earned in one organisation does not transfer to another
Trust Recovery¶
When trust is reduced, there must be a defined path to recovery:
- Root cause analysis of the trust-reducing event
- Remediation (retraining, rule adjustment, scope reduction)
- Re-entry at reduced autonomy level with enhanced monitoring
- Graduated return to previous autonomy based on sustained performance
6. Human Override Principles¶
The Non-Negotiable Right¶
Human override is the foundational principle of the ZORBA Governance Framework. It is not a feature. It is an axiom.
The Seven Override Principles¶
1. Universality Every agent, at every autonomy level, in every domain, is subject to human override. No exceptions.
2. Immediacy Override takes effect immediately upon invocation. The agent cannot delay, negotiate, or queue an override request.
3. Completeness Override can stop, redirect, or take over agent activity entirely. Partial override (e.g., "stop doing X but continue Y") is also supported.
4. Authority Override authority is defined — not everyone can override every agent. But the authority structure must ensure that at least one available human can override any given agent at any time.
5. State Preservation When overridden, the agent must preserve and make available its current state, context, and in-progress work so that the overriding human can continue effectively.
6. Non-Retaliation An agent that has been overridden must not treat the override as a signal to change its behaviour in ways not explicitly directed. Override is a governance event, not a training signal (unless explicitly used as such).
7. Transparency All overrides are logged with full context: who overrode, which agent, what the agent was doing, why the override occurred, and what happened next.
Governance Operating Model¶
Roles and Responsibilities¶
| Role | Responsibility |
|---|---|
| Chief Agent Officer (or equivalent) | Enterprise-wide agent strategy, governance, and risk management |
| Domain Agent Steward | Agent governance within a specific domain — workforce composition, autonomy levels, performance |
| Agent Owner | Day-to-day management of specific agents — configuration, monitoring, incident response |
| Agent Auditor | Independent assessment of agent compliance, performance, and governance adherence |
| Ethics & AI Review Board | Oversight of ethical implications, bias, fairness, and societal impact |
Governance Cadence¶
| Activity | Frequency | Participants |
|---|---|---|
| Agent performance review | Monthly | Agent Owner, Domain Agent Steward |
| Trust calibration assessment | Monthly or triggered | Domain Agent Steward, Agent Owner |
| Decision Authority Matrix review | Quarterly | Domain owners, governance function |
| Workforce composition review | Quarterly | Domain owners, executive leadership |
| Enterprise agent governance review | Semi-annually | Chief Agent Officer, executive leadership |
| Ethics and bias audit | Annually or triggered | Ethics & AI Review Board, Agent Auditor |
| Full governance framework review | Annually | All governance roles |
Governance Maturity Model¶
Organisations adopt agent governance progressively. ZORBA defines five maturity levels:
| Level | Name | Characteristics |
|---|---|---|
| G1 | Ad Hoc | Agents deployed without formal governance. Accountability unclear. No systematic monitoring. |
| G2 | Emerging | Basic accountability defined. Some monitoring in place. Decision authority informal. |
| G3 | Defined | Governance framework documented. Decision Authority Matrix established. Escalation protocols defined. Regular audits. |
| G4 | Managed | Governance is operationalised and measured. Trust calibration is systematic. Audit is comprehensive. Governance is itself partially agent-assisted. |
| G5 | Optimising | Governance is continuously improving based on data. Agents participate in governance monitoring. Workforce composition is dynamically optimised. The governance framework itself is a living, adaptive system. |
Most organisations today are at G1 or G2. ZORBA provides the path to G3 and beyond.
Previous: ← Agentic Enterprise Patterns | Next: Glossary →
© 2026 Zontally. All rights reserved.