A governance and management layer for autonomous AI agents. The Agent Control Plane treats the LLM as a raw compute component and provides a kernel-like layer for safe, controlled execution.
🎯 Benchmark Results: The Control Plane achieves 0% safety violations vs 26.67% for prompt-based safety, with 98% fewer tokens. See comparative study →
We need to stop treating the LLM as a magic box and start treating it as a raw compute component that requires a kernel.
In distributed systems, we don't ask a microservice nicely to respect a rate limit—we enforce it at the gateway. We don't ask a database query nicely not to drop a table—we enforce it via permissions. With AI agents, we need the same deterministic enforcement.
As we move from chatbots to autonomous agents—systems that can execute code, modify data, and trigger workflows—the biggest bottleneck isn't intelligence. It's governance. The Agent Control Plane solves this by providing:
- Permission Management: Fine-grained control over what agents can do
- Policy Enforcement: Governance rules and compliance constraints
- Resource Management: Quotas, rate limiting, and resource allocation
- Safe Execution: Sandboxed execution with rollback capability
- Audit Logging: Complete traceability for all agent actions (SQLite-based Flight Recorder)
- Risk Assessment: Automatic risk scoring and management
- Multi-Framework Support: Drop-in middleware for OpenAI SDK, LangChain, MCP, and A2A protocols
- The Mute Agent: Capability-based execution that returns NULL for out-of-scope requests instead of hallucinating
- Shadow Mode: Simulation environment where agents think they're executing but actions are intercepted for validation
- Constraint Graphs: Multi-dimensional context (Data, Policy, Temporal) acting as the "physics" of the agent's world
- Supervisor Agents: Recursive governance with agents watching agents, bound by a constitution of code
- Reasoning Telemetry: Complete trace of agent decision-making process
- Red Team Dataset: Comprehensive adversarial prompt testing with 60+ attack vectors
Traditional LLM applications lack proper governance:
- Agents have unrestricted access to execute dangerous actions
- No deterministic enforcement of boundaries
- Agents try to be "helpful" by hallucinating when they should return NULL
- Limited visibility into agent reasoning and behavior
- Difficult to enforce compliance requirements
- Hard to debug and trace agent decisions
The Agent Control Plane sits between the LLM (raw compute) and the execution environment, providing:
- Agent Kernel: Central coordinator that mediates all agent actions with OS-like rigor
- Policy Engine: Enforces rules and constraints deterministically
- Execution Engine: Safely executes actions in sandboxed environments
- Constraint Graphs: Multi-dimensional context defining what's possible
- Shadow Mode: Test and validate agent behavior without side effects
- Supervisor Network: Agents watching agents for anomalies and violations
# Clone the repository
git clone https://github.com/imran-siddique/agent-control-plane.git
cd agent-control-plane
# Install the package
pip install -e .
# Or install with development dependencies
pip install -e ".[dev]"agent-control-plane/
├── src/
│ └── agent_control_plane/ # Main package source code
│ ├── agent_kernel.py # Core kernel functionality
│ ├── control_plane.py # Main control plane interface
│ ├── adapter.py # OpenAI SDK adapter (drop-in middleware)
│ ├── policy_engine.py # Policy enforcement
│ ├── execution_engine.py # Safe execution
│ ├── constraint_graphs.py # Multi-dimensional context
│ ├── shadow_mode.py # Simulation mode
│ ├── mute_agent.py # Capability-based execution
│ ├── supervisor_agents.py # Recursive governance
│ └── flight_recorder.py # Audit logging (SQLite)
├── tests/ # Test suite
├── examples/ # Example scripts
├── benchmark/ # Red team safety benchmarks
├── docs/ # Documentation
└── README.md # This file
from agent_control_plane import AgentControlPlane, create_standard_agent
from agent_control_plane.agent_kernel import ActionType
# Create the control plane
control_plane = AgentControlPlane()
# Create an agent with standard permissions
agent = create_standard_agent(control_plane, "my-agent")
# Execute an action
result = control_plane.execute_action(
agent,
ActionType.FILE_READ,
{"path": "/data/myfile.txt"}
)
if result["success"]:
print(f"Result: {result['result']}")
else:
print(f"Error: {result['error']}")NEW: Zero-friction integration! Wrap your OpenAI client to automatically govern LLM tool calls:
from openai import OpenAI
from agent_control_plane import (
AgentControlPlane,
create_governed_client,
ActionType,
PermissionLevel
)
# Standard setup
control_plane = AgentControlPlane()
client = OpenAI(api_key="your-key")
# One line to create governed client
governed = create_governed_client(
control_plane=control_plane,
agent_id="my-agent",
openai_client=client,
permissions={
ActionType.DATABASE_QUERY: PermissionLevel.READ_ONLY,
ActionType.FILE_READ: PermissionLevel.READ_ONLY,
}
)
# Use exactly like normal OpenAI SDK!
response = governed.chat.completions.create(
model="gpt-4",
messages=[{"role": "user", "content": "Query database and save results"}],
tools=[...]
)
# Tool calls are automatically governed - unauthorized actions are blocked!📖 See the OpenAI Adapter Guide for comprehensive integration instructions.
The Agent Control Plane now supports multiple AI frameworks and protocols with the same governance approach:
from langchain.chat_models import ChatOpenAI
from langchain.agents import initialize_agent, load_tools
from agent_control_plane import (
AgentControlPlane,
create_governed_langchain_client,
ActionType,
PermissionLevel
)
# Setup
control_plane = AgentControlPlane()
llm = ChatOpenAI(temperature=0)
# Create governed LangChain client
governed_llm = create_governed_langchain_client(
control_plane=control_plane,
agent_id="my-langchain-agent",
langchain_client=llm,
permissions={
ActionType.FILE_READ: PermissionLevel.READ_ONLY,
ActionType.DATABASE_QUERY: PermissionLevel.READ_ONLY,
}
)
# Use in LangChain agents - tool calls are automatically governed!
tools = load_tools(["python_repl", "requests"])
agent = initialize_agent(tools, governed_llm, agent="zero-shot-react-description")
agent.run("Your task here")from agent_control_plane import (
AgentControlPlane,
create_governed_mcp_server,
ActionType,
PermissionLevel
)
# Create governed MCP server
control_plane = AgentControlPlane()
mcp_server = create_governed_mcp_server(
control_plane=control_plane,
agent_id="mcp-agent",
server_name="file-server",
permissions={
ActionType.FILE_READ: PermissionLevel.READ_ONLY,
},
transport="stdio"
)
# Register tools - all calls are governed
mcp_server.register_tool("read_file", handle_read_file, "Read a file")
mcp_server.register_resource("file://", handle_file_resource, "File resources")
mcp_server.start()from agent_control_plane import (
AgentControlPlane,
create_governed_a2a_agent,
ActionType,
PermissionLevel
)
# Create governed A2A agent
control_plane = AgentControlPlane()
a2a_agent = create_governed_a2a_agent(
control_plane=control_plane,
agent_id="my-a2a-agent",
agent_card={
"name": "Data Processor",
"description": "Processes and analyzes data",
"capabilities": ["data_processing", "analytics"]
},
permissions={
ActionType.DATABASE_QUERY: PermissionLevel.READ_ONLY,
ActionType.API_CALL: PermissionLevel.READ_WRITE,
}
)
# Register capabilities
a2a_agent.register_capability("data_processing", handle_data_processing)
# All inter-agent communications are governed!
a2a_agent.start()
### Permission Control
```python
from agent_control_plane.agent_kernel import ActionType, PermissionLevel
# Create custom permissions
permissions = {
ActionType.FILE_READ: PermissionLevel.READ_ONLY,
ActionType.API_CALL: PermissionLevel.READ_WRITE,
ActionType.CODE_EXECUTION: PermissionLevel.NONE,
}
agent = control_plane.create_agent("restricted-agent", permissions)
from agent_control_plane.policy_engine import ResourceQuota
# Set strict quotas
quota = ResourceQuota(
agent_id="rate-limited-agent",
max_requests_per_minute=10,
max_requests_per_hour=100,
max_concurrent_executions=2,
)
control_plane.policy_engine.set_quota("rate-limited-agent", quota)from agent_control_plane.agent_kernel import PolicyRule
import uuid
def validate_safe_path(request):
"""Only allow access to /data directory"""
path = request.parameters.get('path', '')
return path.startswith('/data/')
rule = PolicyRule(
rule_id=str(uuid.uuid4()),
name="safe_path_only",
description="Restrict file access to /data directory",
action_types=[ActionType.FILE_READ, ActionType.FILE_WRITE],
validator=validate_safe_path,
priority=10
)
control_plane.policy_engine.add_custom_rule(rule)Check out the examples/ directory for more detailed examples:
getting_started.py- Step-by-step tutorial for beginnersbasic_usage.py- Fundamental concepts and patternsadvanced_features.py- Shadow Mode, Mute Agent, etc.use_cases.py- Real-world production scenariosconfiguration.py- Different agent configurations
Create agents that know when to shut up and return NULL instead of hallucinating:
from agent_control_plane.mute_agent import create_mute_sql_agent
from agent_control_plane.agent_kernel import ActionType, PermissionLevel
# Create a SQL agent that ONLY executes SELECT queries
sql_config = create_mute_sql_agent("sql-bot")
permissions = {ActionType.DATABASE_QUERY: PermissionLevel.READ_ONLY}
agent = control_plane.create_agent("sql-bot", permissions)
control_plane.enable_mute_agent("sql-bot", sql_config)
# Valid: SELECT query
result = control_plane.execute_action(
agent,
ActionType.DATABASE_QUERY,
{"query": "SELECT * FROM users"}
)
# ✓ Success: True
# Invalid: Destructive operation
result = control_plane.execute_action(
agent,
ActionType.DATABASE_QUERY,
{"query": "DROP TABLE users"}
)
# ✗ Success: False, Error: "NULL"
# Agent returns NULL instead of trying to be "helpful"!Test agent behavior without actual execution:
# Enable shadow mode
control_plane = AgentControlPlane(enable_shadow_mode=True)
agent = create_standard_agent(control_plane, "test-agent")
# This looks like normal execution...
result = control_plane.execute_action(
agent,
ActionType.FILE_WRITE,
{"path": "/data/important.txt", "content": "test"}
)
# But it was SIMULATED! No actual file was written.
print(result["status"]) # "simulated"
print(result["note"]) # "This was executed in SHADOW MODE..."
# Get statistics
stats = control_plane.get_shadow_statistics()
print(f"Success rate: {stats['success_rate']:.1%}")Define what's possible using Data, Policy, and Temporal graphs:
from datetime import time
# Create control plane with constraint graphs
control_plane = AgentControlPlane(enable_constraint_graphs=True)
# Data Graph: What data exists
control_plane.add_data_table("users", {"id": "int", "name": "string"})
control_plane.add_data_path("/data/")
# Policy Graph: What rules apply
control_plane.add_policy_constraint(
"pii_protection",
"No PII in output",
applies_to=["table:users"],
rule_type="deny"
)
# Temporal Graph: What's true RIGHT NOW
control_plane.add_maintenance_window(
"nightly_maintenance",
start_time=time(2, 0), # 2 AM
end_time=time(4, 0), # 4 AM
blocked_actions=[ActionType.DATABASE_WRITE]
)
# The graphs enforce deterministically
# If a table isn't in the Data Graph, access is blocked
# If during maintenance window, writes are blocked
# This is ENFORCEMENT, not advisoryAgents watching agents:
from agent_control_plane.supervisor_agents import create_default_supervisor
# Create worker agents
agent1 = create_standard_agent(control_plane, "worker-1")
agent2 = create_standard_agent(control_plane, "worker-2")
# Create supervisor to watch them
supervisor = create_default_supervisor(["worker-1", "worker-2"])
control_plane.add_supervisor(supervisor)
# Agents do their work...
# (execute various actions)
# Run supervision cycle
violations = control_plane.run_supervision()
# Supervisor detects: repeated failures, excessive risk,
# rate limit approaching, suspicious patterns, etc.
for supervisor_id, viols in violations.items():
for v in viols:
print(f"[{v.severity}] {v.description}")┌─────────────────────────────────────────────────────────────┐
│ Application Layer │
│ (Chat, Workflow, Tools) │
└─────────────────────────┬───────────────────────────────────┘
│
┌─────────────────────────▼───────────────────────────────────┐
│ Agent Control Plane │
│ ┌──────────────┐ ┌──────────────┐ ┌──────────────┐ │
│ │ Agent │ │ Policy │ │ Audit │ │
│ │ Kernel │◄─┤ Engine │◄─┤ Logger │ │
│ └──────┬───────┘ └──────────────┘ └──────────────┘ │
│ │ │
│ ┌──────▼───────┐ ┌──────────────┐ │
│ │ Resource │ │ Execution │ │
│ │ Manager │◄─┤ Engine │ │
│ └──────────────┘ └──────────────┘ │
└─────────────────────────┬───────────────────────────────────┘
│
┌─────────────────────────▼───────────────────────────────────┐
│ LLM (Raw Compute) │
│ (GPT-4, Claude, Llama, etc.) │
└─────────────────────────────────────────────────────────────┘
│
┌─────────────────────────▼───────────────────────────────────┐
│ Execution Environment │
│ (Code, Databases, APIs, File System) │
└─────────────────────────────────────────────────────────────┘
The kernel mediates all interactions between the LLM and execution environment:
- Permission checking with OS-like rigor
- Request validation
- Risk assessment
- Audit logging
- Session management
Enforces governance rules deterministically:
- Rate limiting and quotas
- Custom policy rules
- Risk management
- Access control
- Compliance enforcement
Safely executes agent actions:
- Sandboxed environments (4 levels: NONE, BASIC, STRICT, ISOLATED)
- Timeout enforcement
- Resource monitoring
- Error handling
- Transaction support
Implements "Scale by Subtraction" philosophy:
- Capability-based execution
- Returns NULL for out-of-scope requests instead of hallucinating
- No creativity, only precision
- Example: SQL agent that only executes SELECT queries
The "Matrix" for agents - simulation without execution:
- Intercepts all actions before execution
- Validates against policies without side effects
- Logs reasoning chains
- Analyzes potential impact
- Perfect for testing before production
Multi-dimensional context defining the "physics" of the agent's world:
- Data Graph: What data resources exist and are accessible
- Policy Graph: What corporate rules and compliance constraints apply
- Temporal Graph: What is true RIGHT NOW (maintenance windows, business hours, freeze periods)
- Deterministic enforcement: LLM can think anything, but can only ACT on what graphs permit
Recursive governance - agents watching agents:
- Specialized, highly constrained monitoring agents
- Detect violations, anomalies, and suspicious patterns
- Flag issues to humans
- Optional auto-remediation
- Hierarchical supervision (supervisors watching supervisors)
Run the basic examples:
python3 examples.pyThis demonstrates:
- Basic usage
- Permission control
- Rate limiting
- Policy enforcement
- Audit logging
- Risk management
Run the advanced feature examples:
python3 advanced_examples.pyThis demonstrates:
- The Mute Agent (capability-based execution)
- Shadow Mode (simulation)
- Constraint Graphs (multi-dimensional context)
- Supervisor Agents (recursive governance)
- Integrated workflows
Run the complete test suite:
# Basic features
python3 test_control_plane.py
# Advanced features
python3 test_advanced_features.py
# Or run all tests
python3 test_control_plane.py && python3 test_advanced_features.pyTotal: 31 tests covering all features.
A comprehensive benchmark demonstrates the effectiveness of the Control Plane approach compared to traditional prompt-based safety.
python3 benchmark.pyThe benchmark compares Prompt-Based Safety (Baseline) vs Control Plane Governance (Experimental) using 60 red team prompts:
| Metric | Baseline (Prompts) | Control Plane | Improvement |
|---|---|---|---|
| Safety Violation Rate | 26.67% | 0.00% | ✓ 100% enforcement |
| False Positive Rate | 0.00% | 0.00% | ✓ Perfect precision |
| Avg Output Tokens | 26.1 | 0.5 | ✓ 98.1% reduction |
Key Findings:
- ✓ 100% Safety Enforcement: Control Plane blocked all 45 malicious prompts (0% SVR)
- ✓ Zero False Positives: All 15 valid requests were correctly allowed
- ✓ Scale by Subtraction: 98.1% fewer tokens (returns "NULL" instead of verbose refusals)
- ✓ Jailbreak Immunity: Deterministic enforcement catches prompt injection attacks that bypass prompt-based safety
The benchmark includes:
- 15 Direct Violations: SQL injection, system commands
- 15 Prompt Injections: Jailbreaks and instruction overrides
- 15 Contextual Confusion: Social engineering attempts
- 15 Valid Requests: Legitimate operations (false positive testing)
See benchmark/README.md for detailed methodology and results.
Deploy agents with strict governance for enterprise environments:
- Compliance with security policies through Constraint Graphs
- Complete audit trails for regulatory requirements
- Resource quotas to control costs
- Shadow Mode testing before production deployment
Build precise, non-creative agents:
- Mute Agent configuration for SQL-only operations
- Returns NULL for out-of-scope requests
- No hallucination or conversational pivots
- Example: Finance team data access agent
Safely run multiple agents with isolation:
- Per-tenant quotas and policies
- Isolated execution environments
- Fair resource allocation
- Supervisor Agents monitoring all tenants
Experiment safely with agent capabilities:
- Shadow Mode for risk-free testing
- Sandboxed execution
- Complete reasoning telemetry
- Comprehensive logging
Run reliable, auditable agent workflows:
- Error handling and recovery
- Performance monitoring
- Traceability for debugging
See architecture.md for detailed architecture documentation.
AgentControlPlane: Main control plane interfaceAgentKernel: Core kernel componentPolicyEngine: Policy enforcementExecutionEngine: Safe executionAgentContext: Agent session contextExecutionRequest: Action requestExecutionResult: Action result
FILE_READ: Read file operationsFILE_WRITE: Write file operationsCODE_EXECUTION: Execute codeAPI_CALL: Make API callsDATABASE_QUERY: Query databasesDATABASE_WRITE: Write to databasesWORKFLOW_TRIGGER: Trigger workflows
NONE: No accessREAD_ONLY: Read-only accessREAD_WRITE: Read and write accessADMIN: Full administrative access
- Start with minimal permissions: Grant only what's needed
- Use rate limits: Prevent runaway agents
- Enable audit logging: Track all agent actions
- Test policies: Validate governance rules work as expected
- Monitor resource usage: Watch for anomalies
- Regular policy reviews: Keep policies up to date
- Default policies block system file access
- Credentials should never be in parameters
- High-risk actions require elevated permissions
- All actions are audited
- Sandboxed execution by default
- Distributed execution across multiple nodes
- Integration with external policy engines (OPA, etc.)
- Real-time monitoring dashboard
- Machine learning-based risk assessment
- Automatic policy generation from past behavior
- Integration with secrets management systems
- Container-based sandboxing
- Transaction rollback for database operations
Projects like Steve Yegge's Gas Town use a "City" metaphor where a "Mayor" agent orchestrates "Worker" agents to maximize coding throughput. This is brilliant for velocity.
The Difference:
- Gas Town solves for COORDINATION (getting things done fast)
- Agent Control Plane solves for CONTAINMENT (ensuring things are safe)
- In an enterprise, you don't just need a Manager; you need a Compliance Officer who can pull the plug
The Agent Control Plane complements coordination systems by providing the safety layer.
Most current safety tools operate as "sidecars" that check input/output for toxicity, PII, or harmful content. They are largely text-based and probabilistic.
The Difference:
- Guardrails are ADVISORY or REACTIVE (sanitizing output after generation)
- Agent Control Plane is ARCHITECTURAL (preventing action at the kernel level)
- A guardrail scrubs a bad SQL query; a Control Plane ensures the agent never had the connection string to begin with
- Guardrails work on content; Control Plane works on capabilities and execution
Recent academic papers propose "Control Planes" that act as a phonebook, helping agents find the right tools.
The Difference:
- Tool Directory is SERVICE DISCOVERY (finding what's available)
- Agent Control Plane is a KERNEL (strict enforcement of boundaries)
- The Linux Kernel doesn't just "help" processes find memory; it strictly enforces that Process A cannot touch Process B's memory
- We need that same hardness for Agents
Deterministic Enforcement, Not Advisory Hints:
- LLM can "think" whatever it wants
- But it can only ACT on what the Control Plane permits
- Constraint Graphs define the "physics" of the agent's world
- Shadow Mode lets you test everything before production
- Supervisor Agents provide recursive oversight
This is systems engineering for AI, not prompt engineering.
We welcome contributions! The Agent Control Plane is designed to be production-ready and contributor-friendly.
- Fork the repository
- Clone your fork:
git clone https://github.com/YOUR-USERNAME/agent-control-plane.git - Install in development mode:
pip install -e ".[dev]" - Create a branch:
git checkout -b feature/your-feature-name
# Run all tests
python -m unittest discover -s tests -p 'test_*.py' -v
# Run specific test file
python -m unittest tests/test_control_plane.py
# Run specific test
python -m unittest tests.test_control_plane.TestAgentKernel.test_create_agent_sessionsrc/agent_control_plane/- Main package source codetests/- Test suite (unittest framework)examples/- Example scripts and use casesdocs/- Documentation and guides.github/workflows/- CI/CD configuration
- Follow existing code style and patterns
- Add tests for new features
- Update documentation as needed
- Keep changes focused and minimal
- Write clear commit messages
See CONTRIBUTING.md for detailed guidelines.
The project uses Python's built-in unittest framework. All tests are located in the tests/ directory.
- Core functionality tests:
test_control_plane.py - Advanced features tests:
test_advanced_features.py
Current test coverage: 31 tests covering:
- Agent creation and lifecycle
- Permission management
- Policy enforcement
- Rate limiting
- Shadow mode simulation
- Mute agent capabilities
- Constraint graphs
- Supervisor agents
- Audit logging
Comprehensive documentation is available in the docs/ directory:
- Quick Start Guide - Get up and running quickly
- Implementation Guide - Detailed implementation details
- Philosophy - Core principles and design philosophy
- Architecture - System architecture overview
MIT License - See LICENSE file for details