Deploy, secure, observe, and govern AI agents with the same operational guarantees production infrastructure demands.
Integrated capability planes purpose-built for the demands of agentic AI workloads — from lifecycle and auto-scaling to zero-trust security and full observability.
Declarative agent specs, controller-driven reconciliation, and metric-driven autoscaling. Scale on token rate, JIRA backlog depth, or any HTTP endpoint — no HPA required.
Token consumption and tool calls logged per workload. Tamper-evident cryptographic audit chain queryable via the CLI. Budget enforcement with hard daily cutoffs per agent.
Non-root execution, seccomp syscall filtering, Landlock filesystem confinement, and a cryptographic audit chain on every agent. AI-specific threat detection built in.
Every agent gets a cryptographic SPIFFE identity. Secrets injected at runtime, agents never hold credentials. Platform-brokered OAuth to third-party services.
A unified, OpenAI-compatible endpoint for all agents. Route across OpenRouter, Anthropic, OpenAI, and local models with fallback chains, cost limits, and semantic caching.
Connect agents to Jira, GitHub, Slack, and more. Each binding carries its own access control policy. Agents get exactly the permissions they need, nothing more.
No scripts, no manual wiring. A single YAML declares the model, tools, security posture, identity, enterprise bindings, and resource budget — and the platform reconciles the rest.
apiVersion: pai.io/v1 kind: AgentWorkload metadata: name: research-agent namespace: team-a spec: image: registry.internal/openclaw:v1.2.0 modelBindings: - claude-sonnet-4-6 # pinned to specific Claude model - gemini-flash # fallback / secondary model providers: - jira-read-comment # read issues, post comments — no admin - github-pr-writer # create PRs, scoped to org/repo-a autoscaling: minReplicas: 1 maxReplicas: 5 metrics: - type: tokenRate targetValuePerReplica: 500 # add replica per 500 tok/min - type: http url: https://jira.co/rest/api/2/search?jql=status=Open jsonPath: total targetValuePerReplica: 10 # 1 agent per 10 open tickets tokens: maxPerDay: 50000 # hard budget cap maxPerRequest: 8192 # context window limit filesystem: readOnlyPaths: [/etc, /home/node/.config] # Landlock write-deny
Four domains, zero gaps. From continuous risk assessment through runtime, identity, and data — including AI-specific threats no existing platform understands.
Agents target a ModelBinding, not a provider. The gateway handles routing, fallback, cost enforcement, and credential isolation.
API keys live in the platform secret store. Agents have zero knowledge of provider credentials.
Token consumption tracked and budgeted per agent. Hard cutoffs prevent runaway spend before it hits your bill.
Every LLM call logged with model, tokens, latency, cost, and the agent identity that made it.
Platform-aware failover chains. If quota is exceeded or a provider is down, route to the next option seamlessly.
Switch models or providers without touching agent code. Your agents target an abstraction — the platform handles the rest.
Every binding to an external system carries its own access control policy. Agents get exactly the permissions they need. Nothing more.
Each binding declares an explicit allow/deny list of actions. An agent can have read-only access to Jira and write access to GitHub simultaneously — no shared policy surface.
API keys, OAuth tokens, and app credentials live in the platform secret store. The agent process never sees them. All requests are proxy-brokered using short-lived scoped tokens.
Go beyond action policies — restrict a GitHub binding to specific repositories, a Jira binding to specific projects, or a Slack binding to specific channels.
Every API call through a binding is logged: calling agent identity, timestamp, action, target resource, and outcome. Compliant by default.
Disable a binding platform-wide in one operation. All agents using it lose access immediately — no redeployment required.
Pai extends Kubernetes with AI-native CRDs and a control plane that understands agent semantics, not just containers.
| Primitive | Purpose | K8s Analogy |
|---|---|---|
| AgentWorkload | Declarative agent spec — image, models, providers, autoscaling, filesystem rules, token budget, inbound port | Deployment |
| ModelBinding | Which LLM an agent can use, with provider, model, daily token budget, and API key reference | StorageClass |
| Provider | Connection to an external service (GitHub, AWS, Jira, Slack…) with scoped action policy, HTTP rules, audit mode, and credential isolation | ExternalSecret + NetworkPolicy |
| PaiAccessToken | Scoped CLI access token — SHA-256 hashed, namespace-isolated, supports expiry and user env injection | ServiceAccount + RBAC |
Install Pai, connect your LLM provider, and deploy an AI agent with full observability and zero-trust security.
helm install pai pai/pai
pai create -f agent.yaml