Models propose. The kernel disposes.
HELM is a kernel-grade execution authority for AI agents. Every tool call, sandbox execution, and self-extension goes through fail-closed governance — producing tamper-proof receipts and deterministic EvidencePacks you can hand to auditors, regulators, or your board.
What you get in 10 minutes:
- 🔒 Fail-closed governance — every action is ALLOW/DENY with a signed receipt
- 📦 Deterministic EvidencePacks — offline-verifiable, air-gapped safe, bit-identical
- 📊 Interactive Proof Report — shareable HTML with causal chain visualization
- 🔌 Works with your stack — OpenAI SDK, LangChain, Claude, Mastra, any MCP client
- 🧱 Kernel-grade trust — Ed25519 signed, Lamport-ordered, replay-from-genesis
📊 What the Proof Report looks like
The
helm demo companycommand generates an interactive HTML proof report with causal chain visualization, receipt details, verification status, and one-click sharing. Opendata/evidence/run-report.htmlafter running the demo.
# Script install (macOS / Linux)
curl -fsSL https://raw.githubusercontent.com/Mindburn-Labs/helm-oss/main/install.sh | bash
# Go
go install github.com/Mindburn-Labs/helm-oss/core/cmd/helm@latest
# Docker
docker run --rm ghcr.io/mindburn-labs/helm-oss:latest --help
# Homebrew (coming soon)
# brew install mindburn-labs/tap/helm# Claude Desktop — one-click .mcpb
helm mcp pack --client claude-desktop --out helm.mcpb
# Claude Code
helm mcp install --client claude-code
# Windsurf / Codex / VS Code / Cursor
helm mcp print-config --client windsurfSDKs ship in-repo and can be used directly:
# TypeScript SDK (from repo)
cd sdk/ts && npm install && npm run build
# Python SDK (from repo)
cd sdk/python && pip install -e .
# Go SDK
go get github.com/Mindburn-Labs/helm-oss/sdk/gonpm packages are live: npm install @mindburn/helm (SDK), npx @mindburn/helm-cli (verifier).
HELM is built for high-stakes, low-latency environments. To measure the overhead on your machine:
./scripts/bench/latency.sh# 1. Setup (SQLite + Ed25519 + config — instant)
helm onboard --yes
# 2. Run governed company demo (15 receipts, 7 phases: approval → sandbox → deny → skill gap → incident)
helm demo company --template starter --provider mock
# 3. Export deterministic EvidencePack + verify offline (air-gapped safe)
helm export --evidence ./data/evidence --out evidence.tar
helm verify --bundle evidence.tar
# 4. Explore skill lifecycle + maintenance loop
helm pack list && helm incident list && helm brief daily→ Full commands: docs/VERIFICATION.md · docs/QUICKSTART.md
npx @mindburn/helm-cliOne command, progressive disclosure, cryptographic proof. Supports interactive and CI modes:
# CI mode — JSON on stdout, exit code 0/1
npx @mindburn/helm-cli --ci --bundle ./evidence 2>/dev/null | jq .verdict→ Full guide: docs/verify.md
| Pain (postmortem you're preventing) | HELM behavior | Receipt reason code | Proof |
|---|---|---|---|
| Tool-call overspend blows budget | ACID budget locks, fail-closed on ceiling breach | DENY_BUDGET_EXCEEDED |
UC-005 |
| Schema drift breaks prod silently | Fail-closed on input AND output schema mismatch | DENY_SCHEMA_MISMATCH |
UC-002, UC-009 |
| Untrusted WASM runs wild | Sandbox: gas + time + memory budgets, deterministic traps | DENY_GAS_EXHAUSTION |
UC-004 |
| "Who approved that?" disputes | Timelock + challenge/response ceremony, Ed25519 signed | DENY_APPROVAL_REQUIRED |
UC-003 |
| No audit trail for regulators | Deterministic EvidencePack, offline verifiable, replay from genesis | — | UC-008 |
| Can't prove compliance to auditors | Conformance L1 + L2 gates, 12 runnable use cases | — | UC-012 |
The only change:
- client = openai.OpenAI()
+ client = openai.OpenAI(base_url="http://localhost:8080/v1")Full snippet:
import openai
client = openai.OpenAI(base_url="http://localhost:8080/v1")
response = client.chat.completions.create(
model="gpt-4",
messages=[{"role": "user", "content": "List files in /tmp"}]
)
print(response.choices[0].message.content)
# Response headers include:
# X-Helm-Decision-ID: dec_a1b2c3...
# X-Helm-Verdict: ALLOW
# X-Helm-Policy-Version: 1.0.0→ Full example: examples/python_openai_baseurl/main.py
The only change:
- const BASE = "https://api.openai.com/v1";
+ const BASE = "http://localhost:8080/v1";Full snippet:
const response = await fetch("http://localhost:8080/v1/chat/completions", {
method: "POST",
headers: { "Content-Type": "application/json" },
body: JSON.stringify({
model: "gpt-4",
messages: [{ role: "user", content: "What time is it?" }],
}),
});
const data = await response.json();
console.log(data.choices[0].message.content);
// X-Helm-Decision-ID: dec_d4e5f6...
// X-Helm-Verdict: ALLOW→ Full example: examples/js_openai_baseurl/main.js
# List governed capabilities
curl -s http://localhost:8080/mcp/v1/capabilities | jq '.tools[].name'
# Execute a governed tool call
curl -s -X POST http://localhost:8080/mcp/v1/execute \
-H 'Content-Type: application/json' \
-d '{"method":"file_read","params":{"path":"/tmp/test.txt"}}' | jq .
# → { "result": ..., "receipt_id": "rec_...", "reason_code": "ALLOW" }
→ Full example: [examples/mcp_client/main.sh](examples/mcp_client/main.sh)
---
## SDKs
Typed clients for 5 languages. All generated from [api/openapi/helm.openapi.yaml](api/openapi/helm.openapi.yaml).
| Language | Status | Path |
| :--- | :--- | :--- |
| **TypeScript** | In-repo | `sdk/ts/` |
| **Python** | In-repo | `sdk/python/` |
| **Go** | In-repo | `sdk/go/` |
| **Rust** | Preview | `sdk/rust/` |
| **Java** | Preview | `sdk/java/` |
Every SDK exposes the same primitives: `chatCompletions`, `approveIntent`, `listSessions`, `getReceipts`, `exportEvidence`, `verifyEvidence`, `conformanceRun`.
Every error includes a typed `reason_code` (e.g. `DENY_TOOL_NOT_FOUND`).
**Go — 10-line denial-handling example:**
```go
c := helm.New("http://localhost:8080")
res, err := c.ChatCompletions(helm.ChatCompletionRequest{
Model: "gpt-4",
Messages: []helm.ChatMessage{{Role: "user", Content: "List /tmp"}},
})
if apiErr, ok := err.(*helm.HelmApiError); ok {
fmt.Println("Denied:", apiErr.ReasonCode) // DENY_TOOL_NOT_FOUND
}Rust:
let c = HelmClient::new("http://localhost:8080");
match c.chat_completions(&req) {
Ok(res) => println!("{:?}", res.choices[0].message.content),
Err(e) => println!("Denied: {:?}", e.reason_code),
}Java:
var helm = new HelmClient("http://localhost:8080");
try { helm.chatCompletions(req); }
catch (HelmApiException e) { System.out.println(e.reasonCode); }Full examples: examples/ · SDK docs: docs/sdks/00_INDEX.md
api/openapi/helm.openapi.yaml — OpenAPI 3.1 spec.
Single source of truth. SDKs are generated from it. CI prevents drift.
Your App (OpenAI SDK)
│
│ base_url = localhost:8080
▼
HELM Proxy ──→ Guardian (policy: allow/deny)
│ │
│ PEP Boundary (JCS canonicalize → SHA-256)
│ │
▼ ▼
Executor ──→ Tool ──→ Receipt (Ed25519 signed)
│ │
▼ ▼
ProofGraph DAG EvidencePack (.tar)
(append-only) (offline verifiable)
│
▼
Replay Verify
(air-gapped safe)
| Shipped in OSS v1.0 |
|---|
| ✅ OpenAI-compatible governed proxy |
| ✅ Schema PEP (input + output) |
| ✅ ProofGraph DAG (Lamport + Ed25519) |
| ✅ WASI sandbox (gas/time/memory) |
| ✅ Approval ceremonies (timelock + challenge) |
| ✅ Trust registry (event-sourced) |
| ✅ EvidencePack export + offline replay |
| ✅ Proof Condensation (Merkle checkpoints) |
| ✅ CPI (Canonical Policy Index) |
| ✅ HSM signing (Ed25519) |
| ✅ Policy Bundles (load, verify, compose) |
| ✅ Conformance L1 + L2 + L3 |
| ✅ 11 CLI commands |
Full scope details in docs/OSS_SCOPE.md
make test # 115 packages, 0 failures
make crucible # 12 use cases + conformance L1/L2
make lint # go vet, clean# Local demo
docker compose up -d
# Production (DigitalOcean / any Docker host)
docker compose -f docker-compose.demo.yml up -d→ deploy/README.md — deploy your own in 3 minutes
helm/
├── api/openapi/ # OpenAPI 3.1 spec (single source of truth)
├── core/ # Go kernel (8-package TCB + executor + ProofGraph)
│ └── cmd/helm/ # CLI: proxy, export, verify, replay, conform, ...
├── packages/
│ └── mindburn-helm-cli/ # @mindburn/helm-cli (npm CLI verifier)
├── sdk/ # Multi-language SDKs (TS, Python, Go, Rust, Java)
├── examples/ # Runnable examples per language + MCP
├── scripts/ # Release, CI, SDK generation
├── deploy/ # Caddy config, demo compose, deploy guide
├── docs/ # Threat model, quickstart, verify, conformance
└── Makefile # build, test, crucible, demo, release-binaries
OSS ships L1 + L2 + L3 conformance. See docs/OSS_SCOPE.md for the shipped-vs-spec boundary.
- TCB isolation gate — 8-package kernel boundary, CI-enforced forbidden imports (TCB Policy)
- Bounded compute gate — WASI sandbox with gas/time/memory caps, deterministic traps on breach (UC-005)
- Schema drift fail-closed — JCS canonicalization + SHA-256 on every tool call, both input and output (UC-002)
See also: SECURITY.md (vulnerability reporting) · Threat Model (9 adversary classes)
See CONTRIBUTING.md. Good first issues: conformance improvements, SDK enhancements, docs truth fixes.
See the GitHub Issues for planned items.
Built by Mindburn Labs.