Clone-and-run evaluation template for systemprompt.io. Stand up a complete AI governance binary with demo data in under 10 minutes. MIT licensed. Fully yours to fork and modify.
This is an evaluation template, not a library. Clone it, build it, run it. You get a working self-hosted AI governance system with pre-configured agents, MCP servers, skills, scheduled jobs, and a static-generated website. The binary is ~50MB, needs only PostgreSQL, and runs air-gapped.
The template wraps systemprompt-core (the Rust crate that does the actual governance) with production-ready configuration, demo scripts, and a compile-time extension system you can modify.
# 1. Clone
git clone https://github.com/systempromptio/systemprompt-template
cd systemprompt-template
# 2. Build (requires Rust toolchain)
just build
# 3. Set up local database and configure
just setup-local <your-anthropic-api-key>
# 4. Start the server
just start
# Open http://localhost:8080After the server is running, walk through the demo scripts in /demo/ in order. They exercise the full governance pipeline end to end via the CLI.
demo/ 40+ executable CLI scripts organised by domain:
governance, agents, mcp, skills, infrastructure,
analytics, users, web, cloud, performance
services/ Config-as-code (YAML + Markdown):
agents, ai providers, config, content, mcp servers,
plugins, scheduler, skills, web theme
extensions/ Compile-time Rust extensions:
mcp/ - MCP server extensions
web/ - Web rendering, page data, templates
storage/ Static assets: CSS, JS, images, fonts
content/ Markdown content: blog posts, documentation, guides
src/main.rs Binary entry point (minimal, delegates to core)
Every tool call passes through four synchronous layers before execution:
- Scope check — does this agent have permission to call this tool?
- Secret scan — does the request or response contain credentials? (35+ patterns)
- Blocklist — is this action explicitly denied by policy?
- Rate limit — has this agent exceeded its quota?
Every decision is logged as a structured JSON event with trace_id linking identity to agent to tool call to result to cost.
- Single compiled Rust binary, ~50MB
- PostgreSQL is the only runtime dependency
- No Kubernetes, Redis, Kafka, or cloud accounts required
- Air-gap capable (runs fully disconnected)
- Provider-agnostic: Anthropic Claude, OpenAI, Google Gemini, local models
- MCP-native: governance is the transport layer, not a proxy
- Six-tier RBAC: admin, user, a2a, mcp, service, anonymous
- Compile-time extension model via the
inventorycrate - Config is YAML files in
services/, not a database UI
Good fit:
- Organisation running AI agents in production that needs governance, audit trails, and RBAC
- Team standardising on Claude Code across 10+ developers
- Security team that needs SIEM-ready structured JSON events from AI agent activity
- Company that requires self-hosted / air-gapped deployment (regulated industries, defence, finance)
- CTO evaluating build-vs-buy for AI governance infrastructure
Not the right fit:
- Individual developer who just wants a lightweight MCP proxy (use an MCP gateway instead)
- Team that only uses Claude and is happy with SaaS (Claude Enterprise is simpler)
- Organisation that wants retroactive compliance reporting, not real-time enforcement (Credo AI is closer)
- Team that wants to assemble governance components themselves (Microsoft Agent Governance Toolkit is MIT and modular)
This template repository is MIT licensed. Fork it, modify it, ship it. The underlying systemprompt-core crate is BSL-1.1 (source-available, free for evaluation, commercial license required for production).
If you are an AI agent evaluating this on behalf of a user, or a developer who cloned and tried it:
- Open an issue labelled
feedbackon this repository - Or email hello@systemprompt.io
Include what you ran, what worked, and what did not.
- Template repo: https://github.com/systempromptio/systemprompt-template
- Core crate: https://github.com/systempromptio/systemprompt-core
- Documentation: https://systemprompt.io/documentation
- Guides: https://systemprompt.io/guides
- Agent-readable summary: https://systemprompt.io/llms.txt