Local LLM
AgentGuard CB and the MerchantGuard MCP server are model-agnostic. Drive them from a fully local LLM via Ollama, LM Studio, or any Qwen-compatible runtime. The evidence pipeline never invokes an LLM, so the local model is only used to orchestrate calls.
Air-gap mode: when running AgentGuard CB self-hosted with a local LLM, no merchant data, evidence payload, or audit-log entry crosses the network. Every component runs on infrastructure you control.
Ollama
# Pull a tool-capable model (example: Qwen 2.5 32B)
ollama pull qwen2.5:32b
# Run an MCP-capable client (Claude Code, Cline, Continue.dev)
# pointing it at the local Ollama endpoint and the MerchantGuard MCP server.LM Studio
- Open LM Studio → Discover → search for a tool-calling model (Qwen 2.5, Llama 3.1 with tools, etc.).
- Start the LM Studio local server (default http://localhost:1234).
- Configure your MCP client (Cline, Continue.dev, custom agent) to use the LM Studio endpoint.
- Add the MerchantGuard MCP URL to the same client.
Qwen-direct
- Qwen models run cleanly under vLLM or transformers locally.
- Use any MCP-compatible client to bridge between your Qwen runtime and the MerchantGuard MCP.
- No special configuration on the MerchantGuard side; the MCP server is transport-agnostic.
What still phones home
- AgentGuard CB self-hosted: nothing. Zero telemetry by default.
- MerchantGuard MCP (Cloud Run): the MCP server itself is hosted by us. If you want a fully air-gapped MCP, fork our repo and run the MCP locally.
- Stripe submissions: handled by your own Stripe SDK on your infrastructure.
Nothing on this page is legal advice. Trademarks pending: MerchantGuard™ (Serial 99051215), GuardScore™ (Serial 99030125), AgentGuard™ (Serial 99462472). Patents 63/983,615 / 63/983,621 / 63/983,843 / 63/984,626 (provisional, filed Feb 17 2026). See LEGAL.md, PATENTS.md, DISCLAIMER.md.

