← All docs

Local LLM

AgentGuard CB and the MerchantGuard MCP server are model-agnostic. Drive them from a fully local LLM via Ollama, LM Studio, or any Qwen-compatible runtime. The evidence pipeline never invokes an LLM, so the local model is only used to orchestrate calls.

Air-gap mode: when running AgentGuard CB self-hosted with a local LLM, no merchant data, evidence payload, or audit-log entry crosses the network. Every component runs on infrastructure you control.

Ollama

# Pull a tool-capable model (example: Qwen 2.5 32B)
ollama pull qwen2.5:32b

# Run an MCP-capable client (Claude Code, Cline, Continue.dev)
# pointing it at the local Ollama endpoint and the MerchantGuard MCP server.

LM Studio

  • Open LM Studio → Discover → search for a tool-calling model (Qwen 2.5, Llama 3.1 with tools, etc.).
  • Start the LM Studio local server (default http://localhost:1234).
  • Configure your MCP client (Cline, Continue.dev, custom agent) to use the LM Studio endpoint.
  • Add the MerchantGuard MCP URL to the same client.

Qwen-direct

  • Qwen models run cleanly under vLLM or transformers locally.
  • Use any MCP-compatible client to bridge between your Qwen runtime and the MerchantGuard MCP.
  • No special configuration on the MerchantGuard side; the MCP server is transport-agnostic.

What still phones home

  • AgentGuard CB self-hosted: nothing. Zero telemetry by default.
  • MerchantGuard MCP (Cloud Run): the MCP server itself is hosted by us. If you want a fully air-gapped MCP, fork our repo and run the MCP locally.
  • Stripe submissions: handled by your own Stripe SDK on your infrastructure.
Nothing on this page is legal advice. Trademarks pending: MerchantGuard™ (Serial 99051215), GuardScore™ (Serial 99030125), AgentGuard™ (Serial 99462472). Patents 63/983,615 / 63/983,621 / 63/983,843 / 63/984,626 (provisional, filed Feb 17 2026). See LEGAL.md, PATENTS.md, DISCLAIMER.md.