SGLang
SGLang
Section titled “SGLang”SGLang can serve open-source models via an OpenAI-compatible HTTP API.
OpenClaw can connect to SGLang using the openai-completions API.
OpenClaw can also auto-discover available models from SGLang when you opt
in with SGLANG_API_KEY (any value works if your server does not enforce auth)
and you do not define an explicit models.providers.sglang entry.
Quick start
Section titled “Quick start”- Start SGLang with an OpenAI-compatible server.
Your base URL should expose /v1 endpoints (for example /v1/models,
/v1/chat/completions). SGLang commonly runs on:
http://127.0.0.1:30000/v1
- Opt in (any value works if no auth is configured):
export SGLANG_API_KEY="sglang-local"- Run onboarding and choose
SGLang, or set a model directly:
openclaw onboard{ agents: { defaults: { model: { primary: "sglang/your-model-id" }, }, },}Model discovery (implicit provider)
Section titled “Model discovery (implicit provider)”When SGLANG_API_KEY is set (or an auth profile exists) and you do not
define models.providers.sglang, OpenClaw will query:
GET http://127.0.0.1:30000/v1/models
and convert the returned IDs into model entries.
If you set models.providers.sglang explicitly, auto-discovery is skipped and
you must define models manually.
Explicit configuration (manual models)
Section titled “Explicit configuration (manual models)”Use explicit config when:
- SGLang runs on a different host/port.
- You want to pin
contextWindow/maxTokensvalues. - Your server requires a real API key (or you want to control headers).
{ models: { providers: { sglang: { baseUrl: "http://127.0.0.1:30000/v1", apiKey: "${SGLANG_API_KEY}", api: "openai-completions", models: [ { id: "your-model-id", name: "Local SGLang Model", reasoning: false, input: ["text"], cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, contextWindow: 128000, maxTokens: 8192, }, ], }, }, },}Troubleshooting
Section titled “Troubleshooting”- Check the server is reachable:
curl http://127.0.0.1:30000/v1/models- If requests fail with auth errors, set a real
SGLANG_API_KEYthat matches your server configuration, or configure the provider explicitly undermodels.providers.sglang.
Proxy-style behavior
Section titled “Proxy-style behavior”SGLang is treated as a proxy-style OpenAI-compatible /v1 backend, not a
native OpenAI endpoint.
- native OpenAI-only request shaping does not apply here
- no
service_tier, no Responsesstore, no prompt-cache hints, and no OpenAI reasoning-compat payload shaping - hidden OpenClaw attribution headers (
originator,version,User-Agent) are not injected on custom SGLang base URLs