chore: Run pnpm format:fix.

This commit is contained in:
cpojer
2026-01-31 21:13:13 +09:00
parent dcc2de15a6
commit 8cab78abbc
624 changed files with 10729 additions and 7514 deletions

View File

@@ -4,6 +4,7 @@ read_when:
- You want to use Anthropic models in OpenClaw
- You want setup-token instead of API keys
---
# Anthropic (Claude)
Anthropic builds the **Claude** model family and provides access via an API.
@@ -29,7 +30,7 @@ openclaw onboard --anthropic-api-key "$ANTHROPIC_API_KEY"
```json5
{
env: { ANTHROPIC_API_KEY: "sk-ant-..." },
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } }
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } },
}
```
@@ -46,11 +47,11 @@ To set the TTL per model, use `cacheControlTtl` in the model `params`:
defaults: {
models: {
"anthropic/claude-opus-4-5": {
params: { cacheControlTtl: "5m" } // or "1h"
}
}
}
}
params: { cacheControlTtl: "5m" }, // or "1h"
},
},
},
},
}
```
@@ -92,7 +93,7 @@ openclaw onboard --auth-choice setup-token
```json5
{
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } }
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } },
}
```
@@ -105,21 +106,25 @@ openclaw onboard --auth-choice setup-token
## Troubleshooting
**401 errors / token suddenly invalid**
- Claude subscription auth can expire or be revoked. Re-run `claude setup-token`
and paste it into the **gateway host**.
- If the Claude CLI login lives on a different machine, use
`openclaw models auth paste-token --provider anthropic` on the gateway host.
**No API key found for provider "anthropic"**
- Auth is **per agent**. New agents dont inherit the main agents keys.
- Re-run onboarding for that agent, or paste a setup-token / API key on the
gateway host, then verify with `openclaw models status`.
**No credentials found for profile `anthropic:default`**
- Run `openclaw models status` to see which auth profile is active.
- Re-run onboarding, or paste a setup-token / API key for that profile.
**No available auth profile (all in cooldown/unavailable)**
- Check `openclaw models status --json` for `auth.unusableProfiles`.
- Add another Anthropic profile or wait for cooldown.

View File

@@ -5,16 +5,17 @@ read_when:
- You want a local API server that wraps Claude Code CLI
- You want to save money by using subscription instead of API keys
---
# Claude Max API Proxy
**claude-max-api-proxy** is a community tool that exposes your Claude Max/Pro subscription as an OpenAI-compatible API endpoint. This allows you to use your subscription with any tool that supports the OpenAI API format.
## Why Use This?
| Approach | Cost | Best For |
|----------|------|----------|
| Anthropic API | Pay per token (~$15/M input, $75/M output for Opus) | Production apps, high volume |
| Claude Max subscription | $200/month flat | Personal use, development, unlimited usage |
| Approach | Cost | Best For |
| ----------------------- | --------------------------------------------------- | ------------------------------------------ |
| Anthropic API | Pay per token (~$15/M input, $75/M output for Opus) | Production apps, high volume |
| Claude Max subscription | $200/month flat | Personal use, development, unlimited usage |
If you have a Claude Max subscription and want to use it with OpenAI-compatible tools, this proxy can save you significant money.
@@ -26,6 +27,7 @@ Your App → claude-max-api-proxy → Claude Code CLI → Anthropic (via subscri
```
The proxy:
1. Accepts OpenAI-format requests at `http://localhost:3456/v1/chat/completions`
2. Converts them to Claude Code CLI commands
3. Returns responses in OpenAI format (streaming supported)
@@ -75,23 +77,23 @@ You can point OpenClaw at the proxy as a custom OpenAI-compatible endpoint:
{
env: {
OPENAI_API_KEY: "not-needed",
OPENAI_BASE_URL: "http://localhost:3456/v1"
OPENAI_BASE_URL: "http://localhost:3456/v1",
},
agents: {
defaults: {
model: { primary: "openai/claude-opus-4" }
}
}
model: { primary: "openai/claude-opus-4" },
},
},
}
```
## Available Models
| Model ID | Maps To |
|----------|---------|
| `claude-opus-4` | Claude Opus 4 |
| Model ID | Maps To |
| ----------------- | --------------- |
| `claude-opus-4` | Claude Opus 4 |
| `claude-sonnet-4` | Claude Sonnet 4 |
| `claude-haiku-4` | Claude Haiku 4 |
| `claude-haiku-4` | Claude Haiku 4 |
## Auto-Start on macOS

View File

@@ -4,6 +4,7 @@ read_when:
- You want Deepgram speech-to-text for audio attachments
- You need a quick Deepgram config example
---
# Deepgram (Audio Transcription)
Deepgram is a speech-to-text API. In OpenClaw it is used for **inbound audio/voice note
@@ -18,22 +19,24 @@ Docs: https://developers.deepgram.com
## Quick start
1) Set your API key:
1. Set your API key:
```
DEEPGRAM_API_KEY=dg_...
```
2) Enable the provider:
2. Enable the provider:
```json5
{
tools: {
media: {
audio: {
enabled: true,
models: [{ provider: "deepgram", model: "nova-3" }]
}
}
}
models: [{ provider: "deepgram", model: "nova-3" }],
},
},
},
}
```
@@ -46,22 +49,22 @@ DEEPGRAM_API_KEY=dg_...
- `tools.media.audio.providerOptions.deepgram.smart_format`: enable smart formatting (optional)
Example with language:
```json5
{
tools: {
media: {
audio: {
enabled: true,
models: [
{ provider: "deepgram", model: "nova-3", language: "en" }
]
}
}
}
models: [{ provider: "deepgram", model: "nova-3", language: "en" }],
},
},
},
}
```
Example with Deepgram options:
```json5
{
tools: {
@@ -72,13 +75,13 @@ Example with Deepgram options:
deepgram: {
detect_language: true,
punctuate: true,
smart_format: true
}
smart_format: true,
},
},
models: [{ provider: "deepgram", model: "nova-3" }]
}
}
}
models: [{ provider: "deepgram", model: "nova-3" }],
},
},
},
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want to use GitHub Copilot as a model provider
- You need the `openclaw models auth login-github-copilot` flow
---
# GitHub Copilot
## What is GitHub Copilot?
@@ -57,7 +58,7 @@ openclaw models set github-copilot/gpt-4o
```json5
{
agents: { defaults: { model: { primary: "github-copilot/gpt-4o" } } }
agents: { defaults: { model: { primary: "github-copilot/gpt-4o" } } },
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want GLM models in OpenClaw
- You need the model naming convention and setup
---
# GLM models
GLM is a **model family** (not a company) available through the Z.AI platform. In OpenClaw, GLM
@@ -20,7 +21,7 @@ openclaw onboard --auth-choice zai-api-key
```json5
{
env: { ZAI_API_KEY: "sk-..." },
agents: { defaults: { model: { primary: "zai/glm-4.7" } } }
agents: { defaults: { model: { primary: "zai/glm-4.7" } } },
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want to choose a model provider
- You need a quick overview of supported LLM backends
---
# Model Providers
OpenClaw can use many LLM providers. Pick a provider, authenticate, then set the
@@ -22,12 +23,12 @@ See [Venice AI](/providers/venice).
## Quick start
1) Authenticate with the provider (usually via `openclaw onboard`).
2) Set the default model:
1. Authenticate with the provider (usually via `openclaw onboard`).
2. Set the default model:
```json5
{
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } }
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } },
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want MiniMax models in OpenClaw
- You need MiniMax setup guidance
---
# MiniMax
MiniMax is an AI company that builds the **M2/M2.1** model family. The current
@@ -46,17 +47,20 @@ moltbot plugins enable minimax-portal-auth # skip if already loaded.
moltbot gateway restart # restart if gateway is already running
moltbot onboard --auth-choice minimax-portal
```
You will be prompted to select an endpoint:
- **Global** - International users (`api.minimax.io`)
- **CN** - Users in China (`api.minimaxi.com`)
See [MiniMax OAuth plugin README](https://github.com/moltbot/moltbot/tree/main/extensions/minimax-portal-auth) for details.
### MiniMax M2.1 (API key)
### MiniMax M2.1 (API key)
**Best for:** hosted MiniMax with Anthropic-compatible API.
Configure via CLI:
- Run `openclaw configure`
- Select **Model/auth**
- Choose **MiniMax M2.1**
@@ -80,12 +84,12 @@ Configure via CLI:
input: ["text"],
cost: { input: 15, output: 60, cacheRead: 2, cacheWrite: 10 },
contextWindow: 200000,
maxTokens: 8192
}
]
}
}
}
maxTokens: 8192,
},
],
},
},
},
}
```
@@ -100,14 +104,14 @@ Configure via CLI:
defaults: {
models: {
"anthropic/claude-opus-4-5": { alias: "opus" },
"minimax/MiniMax-M2.1": { alias: "minimax" }
"minimax/MiniMax-M2.1": { alias: "minimax" },
},
model: {
primary: "anthropic/claude-opus-4-5",
fallbacks: ["minimax/MiniMax-M2.1"]
}
}
}
fallbacks: ["minimax/MiniMax-M2.1"],
},
},
},
}
```
@@ -124,8 +128,8 @@ Configure manually via `openclaw.json`:
agents: {
defaults: {
model: { primary: "lmstudio/minimax-m2.1-gs32" },
models: { "lmstudio/minimax-m2.1-gs32": { alias: "Minimax" } }
}
models: { "lmstudio/minimax-m2.1-gs32": { alias: "Minimax" } },
},
},
models: {
mode: "merge",
@@ -142,12 +146,12 @@ Configure manually via `openclaw.json`:
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 196608,
maxTokens: 8192
}
]
}
}
}
maxTokens: 8192,
},
],
},
},
},
}
```
@@ -155,11 +159,10 @@ Configure manually via `openclaw.json`:
Use the interactive config wizard to set MiniMax without editing JSON:
1) Run `openclaw configure`.
2) Select **Model/auth**.
3) Choose **MiniMax M2.1**.
4) Pick your default model when prompted.
1. Run `openclaw configure`.
2. Select **Model/auth**.
3. Choose **MiniMax M2.1**.
4. Pick your default model when prompted.
## Configuration options
@@ -186,16 +189,19 @@ Use the interactive config wizard to set MiniMax without editing JSON:
This usually means the **MiniMax provider isnt configured** (no provider entry
and no MiniMax auth profile/env key found). A fix for this detection is in
**2026.1.12** (unreleased at the time of writing). Fix by:
- Upgrading to **2026.1.12** (or run from source `main`), then restarting the gateway.
- Running `openclaw configure` and selecting **MiniMax M2.1**, or
- Adding the `models.providers.minimax` block manually, or
- Setting `MINIMAX_API_KEY` (or a MiniMax auth profile) so the provider can be injected.
Make sure the model id is **casesensitive**:
- `minimax/MiniMax-M2.1`
- `minimax/MiniMax-M2.1-lightning`
Then recheck with:
```bash
openclaw models list
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want to choose a model provider
- You want quick setup examples for LLM auth + model selection
---
# Model Providers
OpenClaw can use many LLM providers. Pick one, authenticate, then set the default
@@ -20,12 +21,12 @@ See [Venice AI](/providers/venice).
## Quick start (two steps)
1) Authenticate with the provider (usually via `openclaw onboard`).
2) Set the default model:
1. Authenticate with the provider (usually via `openclaw onboard`).
2. Set the default model:
```json5
{
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } }
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } },
}
```

View File

@@ -13,13 +13,14 @@ provider and set the default model to `moonshot/kimi-k2.5`, or use
Kimi Coding with `kimi-coding/k2p5`.
Current Kimi K2 model IDs:
{/* moonshot-kimi-k2-ids:start */}
{/_ moonshot-kimi-k2-ids:start _/}
- `kimi-k2.5`
- `kimi-k2-0905-preview`
- `kimi-k2-turbo-preview`
- `kimi-k2-thinking`
- `kimi-k2-thinking-turbo`
{/* moonshot-kimi-k2-ids:end */}
{/_ moonshot-kimi-k2-ids:end _/}
```bash
openclaw onboard --auth-choice moonshot-api-key
@@ -47,10 +48,10 @@ Note: Moonshot and Kimi Coding are separate providers. Keys are not interchangea
"moonshot/kimi-k2-0905-preview": { alias: "Kimi K2" },
"moonshot/kimi-k2-turbo-preview": { alias: "Kimi K2 Turbo" },
"moonshot/kimi-k2-thinking": { alias: "Kimi K2 Thinking" },
"moonshot/kimi-k2-thinking-turbo": { alias: "Kimi K2 Thinking Turbo" }
"moonshot/kimi-k2-thinking-turbo": { alias: "Kimi K2 Thinking Turbo" },
// moonshot-kimi-k2-aliases:end
}
}
},
},
},
models: {
mode: "merge",
@@ -68,7 +69,7 @@ Note: Moonshot and Kimi Coding are separate providers. Keys are not interchangea
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 256000,
maxTokens: 8192
maxTokens: 8192,
},
{
id: "kimi-k2-0905-preview",
@@ -77,7 +78,7 @@ Note: Moonshot and Kimi Coding are separate providers. Keys are not interchangea
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 256000,
maxTokens: 8192
maxTokens: 8192,
},
{
id: "kimi-k2-turbo-preview",
@@ -86,7 +87,7 @@ Note: Moonshot and Kimi Coding are separate providers. Keys are not interchangea
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 256000,
maxTokens: 8192
maxTokens: 8192,
},
{
id: "kimi-k2-thinking",
@@ -95,7 +96,7 @@ Note: Moonshot and Kimi Coding are separate providers. Keys are not interchangea
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 256000,
maxTokens: 8192
maxTokens: 8192,
},
{
id: "kimi-k2-thinking-turbo",
@@ -104,13 +105,13 @@ Note: Moonshot and Kimi Coding are separate providers. Keys are not interchangea
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 256000,
maxTokens: 8192
}
maxTokens: 8192,
},
// moonshot-kimi-k2-models:end
]
}
}
}
],
},
},
},
}
```
@@ -123,16 +124,16 @@ Note: Moonshot and Kimi Coding are separate providers. Keys are not interchangea
defaults: {
model: { primary: "kimi-coding/k2p5" },
models: {
"kimi-coding/k2p5": { alias: "Kimi K2.5" }
}
}
}
"kimi-coding/k2p5": { alias: "Kimi K2.5" },
},
},
},
}
```
## Notes
- Moonshot model refs use `moonshot/<modelId>`. Kimi Coding model refs use `kimi-coding/<modelId>`.
- Moonshot model refs use `moonshot/<modelId>`. Kimi Coding model refs use `kimi-coding/<modelId>`.
- Override pricing and context metadata in `models.providers` if needed.
- If Moonshot publishes different context limits for a model, adjust
`contextWindow` accordingly.

View File

@@ -4,15 +4,16 @@ read_when:
- You want to run OpenClaw with local models via Ollama
- You need Ollama setup and configuration guidance
---
# Ollama
Ollama is a local LLM runtime that makes it easy to run open-source models on your machine. OpenClaw integrates with Ollama's OpenAI-compatible API and can **auto-discover tool-capable models** when you opt in with `OLLAMA_API_KEY` (or an auth profile) and do not define an explicit `models.providers.ollama` entry.
## Quick start
1) Install Ollama: https://ollama.ai
1. Install Ollama: https://ollama.ai
2) Pull a model:
2. Pull a model:
```bash
ollama pull llama3.3
@@ -22,7 +23,7 @@ ollama pull qwen2.5-coder:32b
ollama pull deepseek-r1:32b
```
3) Enable Ollama for OpenClaw (any value works; Ollama doesn't require a real key):
3. Enable Ollama for OpenClaw (any value works; Ollama doesn't require a real key):
```bash
# Set environment variable
@@ -32,15 +33,15 @@ export OLLAMA_API_KEY="ollama-local"
openclaw config set models.providers.ollama.apiKey "ollama-local"
```
4) Use Ollama models:
4. Use Ollama models:
```json5
{
agents: {
defaults: {
model: { primary: "ollama/llama3.3" }
}
}
model: { primary: "ollama/llama3.3" },
},
},
}
```
@@ -87,6 +88,7 @@ export OLLAMA_API_KEY="ollama-local"
### Explicit setup (manual models)
Use explicit config when:
- Ollama runs on another host/port.
- You want to force specific context windows or model lists.
- You want to include models that do not report tool support.
@@ -129,10 +131,10 @@ If Ollama is running on a different host or port (explicit config disables auto-
providers: {
ollama: {
apiKey: "ollama-local",
baseUrl: "http://ollama-host:11434/v1"
}
}
}
baseUrl: "http://ollama-host:11434/v1",
},
},
},
}
```
@@ -146,10 +148,10 @@ Once configured, all your Ollama models are available:
defaults: {
model: {
primary: "ollama/llama3.3",
fallback: ["ollama/qwen2.5-coder:32b"]
}
}
}
fallback: ["ollama/qwen2.5-coder:32b"],
},
},
},
}
```
@@ -190,6 +192,7 @@ curl http://localhost:11434/api/tags
### No models available
OpenClaw only auto-discovers models that report tool support. If your model isn't listed, either:
- Pull a tool-capable model, or
- Define the model explicitly in `models.providers.ollama`.

View File

@@ -4,6 +4,7 @@ read_when:
- You want to use OpenAI models in OpenClaw
- You want Codex subscription auth instead of API keys
---
# OpenAI
OpenAI provides developer APIs for GPT models. Codex supports **ChatGPT sign-in** for subscription
@@ -27,7 +28,7 @@ openclaw onboard --openai-api-key "$OPENAI_API_KEY"
```json5
{
env: { OPENAI_API_KEY: "sk-..." },
agents: { defaults: { model: { primary: "openai/gpt-5.2" } } }
agents: { defaults: { model: { primary: "openai/gpt-5.2" } } },
}
```
@@ -50,7 +51,7 @@ openclaw models auth login --provider openai-codex
```json5
{
agents: { defaults: { model: { primary: "openai-codex/gpt-5.2" } } }
agents: { defaults: { model: { primary: "openai-codex/gpt-5.2" } } },
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want OpenCode Zen for model access
- You want a curated list of coding-friendly models
---
# OpenCode Zen
OpenCode Zen is a **curated list of models** recommended by the OpenCode team for coding agents.
@@ -23,7 +24,7 @@ openclaw onboard --opencode-zen-api-key "$OPENCODE_API_KEY"
```json5
{
env: { OPENCODE_API_KEY: "sk-..." },
agents: { defaults: { model: { primary: "opencode/claude-opus-4-5" } } }
agents: { defaults: { model: { primary: "opencode/claude-opus-4-5" } } },
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want a single API key for many LLMs
- You want to run models via OpenRouter in OpenClaw
---
# OpenRouter
OpenRouter provides a **unified API** that routes requests to many models behind a single
@@ -22,9 +23,9 @@ openclaw onboard --auth-choice apiKey --token-provider openrouter --token "$OPEN
env: { OPENROUTER_API_KEY: "sk-or-..." },
agents: {
defaults: {
model: { primary: "openrouter/anthropic/claude-sonnet-4-5" }
}
}
model: { primary: "openrouter/anthropic/claude-sonnet-4-5" },
},
},
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want to use Qwen with OpenClaw
- You want free-tier OAuth access to Qwen Coder
---
# Qwen
Qwen provides a free-tier OAuth flow for Qwen Coder and Qwen Vision models

View File

@@ -4,6 +4,7 @@ read_when:
- You want to use Synthetic as a model provider
- You need a Synthetic API key or base URL setup
---
# Synthetic
Synthetic exposes Anthropic-compatible endpoints. OpenClaw registers it as the
@@ -11,8 +12,8 @@ Synthetic exposes Anthropic-compatible endpoints. OpenClaw registers it as the
## Quick setup
1) Set `SYNTHETIC_API_KEY` (or run the wizard below).
2) Run onboarding:
1. Set `SYNTHETIC_API_KEY` (or run the wizard below).
2. Run onboarding:
```bash
openclaw onboard --auth-choice synthetic-api-key
@@ -32,8 +33,8 @@ synthetic/hf:MiniMaxAI/MiniMax-M2.1
agents: {
defaults: {
model: { primary: "synthetic/hf:MiniMaxAI/MiniMax-M2.1" },
models: { "synthetic/hf:MiniMaxAI/MiniMax-M2.1": { alias: "MiniMax M2.1" } }
}
models: { "synthetic/hf:MiniMaxAI/MiniMax-M2.1": { alias: "MiniMax M2.1" } },
},
},
models: {
mode: "merge",
@@ -50,12 +51,12 @@ synthetic/hf:MiniMaxAI/MiniMax-M2.1
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 192000,
maxTokens: 65536
}
]
}
}
}
maxTokens: 65536,
},
],
},
},
},
}
```
@@ -67,27 +68,27 @@ its base URL, override `models.providers.synthetic.baseUrl`.
All models below use cost `0` (input/output/cache).
| Model ID | Context window | Max tokens | Reasoning | Input |
| --- | --- | --- | --- | --- |
| `hf:MiniMaxAI/MiniMax-M2.1` | 192000 | 65536 | false | text |
| `hf:moonshotai/Kimi-K2-Thinking` | 256000 | 8192 | true | text |
| `hf:zai-org/GLM-4.7` | 198000 | 128000 | false | text |
| `hf:deepseek-ai/DeepSeek-R1-0528` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3-0324` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3.1` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3.1-Terminus` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3.2` | 159000 | 8192 | false | text |
| `hf:meta-llama/Llama-3.3-70B-Instruct` | 128000 | 8192 | false | text |
| `hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | 524000 | 8192 | false | text |
| `hf:moonshotai/Kimi-K2-Instruct-0905` | 256000 | 8192 | false | text |
| `hf:openai/gpt-oss-120b` | 128000 | 8192 | false | text |
| `hf:Qwen/Qwen3-235B-A22B-Instruct-2507` | 256000 | 8192 | false | text |
| `hf:Qwen/Qwen3-Coder-480B-A35B-Instruct` | 256000 | 8192 | false | text |
| `hf:Qwen/Qwen3-VL-235B-A22B-Instruct` | 250000 | 8192 | false | text + image |
| `hf:zai-org/GLM-4.5` | 128000 | 128000 | false | text |
| `hf:zai-org/GLM-4.6` | 198000 | 128000 | false | text |
| `hf:deepseek-ai/DeepSeek-V3` | 128000 | 8192 | false | text |
| `hf:Qwen/Qwen3-235B-A22B-Thinking-2507` | 256000 | 8192 | true | text |
| Model ID | Context window | Max tokens | Reasoning | Input |
| ------------------------------------------------------ | -------------- | ---------- | --------- | ------------ |
| `hf:MiniMaxAI/MiniMax-M2.1` | 192000 | 65536 | false | text |
| `hf:moonshotai/Kimi-K2-Thinking` | 256000 | 8192 | true | text |
| `hf:zai-org/GLM-4.7` | 198000 | 128000 | false | text |
| `hf:deepseek-ai/DeepSeek-R1-0528` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3-0324` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3.1` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3.1-Terminus` | 128000 | 8192 | false | text |
| `hf:deepseek-ai/DeepSeek-V3.2` | 159000 | 8192 | false | text |
| `hf:meta-llama/Llama-3.3-70B-Instruct` | 128000 | 8192 | false | text |
| `hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | 524000 | 8192 | false | text |
| `hf:moonshotai/Kimi-K2-Instruct-0905` | 256000 | 8192 | false | text |
| `hf:openai/gpt-oss-120b` | 128000 | 8192 | false | text |
| `hf:Qwen/Qwen3-235B-A22B-Instruct-2507` | 256000 | 8192 | false | text |
| `hf:Qwen/Qwen3-Coder-480B-A35B-Instruct` | 256000 | 8192 | false | text |
| `hf:Qwen/Qwen3-VL-235B-A22B-Instruct` | 250000 | 8192 | false | text + image |
| `hf:zai-org/GLM-4.5` | 128000 | 128000 | false | text |
| `hf:zai-org/GLM-4.6` | 198000 | 128000 | false | text |
| `hf:deepseek-ai/DeepSeek-V3` | 128000 | 8192 | false | text |
| `hf:Qwen/Qwen3-235B-A22B-Thinking-2507` | 256000 | 8192 | true | text |
## Notes

View File

@@ -4,6 +4,7 @@ read_when:
- You want privacy-focused inference in OpenClaw
- You want Venice AI setup guidance
---
# Venice AI (Venice highlight)
**Venice** is our highlight Venice setup for privacy-first inference with optional anonymized access to proprietary models.
@@ -21,10 +22,10 @@ Venice AI provides privacy-focused AI inference with support for uncensored mode
Venice offers two privacy levels — understanding this is key to choosing your model:
| Mode | Description | Models |
|------|-------------|--------|
| **Private** | Fully private. Prompts/responses are **never stored or logged**. Ephemeral. | Llama, Qwen, DeepSeek, Venice Uncensored, etc. |
| **Anonymized** | Proxied through Venice with metadata stripped. The underlying provider (OpenAI, Anthropic) sees anonymized requests. | Claude, GPT, Gemini, Grok, Kimi, MiniMax |
| Mode | Description | Models |
| -------------- | -------------------------------------------------------------------------------------------------------------------- | ---------------------------------------------- |
| **Private** | Fully private. Prompts/responses are **never stored or logged**. Ephemeral. | Llama, Qwen, DeepSeek, Venice Uncensored, etc. |
| **Anonymized** | Proxied through Venice with metadata stripped. The underlying provider (OpenAI, Anthropic) sees anonymized requests. | Claude, GPT, Gemini, Grok, Kimi, MiniMax |
## Features
@@ -60,6 +61,7 @@ openclaw onboard --auth-choice venice-api-key
```
This will:
1. Prompt for your API key (or use existing `VENICE_API_KEY`)
2. Show all available Venice models
3. Let you pick your default model
@@ -109,53 +111,53 @@ openclaw models list | grep venice
## Which Model Should I Use?
| Use Case | Recommended Model | Why |
|----------|-------------------|-----|
| **General chat** | `llama-3.3-70b` | Good all-around, fully private |
| **Best overall quality** | `claude-opus-45` | Opus remains the strongest for hard tasks |
| **Privacy + Claude quality** | `claude-opus-45` | Best reasoning via anonymized proxy |
| **Coding** | `qwen3-coder-480b-a35b-instruct` | Code-optimized, 262k context |
| **Vision tasks** | `qwen3-vl-235b-a22b` | Best private vision model |
| **Uncensored** | `venice-uncensored` | No content restrictions |
| **Fast + cheap** | `qwen3-4b` | Lightweight, still capable |
| **Complex reasoning** | `deepseek-v3.2` | Strong reasoning, private |
| Use Case | Recommended Model | Why |
| ---------------------------- | -------------------------------- | ----------------------------------------- |
| **General chat** | `llama-3.3-70b` | Good all-around, fully private |
| **Best overall quality** | `claude-opus-45` | Opus remains the strongest for hard tasks |
| **Privacy + Claude quality** | `claude-opus-45` | Best reasoning via anonymized proxy |
| **Coding** | `qwen3-coder-480b-a35b-instruct` | Code-optimized, 262k context |
| **Vision tasks** | `qwen3-vl-235b-a22b` | Best private vision model |
| **Uncensored** | `venice-uncensored` | No content restrictions |
| **Fast + cheap** | `qwen3-4b` | Lightweight, still capable |
| **Complex reasoning** | `deepseek-v3.2` | Strong reasoning, private |
## Available Models (25 Total)
### Private Models (15) — Fully Private, No Logging
| Model ID | Name | Context (tokens) | Features |
|----------|------|------------------|----------|
| `llama-3.3-70b` | Llama 3.3 70B | 131k | General |
| `llama-3.2-3b` | Llama 3.2 3B | 131k | Fast, lightweight |
| `hermes-3-llama-3.1-405b` | Hermes 3 Llama 3.1 405B | 131k | Complex tasks |
| `qwen3-235b-a22b-thinking-2507` | Qwen3 235B Thinking | 131k | Reasoning |
| `qwen3-235b-a22b-instruct-2507` | Qwen3 235B Instruct | 131k | General |
| `qwen3-coder-480b-a35b-instruct` | Qwen3 Coder 480B | 262k | Code |
| `qwen3-next-80b` | Qwen3 Next 80B | 262k | General |
| `qwen3-vl-235b-a22b` | Qwen3 VL 235B | 262k | Vision |
| `qwen3-4b` | Venice Small (Qwen3 4B) | 32k | Fast, reasoning |
| `deepseek-v3.2` | DeepSeek V3.2 | 163k | Reasoning |
| `venice-uncensored` | Venice Uncensored | 32k | Uncensored |
| `mistral-31-24b` | Venice Medium (Mistral) | 131k | Vision |
| `google-gemma-3-27b-it` | Gemma 3 27B Instruct | 202k | Vision |
| `openai-gpt-oss-120b` | OpenAI GPT OSS 120B | 131k | General |
| `zai-org-glm-4.7` | GLM 4.7 | 202k | Reasoning, multilingual |
| Model ID | Name | Context (tokens) | Features |
| -------------------------------- | ----------------------- | ---------------- | ----------------------- |
| `llama-3.3-70b` | Llama 3.3 70B | 131k | General |
| `llama-3.2-3b` | Llama 3.2 3B | 131k | Fast, lightweight |
| `hermes-3-llama-3.1-405b` | Hermes 3 Llama 3.1 405B | 131k | Complex tasks |
| `qwen3-235b-a22b-thinking-2507` | Qwen3 235B Thinking | 131k | Reasoning |
| `qwen3-235b-a22b-instruct-2507` | Qwen3 235B Instruct | 131k | General |
| `qwen3-coder-480b-a35b-instruct` | Qwen3 Coder 480B | 262k | Code |
| `qwen3-next-80b` | Qwen3 Next 80B | 262k | General |
| `qwen3-vl-235b-a22b` | Qwen3 VL 235B | 262k | Vision |
| `qwen3-4b` | Venice Small (Qwen3 4B) | 32k | Fast, reasoning |
| `deepseek-v3.2` | DeepSeek V3.2 | 163k | Reasoning |
| `venice-uncensored` | Venice Uncensored | 32k | Uncensored |
| `mistral-31-24b` | Venice Medium (Mistral) | 131k | Vision |
| `google-gemma-3-27b-it` | Gemma 3 27B Instruct | 202k | Vision |
| `openai-gpt-oss-120b` | OpenAI GPT OSS 120B | 131k | General |
| `zai-org-glm-4.7` | GLM 4.7 | 202k | Reasoning, multilingual |
### Anonymized Models (10) — Via Venice Proxy
| Model ID | Original | Context (tokens) | Features |
|----------|----------|------------------|----------|
| `claude-opus-45` | Claude Opus 4.5 | 202k | Reasoning, vision |
| `claude-sonnet-45` | Claude Sonnet 4.5 | 202k | Reasoning, vision |
| `openai-gpt-52` | GPT-5.2 | 262k | Reasoning |
| `openai-gpt-52-codex` | GPT-5.2 Codex | 262k | Reasoning, vision |
| `gemini-3-pro-preview` | Gemini 3 Pro | 202k | Reasoning, vision |
| `gemini-3-flash-preview` | Gemini 3 Flash | 262k | Reasoning, vision |
| `grok-41-fast` | Grok 4.1 Fast | 262k | Reasoning, vision |
| `grok-code-fast-1` | Grok Code Fast 1 | 262k | Reasoning, code |
| `kimi-k2-thinking` | Kimi K2 Thinking | 262k | Reasoning |
| `minimax-m21` | MiniMax M2.1 | 202k | Reasoning |
| Model ID | Original | Context (tokens) | Features |
| ------------------------ | ----------------- | ---------------- | ----------------- |
| `claude-opus-45` | Claude Opus 4.5 | 202k | Reasoning, vision |
| `claude-sonnet-45` | Claude Sonnet 4.5 | 202k | Reasoning, vision |
| `openai-gpt-52` | GPT-5.2 | 262k | Reasoning |
| `openai-gpt-52-codex` | GPT-5.2 Codex | 262k | Reasoning, vision |
| `gemini-3-pro-preview` | Gemini 3 Pro | 202k | Reasoning, vision |
| `gemini-3-flash-preview` | Gemini 3 Flash | 262k | Reasoning, vision |
| `grok-41-fast` | Grok 4.1 Fast | 262k | Reasoning, vision |
| `grok-code-fast-1` | Grok Code Fast 1 | 262k | Reasoning, code |
| `kimi-k2-thinking` | Kimi K2 Thinking | 262k | Reasoning |
| `minimax-m21` | MiniMax M2.1 | 202k | Reasoning |
## Model Discovery
@@ -165,12 +167,12 @@ The `/models` endpoint is public (no auth needed for listing), but inference req
## Streaming & Tool Support
| Feature | Support |
|---------|---------|
| **Streaming** | ✅ All models |
| Feature | Support |
| -------------------- | ------------------------------------------------------- |
| **Streaming** | ✅ All models |
| **Function calling** | ✅ Most models (check `supportsFunctionCalling` in API) |
| **Vision/Images** | ✅ Models marked with "Vision" feature |
| **JSON mode** | ✅ Supported via `response_format` |
| **Vision/Images** | ✅ Models marked with "Vision" feature |
| **JSON mode** | ✅ Supported via `response_format` |
## Pricing
@@ -181,12 +183,12 @@ Venice uses a credit-based system. Check [venice.ai/pricing](https://venice.ai/p
## Comparison: Venice vs Direct API
| Aspect | Venice (Anonymized) | Direct API |
|--------|---------------------|------------|
| **Privacy** | Metadata stripped, anonymized | Your account linked |
| **Latency** | +10-50ms (proxy) | Direct |
| **Features** | Most features supported | Full features |
| **Billing** | Venice credits | Provider billing |
| Aspect | Venice (Anonymized) | Direct API |
| ------------ | ----------------------------- | ------------------- |
| **Privacy** | Metadata stripped, anonymized | Your account linked |
| **Latency** | +10-50ms (proxy) | Direct |
| **Features** | Most features supported | Full features |
| **Billing** | Venice credits | Provider billing |
## Usage Examples
@@ -247,12 +249,12 @@ Venice API is at `https://api.venice.ai/api/v1`. Ensure your network allows HTTP
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 131072,
maxTokens: 8192
}
]
}
}
}
maxTokens: 8192,
},
],
},
},
},
}
```

View File

@@ -5,10 +5,10 @@ read_when:
- You want to use Vercel AI Gateway with OpenClaw
- You need the API key env var or CLI auth choice
---
# Vercel AI Gateway
The [Vercel AI Gateway](https://vercel.com/ai-gateway) provides a unified API to access hundreds of models through a single endpoint.
The [Vercel AI Gateway](https://vercel.com/ai-gateway) provides a unified API to access hundreds of models through a single endpoint.
- Provider: `vercel-ai-gateway`
- Auth: `AI_GATEWAY_API_KEY`
@@ -16,21 +16,21 @@ The [Vercel AI Gateway](https://vercel.com/ai-gateway) provides a unified API to
## Quick start
1) Set the API key (recommended: store it for the Gateway):
1. Set the API key (recommended: store it for the Gateway):
```bash
openclaw onboard --auth-choice ai-gateway-api-key
```
2) Set a default model:
2. Set a default model:
```json5
{
agents: {
defaults: {
model: { primary: "vercel-ai-gateway/anthropic/claude-opus-4.5" }
}
}
model: { primary: "vercel-ai-gateway/anthropic/claude-opus-4.5" },
},
},
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want Xiaomi MiMo models in OpenClaw
- You need XIAOMI_API_KEY setup
---
# Xiaomi MiMo
Xiaomi MiMo is the API platform for **MiMo** models. It provides REST APIs compatible with
@@ -46,12 +47,12 @@ openclaw onboard --auth-choice xiaomi-api-key --xiaomi-api-key "$XIAOMI_API_KEY"
input: ["text"],
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
contextWindow: 262144,
maxTokens: 8192
}
]
}
}
}
maxTokens: 8192,
},
],
},
},
},
}
```

View File

@@ -4,6 +4,7 @@ read_when:
- You want Z.AI / GLM models in OpenClaw
- You need a simple ZAI_API_KEY setup
---
# Z.AI
Z.AI is the API platform for **GLM** models. It provides REST APIs for GLM and uses API keys
@@ -23,7 +24,7 @@ openclaw onboard --zai-api-key "$ZAI_API_KEY"
```json5
{
env: { ZAI_API_KEY: "sk-..." },
agents: { defaults: { model: { primary: "zai/glm-4.7" } } }
agents: { defaults: { model: { primary: "zai/glm-4.7" } } },
}
```