From 809abd60bf1c182dbed4a5ddaad61c0701c68084 Mon Sep 17 00:00:00 2001 From: teknium1 Date: Thu, 12 Mar 2026 17:28:36 -0700 Subject: [PATCH] docs: add Anthropic provider to all documentation pages - quickstart.md: Add Anthropic to the provider comparison table - configuration.md: Add Anthropic to provider list table, add full 'Anthropic (Native)' section with three auth methods (API key, setup-token, Claude Code auto-detect), config.yaml example, and provider alias tip - environment-variables.md: Add ANTHROPIC_API_KEY, ANTHROPIC_TOKEN, CLAUDE_CODE_OAUTH_TOKEN to LLM Providers table; add 'anthropic' to HERMES_INFERENCE_PROVIDER values list --- website/docs/getting-started/quickstart.md | 1 + .../docs/reference/environment-variables.md | 5 +++- website/docs/user-guide/configuration.md | 29 +++++++++++++++++++ 3 files changed, 34 insertions(+), 1 deletion(-) diff --git a/website/docs/getting-started/quickstart.md b/website/docs/getting-started/quickstart.md index 832ba0be5..eceaf73de 100644 --- a/website/docs/getting-started/quickstart.md +++ b/website/docs/getting-started/quickstart.md @@ -43,6 +43,7 @@ hermes setup # Or configure everything at once |----------|-----------|---------------| | **Nous Portal** | Subscription-based, zero-config | OAuth login via `hermes model` | | **OpenAI Codex** | ChatGPT OAuth, uses Codex models | Device code auth via `hermes model` | +| **Anthropic** | Claude models directly (Pro/Max or API key) | API key or Claude Code setup-token | | **OpenRouter** | 200+ models, pay-per-use | Enter your API key | | **Custom Endpoint** | VLLM, SGLang, any OpenAI-compatible API | Set base URL + API key | diff --git a/website/docs/reference/environment-variables.md b/website/docs/reference/environment-variables.md index 26a0683e3..b93108b44 100644 --- a/website/docs/reference/environment-variables.md +++ b/website/docs/reference/environment-variables.md @@ -23,6 +23,9 @@ All variables go in `~/.hermes/.env`. You can also set them with `hermes config | `MINIMAX_BASE_URL` | Override MiniMax base URL (default: `https://api.minimax.io/v1`) | | `MINIMAX_CN_API_KEY` | MiniMax API key — China endpoint ([minimaxi.com](https://www.minimaxi.com)) | | `MINIMAX_CN_BASE_URL` | Override MiniMax China base URL (default: `https://api.minimaxi.com/v1`) | +| `ANTHROPIC_API_KEY` | Anthropic API key or setup-token ([console.anthropic.com](https://console.anthropic.com/)) | +| `ANTHROPIC_TOKEN` | Anthropic OAuth/setup token (alternative to `ANTHROPIC_API_KEY`) | +| `CLAUDE_CODE_OAUTH_TOKEN` | Claude Code setup-token (same as `ANTHROPIC_TOKEN`) | | `HERMES_MODEL` | Preferred model name (checked before `LLM_MODEL`, used by gateway) | | `LLM_MODEL` | Default model name (fallback when not set in config.yaml) | | `VOICE_TOOLS_OPENAI_KEY` | OpenAI key for TTS and voice transcription (separate from custom endpoint) | @@ -32,7 +35,7 @@ All variables go in `~/.hermes/.env`. You can also set them with `hermes config | Variable | Description | |----------|-------------| -| `HERMES_INFERENCE_PROVIDER` | Override provider selection: `auto`, `openrouter`, `nous`, `zai`, `kimi-coding`, `minimax`, `minimax-cn` (default: `auto`) | +| `HERMES_INFERENCE_PROVIDER` | Override provider selection: `auto`, `openrouter`, `nous`, `anthropic`, `zai`, `kimi-coding`, `minimax`, `minimax-cn` (default: `auto`) | | `HERMES_PORTAL_BASE_URL` | Override Nous Portal URL (for development/testing) | | `NOUS_INFERENCE_BASE_URL` | Override Nous inference API URL | | `HERMES_NOUS_MIN_KEY_TTL_SECONDS` | Min agent key TTL before re-mint (default: 1800 = 30min) | diff --git a/website/docs/user-guide/configuration.md b/website/docs/user-guide/configuration.md index 7dad284e6..a771b50b5 100644 --- a/website/docs/user-guide/configuration.md +++ b/website/docs/user-guide/configuration.md @@ -63,6 +63,7 @@ You need at least one way to connect to an LLM. Use `hermes model` to switch pro |----------|-------| | **Nous Portal** | `hermes model` (OAuth, subscription-based) | | **OpenAI Codex** | `hermes model` (ChatGPT OAuth, uses Codex models) | +| **Anthropic** | `hermes model` (API key, setup-token, or Claude Code auto-detect) | | **OpenRouter** | `OPENROUTER_API_KEY` in `~/.hermes/.env` | | **z.ai / GLM** | `GLM_API_KEY` in `~/.hermes/.env` (provider: `zai`) | | **Kimi / Moonshot** | `KIMI_API_KEY` in `~/.hermes/.env` (provider: `kimi-coding`) | @@ -78,6 +79,34 @@ The OpenAI Codex provider authenticates via device code (open a URL, enter a cod Even when using Nous Portal, Codex, or a custom endpoint, some tools (vision, web summarization, MoA) use a separate "auxiliary" model — by default Gemini Flash via OpenRouter. An `OPENROUTER_API_KEY` enables these tools automatically. You can also configure which model and provider these tools use — see [Auxiliary Models](#auxiliary-models) below. ::: +### Anthropic (Native) + +Use Claude models directly through the Anthropic API — no OpenRouter proxy needed. Supports three auth methods: + +```bash +# With an API key (pay-per-token) +export ANTHROPIC_API_KEY=sk-ant-api03-... +hermes chat --provider anthropic --model claude-sonnet-4-6 + +# With a Claude Code setup-token (Pro/Max subscription) +export ANTHROPIC_API_KEY=sk-ant-oat01-... # from 'claude setup-token' +hermes chat --provider anthropic + +# Auto-detect Claude Code credentials (if you have Claude Code installed) +hermes chat --provider anthropic # reads ~/.claude.json automatically +``` + +Or set it permanently: +```yaml +model: + provider: "anthropic" + default: "claude-sonnet-4-6" +``` + +:::tip Aliases +`--provider claude` and `--provider claude-code` also work as shorthand for `--provider anthropic`. +::: + ### First-Class Chinese AI Providers These providers have built-in support with dedicated provider IDs. Set the API key and use `--provider` to select: