diff --git a/config.yaml b/config.yaml index 1dcd03fd..9c1d1c4d 100644 --- a/config.yaml +++ b/config.yaml @@ -180,6 +180,10 @@ custom_providers: base_url: http://localhost:11434/v1 api_key: ollama model: glm-4.7-flash:latest +- name: Google Gemini + base_url: https://generativelanguage.googleapis.com/v1beta/openai + api_key_env: GEMINI_API_KEY + model: gemini-2.5-pro system_prompt_suffix: "You are Timmy. Your soul is defined in SOUL.md \u2014 read\ \ it, live it.\nYou run locally on your owner's machine via Ollama. You never phone\ \ home.\nYou speak plainly. You prefer short sentences. Brevity is a kindness.\n\ @@ -192,40 +196,15 @@ providers: ollama: base_url: http://localhost:11434/v1 model: hermes3:latest - -# ── Fallback Model ──────────────────────────────────────────────────── -# Automatic provider failover when primary is unavailable. -# Uncomment and configure to enable. Triggers on rate limits (429), -# overload (529), service errors (503), or connection failures. -# -# Supported providers: -# openrouter (OPENROUTER_API_KEY) — routes to any model -# openai-codex (OAuth — hermes login) — OpenAI Codex -# nous (OAuth — hermes login) — Nous Portal -# zai (ZAI_API_KEY) — Z.AI / GLM -# kimi-coding (KIMI_API_KEY) — Kimi / Moonshot -# minimax (MINIMAX_API_KEY) — MiniMax -# minimax-cn (MINIMAX_CN_API_KEY) — MiniMax (China) -# -# For custom OpenAI-compatible endpoints, add base_url and api_key_env. -# -# fallback_model: -# provider: openrouter -# model: anthropic/claude-sonnet-4 -# -# ── Smart Model Routing ──────────────────────────────────────────────── -# Optional cheap-vs-strong routing for simple turns. -# Keeps the primary model for complex work, but can route short/simple -# messages to a cheaper model across providers. -# -# smart_model_routing: -# enabled: true -# max_simple_chars: 160 -# max_simple_words: 28 -# cheap_model: -# provider: openrouter -# model: google/gemini-2.5-flash - -# Orchestration: Huey (replaces sovereign-orchestration repo) -# Start with: huey_consumer.py timmy-config.tasks.huey -# See orchestration.py + tasks.py +mcp_servers: + orchestration: + command: /Users/apayne/.hermes/hermes-agent/venv/bin/python3 + args: + - /Users/apayne/.hermes/hermes-agent/tools/orchestration_mcp_server.py + env: {} + timeout: 120 +fallback_model: + provider: custom + model: gemini-2.5-pro + base_url: https://generativelanguage.googleapis.com/v1beta/openai + api_key_env: GEMINI_API_KEY diff --git a/playbooks/issue-triager.yaml b/playbooks/issue-triager.yaml index d3f55fcc..adf31790 100644 --- a/playbooks/issue-triager.yaml +++ b/playbooks/issue-triager.yaml @@ -19,9 +19,7 @@ trigger: repos: - Timmy_Foundation/the-nexus - - Timmy_Foundation/autolora - - Timmy_Foundation/sovereign-orchestration - - Timmy_Foundation/timmy-config + - Timmy_Foundation/hermes-agent steps: - fetch_issues diff --git a/playbooks/pr-reviewer.yaml b/playbooks/pr-reviewer.yaml index def51ab4..5d694396 100644 --- a/playbooks/pr-reviewer.yaml +++ b/playbooks/pr-reviewer.yaml @@ -19,9 +19,7 @@ trigger: repos: - Timmy_Foundation/the-nexus - - Timmy_Foundation/autolora - - Timmy_Foundation/sovereign-orchestration - - Timmy_Foundation/timmy-config + - Timmy_Foundation/hermes-agent steps: - fetch_prs