feat(fleet): promote Ollama to first-class provider, assign Gemma 4 across fleet
Some checks failed
Deploy Nexus / deploy (push) Failing after 3s
Staging Verification Gate / verify-staging (push) Failing after 3s

- lazarus-registry.yaml: replace big_brain/RunPod with local ollama/gemma4:12b
- fleet-routing.json: assign ollama:gemma4:12b to carnice, bilbobagginshire, substratum
- intelligence/deepdive/config.yaml: local model -> gemma4:12b
This commit is contained in:
2026-04-07 15:55:52 +00:00
parent 5275c96e52
commit 34862cf5e5
3 changed files with 11 additions and 11 deletions

View File

@@ -9,7 +9,7 @@
"id": 27, "id": 27,
"name": "carnice", "name": "carnice",
"gitea_user": "carnice", "gitea_user": "carnice",
"model": "qwen3.5-9b", "model": "ollama:gemma4:12b",
"tier": "free", "tier": "free",
"location": "Local Metal", "location": "Local Metal",
"description": "Local Hermes agent, fine-tuned on Hermes traces. Runs on local hardware.", "description": "Local Hermes agent, fine-tuned on Hermes traces. Runs on local hardware.",
@@ -41,7 +41,7 @@
"id": 25, "id": 25,
"name": "bilbobagginshire", "name": "bilbobagginshire",
"gitea_user": "bilbobagginshire", "gitea_user": "bilbobagginshire",
"model": "ollama", "model": "ollama:gemma4:12b",
"tier": "free", "tier": "free",
"location": "Bag End, The Shire (VPS)", "location": "Bag End, The Shire (VPS)",
"description": "Ollama on VPS. Speaks when spoken to. Prefers quiet. Not for delegated work.", "description": "Ollama on VPS. Speaks when spoken to. Prefers quiet. Not for delegated work.",
@@ -74,7 +74,7 @@
"id": 23, "id": 23,
"name": "substratum", "name": "substratum",
"gitea_user": "substratum", "gitea_user": "substratum",
"model": "unassigned", "model": "ollama:gemma4:12b",
"tier": "unknown", "tier": "unknown",
"location": "Below the Surface", "location": "Below the Surface",
"description": "Infrastructure, deployments, bedrock services. Needs model assignment before activation.", "description": "Infrastructure, deployments, bedrock services. Needs model assignment before activation.",

View File

@@ -76,7 +76,7 @@ deepdive:
# Phase 3: Synthesis # Phase 3: Synthesis
synthesis: synthesis:
llm_endpoint: "http://localhost:4000/v1" # Local llama-server llm_endpoint: "http://localhost:4000/v1" # Local llama-server
llm_model: "gemma-4-it" llm_model: "gemma4:12b"
max_summary_length: 800 max_summary_length: 800
temperature: 0.7 temperature: 0.7

View File

@@ -20,8 +20,8 @@ fleet:
- provider: openrouter - provider: openrouter
model: anthropic/claude-sonnet-4-20250514 model: anthropic/claude-sonnet-4-20250514
timeout: 120 timeout: 120
- provider: big_brain - provider: ollama
model: gemma3:27b-instruct-q8_0 model: gemma4:12b
timeout: 300 timeout: 300
health_endpoints: health_endpoints:
gateway: http://127.0.0.1:8646 gateway: http://127.0.0.1:8646
@@ -100,11 +100,11 @@ provider_health_matrix:
last_checked: '2026-04-07T02:55:00Z' last_checked: '2026-04-07T02:55:00Z'
rate_limited: false rate_limited: false
dead: false dead: false
big_brain: ollama:
status: provisioning status: healthy
note: RunPod L40S instance big-brain-bezalel deployed; Ollama endpoint propagating note: Local Ollama endpoint with Gemma 4 support
last_checked: '2026-04-07T02:55:00Z' last_checked: '2026-04-07T15:09:53.385047+00:00'
endpoint: http://yxw29g3excyddq-64411cd0-11434.tcp.runpod.net:11434/v1 endpoint: http://localhost:11434/v1
rate_limited: false rate_limited: false
dead: false dead: false
timeout_policies: timeout_policies: