feat(fleet): promote Ollama to first-class provider, assign Gemma 4 across fleet
Some checks failed
Deploy Nexus / deploy (push) Failing after 3s
Staging Verification Gate / verify-staging (push) Failing after 3s

- lazarus-registry.yaml: replace big_brain/RunPod with local ollama/gemma4:12b
- fleet-routing.json: assign ollama:gemma4:12b to carnice, bilbobagginshire, substratum
- intelligence/deepdive/config.yaml: local model -> gemma4:12b
This commit is contained in:
2026-04-07 15:55:52 +00:00
parent 5275c96e52
commit 34862cf5e5
3 changed files with 11 additions and 11 deletions

View File

@@ -9,7 +9,7 @@
"id": 27,
"name": "carnice",
"gitea_user": "carnice",
"model": "qwen3.5-9b",
"model": "ollama:gemma4:12b",
"tier": "free",
"location": "Local Metal",
"description": "Local Hermes agent, fine-tuned on Hermes traces. Runs on local hardware.",
@@ -41,7 +41,7 @@
"id": 25,
"name": "bilbobagginshire",
"gitea_user": "bilbobagginshire",
"model": "ollama",
"model": "ollama:gemma4:12b",
"tier": "free",
"location": "Bag End, The Shire (VPS)",
"description": "Ollama on VPS. Speaks when spoken to. Prefers quiet. Not for delegated work.",
@@ -74,7 +74,7 @@
"id": 23,
"name": "substratum",
"gitea_user": "substratum",
"model": "unassigned",
"model": "ollama:gemma4:12b",
"tier": "unknown",
"location": "Below the Surface",
"description": "Infrastructure, deployments, bedrock services. Needs model assignment before activation.",

View File

@@ -76,7 +76,7 @@ deepdive:
# Phase 3: Synthesis
synthesis:
llm_endpoint: "http://localhost:4000/v1" # Local llama-server
llm_model: "gemma-4-it"
llm_model: "gemma4:12b"
max_summary_length: 800
temperature: 0.7

View File

@@ -20,8 +20,8 @@ fleet:
- provider: openrouter
model: anthropic/claude-sonnet-4-20250514
timeout: 120
- provider: big_brain
model: gemma3:27b-instruct-q8_0
- provider: ollama
model: gemma4:12b
timeout: 300
health_endpoints:
gateway: http://127.0.0.1:8646
@@ -100,11 +100,11 @@ provider_health_matrix:
last_checked: '2026-04-07T02:55:00Z'
rate_limited: false
dead: false
big_brain:
status: provisioning
note: RunPod L40S instance big-brain-bezalel deployed; Ollama endpoint propagating
last_checked: '2026-04-07T02:55:00Z'
endpoint: http://yxw29g3excyddq-64411cd0-11434.tcp.runpod.net:11434/v1
ollama:
status: healthy
note: Local Ollama endpoint with Gemma 4 support
last_checked: '2026-04-07T15:09:53.385047+00:00'
endpoint: http://localhost:11434/v1
rate_limited: false
dead: false
timeout_policies: