From 93ef328ff3723aa56955f183e15dabed76a0c4b8 Mon Sep 17 00:00:00 2001 From: Alexander Whitestone Date: Tue, 24 Mar 2026 10:26:54 -0400 Subject: [PATCH] feat: onboard ollama as local agent in status board (#397) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Add ollama to the AGENT_STATUS_STUB in the 3D agent panel display. Ollama runs LLMs locally (OpenAI-compatible API at localhost:11434), so it is marked local: true — consistent with how claude is tagged. Refs #397 --- app.js | 1 + 1 file changed, 1 insertion(+) diff --git a/app.js b/app.js index 61a2f03..3ad6292 100644 --- a/app.js +++ b/app.js @@ -4000,6 +4000,7 @@ const AGENT_STATUS_STUB = { { name: 'kimi', status: 'working', issue: 'Portal system YAML registry (#5)', prs_today: 2, local: false }, { name: 'groq', status: 'idle', issue: null, prs_today: 0, local: false }, { name: 'grok', status: 'dead', issue: null, prs_today: 0, local: false }, + { name: 'ollama', status: 'idle', issue: null, prs_today: 0, local: true }, ] }; -- 2.43.0