[P2-3] Fix Ollama install and build custom Ollama with TurboQuant fork #23
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Parent: #1 | Depends on: P2-1 passing quality gate
Why
Timmy runs inference through Ollama. The TurboQuant fork's
llama-serverworks standalone but isn't wired into the stack Timmy actually uses. Until Ollama speaks turbo4, the wand isn't in Timmy's hand.The Problem (from Phase 1 report)
Steps
Alternative: Skip Ollama, use llama-server directly
If the Ollama build is too complex, the fork's
llama-serverbinary is a drop-in replacement:Hermes config already points at localhost:8081. This would work TODAY.
Acceptance Criteria
The fast path is the llama-server alternative. Do that first, Ollama build second.
Closed per new fleet policy: no local llama-server for models >5GB. RunPod serverless endpoints only. See Timmy_Foundation/timmy-home#409.