[POLICY] RunPod Serverless Mandate — No Local Llama for Models >5GB #409
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Directive
Hard disk is filling up repeatedly because we keep downloading large models to run local llama-server. This is unsustainable.
New Policy
Acceptance Criteria
wontfixwith a comment linking to this policy.localhost/127.0.0.1.Obsolete Issues to Close
Owner
@allegro — execute the burn-down.
Initial Burn-Down Complete — 11 Issues Closed
Ezra closed the following obsolete issues per this policy:
timmy-home#346— Local-First Claw Code: Llama 1.5B-7B + Telegram + TV Code OAuthtimmy-home#307— VPS Ollama stuck in swap death spiraltimmy-home#297— Bilbo Memory Recovery - Free 2GB+ for Ollamahermes-agent#104— Morrowind: Local brain parity — Hermes 4 14B vs Cloud Claudeturboquant#30— Epic: TurboQuant + Gemma 4 Local Mac Deploymentturboquant#31— Download Gemma 4 via Ollama on Macturboquant#27— Build TheTom/llama-cpp-turboquant for M3 Max Metalturboquant#26— Production cutover: swap Timmy's llama-server to TurboQuantturboquant#25— Download qwen3.5:27b and benchmark turbo4turboquant#23— Fix Ollama install and build custom Ollama with TurboQuant forkturboquant#1— TurboQuant — KV Cache Compression for Local Inference on M4 MaxRemaining work: disk audit, model deletion, RunPod endpoint creation, profile updates.