feat: standardize llama.cpp backend for sovereign local inference (#1123)
This commit is contained in:
@@ -46,5 +46,3 @@ Standardizes local LLM inference across the fleet using llama.cpp.
|
||||
- Slow → -t to core count
|
||||
- OOM → reduce -c
|
||||
- Port conflict → lsof -i :11435
|
||||
|
||||
See systemd/llama-server.service for production deployment.
|
||||
|
||||
Reference in New Issue
Block a user