feat: standardize llama.cpp backend for sovereign local inference (#1123)
Some checks failed
Deploy Nexus / deploy (push) Has been cancelled
Staging Verification Gate / verify-staging (push) Has been cancelled

This commit is contained in:
2026-04-14 01:48:26 +00:00
parent ed79826608
commit e27c51c6da

View File

@@ -1,4 +1,4 @@
"""llama_provider.py — Hermes inference router provider for llama.cpp local server."""
"""llama_provider.py — Hermes inference router provider for llama.cpp."""
import logging, os, time
from dataclasses import dataclass
from typing import Optional