feat: standardize llama.cpp backend for sovereign local inference (#1123)
This commit is contained in:
@@ -1,4 +1,4 @@
|
||||
"""llama_provider.py — Hermes inference router provider for llama.cpp local server."""
|
||||
"""llama_provider.py — Hermes inference router provider for llama.cpp."""
|
||||
import logging, os, time
|
||||
from dataclasses import dataclass
|
||||
from typing import Optional
|
||||
|
||||
Reference in New Issue
Block a user