feat: standardize llama.cpp backend for sovereign local inference (#1123)
This commit is contained in:
@@ -2,7 +2,6 @@
|
||||
"""llama_client.py — OpenAI-compatible client for llama.cpp HTTP API."""
|
||||
import argparse, json, os, sys, time
|
||||
from dataclasses import dataclass
|
||||
from typing import Generator, Optional
|
||||
import urllib.request, urllib.error
|
||||
|
||||
DEFAULT_ENDPOINT = os.environ.get("LLAMA_ENDPOINT", "http://localhost:11435")
|
||||
|
||||
Reference in New Issue
Block a user