forked from Rockachopa/Timmy-time-dashboard
Compare commits
5 Commits
kimi/issue
...
kimi/issue
| Author | SHA1 | Date | |
|---|---|---|---|
| 4d8272f64e | |||
| d70e4f810a | |||
|
|
0b81a01c90 | ||
| 7f20742fcf | |||
| 15eb7c3b45 |
@@ -54,19 +54,6 @@ providers:
|
|||||||
context_window: 2048
|
context_window: 2048
|
||||||
capabilities: [text, vision, streaming]
|
capabilities: [text, vision, streaming]
|
||||||
|
|
||||||
# Secondary: Local AirLLM (if installed)
|
|
||||||
- name: airllm-local
|
|
||||||
type: airllm
|
|
||||||
enabled: false # Enable if pip install airllm
|
|
||||||
priority: 2
|
|
||||||
models:
|
|
||||||
- name: 70b
|
|
||||||
default: true
|
|
||||||
capabilities: [text, tools, json, streaming]
|
|
||||||
- name: 8b
|
|
||||||
capabilities: [text, tools, json, streaming]
|
|
||||||
- name: 405b
|
|
||||||
capabilities: [text, tools, json, streaming]
|
|
||||||
|
|
||||||
# Tertiary: OpenAI (if API key available)
|
# Tertiary: OpenAI (if API key available)
|
||||||
- name: openai-backup
|
- name: openai-backup
|
||||||
|
|||||||
@@ -100,7 +100,7 @@ class CSRFMiddleware(BaseHTTPMiddleware):
|
|||||||
...
|
...
|
||||||
|
|
||||||
Usage:
|
Usage:
|
||||||
app.add_middleware(CSRFMiddleware, secret="your-secret-key")
|
app.add_middleware(CSRFMiddleware, secret=settings.csrf_secret)
|
||||||
|
|
||||||
Attributes:
|
Attributes:
|
||||||
secret: Secret key for token signing (optional, for future use).
|
secret: Secret key for token signing (optional, for future use).
|
||||||
|
|||||||
@@ -18,6 +18,8 @@ from enum import Enum
|
|||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Any
|
from typing import Any
|
||||||
|
|
||||||
|
from config import settings
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import yaml
|
import yaml
|
||||||
except ImportError:
|
except ImportError:
|
||||||
@@ -100,7 +102,7 @@ class Provider:
|
|||||||
"""LLM provider configuration and state."""
|
"""LLM provider configuration and state."""
|
||||||
|
|
||||||
name: str
|
name: str
|
||||||
type: str # ollama, openai, anthropic, airllm
|
type: str # ollama, openai, anthropic
|
||||||
enabled: bool
|
enabled: bool
|
||||||
priority: int
|
priority: int
|
||||||
url: str | None = None
|
url: str | None = None
|
||||||
@@ -301,22 +303,13 @@ class CascadeRouter:
|
|||||||
# Can't check without requests, assume available
|
# Can't check without requests, assume available
|
||||||
return True
|
return True
|
||||||
try:
|
try:
|
||||||
url = provider.url or "http://localhost:11434"
|
url = provider.url or settings.ollama_url
|
||||||
response = requests.get(f"{url}/api/tags", timeout=5)
|
response = requests.get(f"{url}/api/tags", timeout=5)
|
||||||
return response.status_code == 200
|
return response.status_code == 200
|
||||||
except Exception as exc:
|
except Exception as exc:
|
||||||
logger.debug("Ollama provider check error: %s", exc)
|
logger.debug("Ollama provider check error: %s", exc)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
elif provider.type == "airllm":
|
|
||||||
# Check if airllm is installed
|
|
||||||
try:
|
|
||||||
import importlib.util
|
|
||||||
|
|
||||||
return importlib.util.find_spec("airllm") is not None
|
|
||||||
except (ImportError, ModuleNotFoundError):
|
|
||||||
return False
|
|
||||||
|
|
||||||
elif provider.type in ("openai", "anthropic", "grok"):
|
elif provider.type in ("openai", "anthropic", "grok"):
|
||||||
# Check if API key is set
|
# Check if API key is set
|
||||||
return provider.api_key is not None and provider.api_key != ""
|
return provider.api_key is not None and provider.api_key != ""
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
|
|
||||||
import time
|
import time
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from unittest.mock import AsyncMock, MagicMock, patch
|
from unittest.mock import AsyncMock, patch
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
import yaml
|
import yaml
|
||||||
@@ -489,34 +489,6 @@ class TestProviderAvailabilityCheck:
|
|||||||
|
|
||||||
assert router._check_provider_available(provider) is False
|
assert router._check_provider_available(provider) is False
|
||||||
|
|
||||||
def test_check_airllm_installed(self):
|
|
||||||
"""Test AirLLM when installed."""
|
|
||||||
router = CascadeRouter(config_path=Path("/nonexistent"))
|
|
||||||
|
|
||||||
provider = Provider(
|
|
||||||
name="airllm",
|
|
||||||
type="airllm",
|
|
||||||
enabled=True,
|
|
||||||
priority=1,
|
|
||||||
)
|
|
||||||
|
|
||||||
with patch("importlib.util.find_spec", return_value=MagicMock()):
|
|
||||||
assert router._check_provider_available(provider) is True
|
|
||||||
|
|
||||||
def test_check_airllm_not_installed(self):
|
|
||||||
"""Test AirLLM when not installed."""
|
|
||||||
router = CascadeRouter(config_path=Path("/nonexistent"))
|
|
||||||
|
|
||||||
provider = Provider(
|
|
||||||
name="airllm",
|
|
||||||
type="airllm",
|
|
||||||
enabled=True,
|
|
||||||
priority=1,
|
|
||||||
)
|
|
||||||
|
|
||||||
with patch("importlib.util.find_spec", return_value=None):
|
|
||||||
assert router._check_provider_available(provider) is False
|
|
||||||
|
|
||||||
|
|
||||||
class TestCascadeRouterReload:
|
class TestCascadeRouterReload:
|
||||||
"""Test hot-reload of providers.yaml."""
|
"""Test hot-reload of providers.yaml."""
|
||||||
|
|||||||
Reference in New Issue
Block a user