forked from Rockachopa/Timmy-time-dashboard
Move 97 test files from flat tests/ into 13 subdirectories: tests/dashboard/ (8 files — routes, mobile, mission control) tests/swarm/ (17 files — coordinator, docker, routing, tasks) tests/timmy/ (12 files — agent, backends, CLI, tools) tests/self_coding/ (14 files — git safety, indexer, self-modify) tests/lightning/ (3 files — L402, LND, interface) tests/creative/ (8 files — assembler, director, image/music/video) tests/integrations/ (10 files — chat bridge, telegram, voice, websocket) tests/mcp/ (4 files — bootstrap, discovery, executor) tests/spark/ (3 files — engine, tools, events) tests/hands/ (3 files — registry, oracle, phase5) tests/scripture/ (1 file) tests/infrastructure/ (3 files — router cascade, API) tests/security/ (3 files — XSS, regression) Fix Path(__file__) reference in test_mobile_scenarios.py for new depth. Add __init__.py to all test subdirectories. Tests: 1503 passed, 9 failed (pre-existing), 53 errors (pre-existing) https://claude.ai/code/session_019oMFNvD8uSGSSmBMGkBfQN
144 lines
5.8 KiB
Python
144 lines
5.8 KiB
Python
"""Tests for src/timmy/backends.py — AirLLM wrapper and helpers."""
|
|
|
|
import sys
|
|
from unittest.mock import MagicMock, patch
|
|
|
|
import pytest
|
|
|
|
|
|
# ── is_apple_silicon ──────────────────────────────────────────────────────────
|
|
|
|
def test_is_apple_silicon_true_on_arm_darwin():
|
|
with patch("timmy.backends.platform.system", return_value="Darwin"), \
|
|
patch("timmy.backends.platform.machine", return_value="arm64"):
|
|
from timmy.backends import is_apple_silicon
|
|
assert is_apple_silicon() is True
|
|
|
|
|
|
def test_is_apple_silicon_false_on_linux():
|
|
with patch("timmy.backends.platform.system", return_value="Linux"), \
|
|
patch("timmy.backends.platform.machine", return_value="x86_64"):
|
|
from timmy.backends import is_apple_silicon
|
|
assert is_apple_silicon() is False
|
|
|
|
|
|
def test_is_apple_silicon_false_on_intel_mac():
|
|
with patch("timmy.backends.platform.system", return_value="Darwin"), \
|
|
patch("timmy.backends.platform.machine", return_value="x86_64"):
|
|
from timmy.backends import is_apple_silicon
|
|
assert is_apple_silicon() is False
|
|
|
|
|
|
# ── airllm_available ─────────────────────────────────────────────────────────
|
|
|
|
def test_airllm_available_true_when_stub_in_sys_modules():
|
|
# conftest already stubs 'airllm' — importable → True.
|
|
from timmy.backends import airllm_available
|
|
assert airllm_available() is True
|
|
|
|
|
|
def test_airllm_available_false_when_not_importable():
|
|
# Temporarily remove the stub to simulate airllm not installed.
|
|
saved = sys.modules.pop("airllm", None)
|
|
try:
|
|
from timmy.backends import airllm_available
|
|
assert airllm_available() is False
|
|
finally:
|
|
if saved is not None:
|
|
sys.modules["airllm"] = saved
|
|
|
|
|
|
# ── TimmyAirLLMAgent construction ────────────────────────────────────────────
|
|
|
|
def test_airllm_agent_raises_on_unknown_size():
|
|
from timmy.backends import TimmyAirLLMAgent
|
|
with pytest.raises(ValueError, match="Unknown model size"):
|
|
TimmyAirLLMAgent(model_size="3b")
|
|
|
|
|
|
def test_airllm_agent_uses_automodel_on_non_apple():
|
|
"""Non-Apple-Silicon path uses AutoModel.from_pretrained."""
|
|
with patch("timmy.backends.is_apple_silicon", return_value=False):
|
|
from timmy.backends import TimmyAirLLMAgent
|
|
agent = TimmyAirLLMAgent(model_size="8b")
|
|
# sys.modules["airllm"] is a MagicMock; AutoModel.from_pretrained was called.
|
|
assert sys.modules["airllm"].AutoModel.from_pretrained.called
|
|
|
|
|
|
def test_airllm_agent_uses_mlx_on_apple_silicon():
|
|
"""Apple Silicon path uses AirLLMMLX, not AutoModel."""
|
|
with patch("timmy.backends.is_apple_silicon", return_value=True):
|
|
from timmy.backends import TimmyAirLLMAgent
|
|
agent = TimmyAirLLMAgent(model_size="8b")
|
|
assert sys.modules["airllm"].AirLLMMLX.called
|
|
|
|
|
|
def test_airllm_agent_resolves_correct_model_id_for_70b():
|
|
with patch("timmy.backends.is_apple_silicon", return_value=False):
|
|
from timmy.backends import TimmyAirLLMAgent, _AIRLLM_MODELS
|
|
TimmyAirLLMAgent(model_size="70b")
|
|
sys.modules["airllm"].AutoModel.from_pretrained.assert_called_with(
|
|
_AIRLLM_MODELS["70b"]
|
|
)
|
|
|
|
|
|
# ── TimmyAirLLMAgent.print_response ──────────────────────────────────────────
|
|
|
|
def _make_agent(model_size: str = "8b") -> "TimmyAirLLMAgent":
|
|
"""Helper: create an agent with a fully mocked underlying model."""
|
|
with patch("timmy.backends.is_apple_silicon", return_value=False):
|
|
from timmy.backends import TimmyAirLLMAgent
|
|
agent = TimmyAirLLMAgent(model_size=model_size)
|
|
|
|
# Replace the underlying model with a clean mock that returns predictable output.
|
|
mock_model = MagicMock()
|
|
mock_tokenizer = MagicMock()
|
|
# tokenizer() returns a dict-like object with an "input_ids" tensor mock.
|
|
input_ids_mock = MagicMock()
|
|
input_ids_mock.shape = [1, 10] # shape[1] = prompt token count = 10
|
|
token_dict = {"input_ids": input_ids_mock}
|
|
mock_tokenizer.return_value = token_dict
|
|
# generate() returns a list of token sequences.
|
|
mock_tokenizer.decode.return_value = "Sir, affirmative."
|
|
mock_model.tokenizer = mock_tokenizer
|
|
mock_model.generate.return_value = [list(range(15))] # 15 tokens total
|
|
agent._model = mock_model
|
|
return agent
|
|
|
|
|
|
def test_print_response_calls_generate():
|
|
agent = _make_agent()
|
|
agent.print_response("What is sovereignty?", stream=True)
|
|
agent._model.generate.assert_called_once()
|
|
|
|
|
|
def test_print_response_decodes_only_generated_tokens():
|
|
agent = _make_agent()
|
|
agent.print_response("Hello", stream=False)
|
|
# decode should be called with tokens starting at index 10 (prompt length).
|
|
decode_call = agent._model.tokenizer.decode.call_args
|
|
token_slice = decode_call[0][0]
|
|
assert list(token_slice) == list(range(10, 15))
|
|
|
|
|
|
def test_print_response_updates_history():
|
|
agent = _make_agent()
|
|
agent.print_response("First message")
|
|
assert any("First message" in turn for turn in agent._history)
|
|
assert any("Timmy:" in turn for turn in agent._history)
|
|
|
|
|
|
def test_print_response_history_included_in_second_prompt():
|
|
agent = _make_agent()
|
|
agent.print_response("First")
|
|
# Build the prompt for the second call — history should appear.
|
|
prompt = agent._build_prompt("Second")
|
|
assert "First" in prompt
|
|
assert "Second" in prompt
|
|
|
|
|
|
def test_print_response_stream_flag_accepted():
|
|
"""stream=False should not raise — it's accepted for API compatibility."""
|
|
agent = _make_agent()
|
|
agent.print_response("hello", stream=False) # no error
|