Merge pull request #1609 from 0xbyt4/fix/context-counter-cache-tokens
fix: context counter shows cached token count in status bar
This commit is contained in:
@@ -5258,6 +5258,15 @@ class AIAgent:
|
||||
if hasattr(response, 'usage') and response.usage:
|
||||
if self.api_mode in ("codex_responses", "anthropic_messages"):
|
||||
prompt_tokens = getattr(response.usage, 'input_tokens', 0) or 0
|
||||
if self.api_mode == "anthropic_messages":
|
||||
# Anthropic splits input into cache_read + cache_creation
|
||||
# + non-cached input_tokens. Without adding the cached
|
||||
# portions, the context bar shows only the tiny non-cached
|
||||
# portion (e.g. 3 tokens) instead of the real total (~18K).
|
||||
# Other providers (OpenAI/Codex) already include cached
|
||||
# tokens in their input_tokens/prompt_tokens field.
|
||||
prompt_tokens += getattr(response.usage, 'cache_read_input_tokens', 0) or 0
|
||||
prompt_tokens += getattr(response.usage, 'cache_creation_input_tokens', 0) or 0
|
||||
completion_tokens = getattr(response.usage, 'output_tokens', 0) or 0
|
||||
total_tokens = (
|
||||
getattr(response.usage, 'total_tokens', None)
|
||||
|
||||
115
tests/test_context_token_tracking.py
Normal file
115
tests/test_context_token_tracking.py
Normal file
@@ -0,0 +1,115 @@
|
||||
"""Tests for context token tracking in run_agent.py's usage extraction.
|
||||
|
||||
The context counter (status bar) must show the TOTAL prompt tokens including
|
||||
Anthropic's cached portions. This is an integration test for the token
|
||||
extraction in run_conversation(), not the ContextCompressor itself (which
|
||||
is tested in tests/agent/test_context_compressor.py).
|
||||
"""
|
||||
|
||||
import sys
|
||||
import types
|
||||
from types import SimpleNamespace
|
||||
|
||||
sys.modules.setdefault("fire", types.SimpleNamespace(Fire=lambda *a, **k: None))
|
||||
sys.modules.setdefault("firecrawl", types.SimpleNamespace(Firecrawl=object))
|
||||
sys.modules.setdefault("fal_client", types.SimpleNamespace())
|
||||
|
||||
import run_agent
|
||||
|
||||
|
||||
def _patch_bootstrap(monkeypatch):
|
||||
monkeypatch.setattr(run_agent, "get_tool_definitions", lambda **kwargs: [{
|
||||
"type": "function",
|
||||
"function": {"name": "t", "description": "t", "parameters": {"type": "object", "properties": {}}},
|
||||
}])
|
||||
monkeypatch.setattr(run_agent, "check_toolset_requirements", lambda: {})
|
||||
|
||||
|
||||
class _FakeAnthropicClient:
|
||||
def close(self):
|
||||
pass
|
||||
|
||||
|
||||
def _make_agent(monkeypatch, api_mode, provider, response_fn):
|
||||
_patch_bootstrap(monkeypatch)
|
||||
if api_mode == "anthropic_messages":
|
||||
monkeypatch.setattr("agent.anthropic_adapter.build_anthropic_client", lambda k, b=None: _FakeAnthropicClient())
|
||||
|
||||
class _A(run_agent.AIAgent):
|
||||
def __init__(self, *a, **kw):
|
||||
kw.update(skip_context_files=True, skip_memory=True, max_iterations=4)
|
||||
super().__init__(*a, **kw)
|
||||
self._cleanup_task_resources = self._persist_session = lambda *a, **k: None
|
||||
self._save_trajectory = self._save_session_log = lambda *a, **k: None
|
||||
|
||||
def run_conversation(self, msg, conversation_history=None, task_id=None):
|
||||
self._interruptible_api_call = lambda kw: response_fn()
|
||||
return super().run_conversation(msg, conversation_history=conversation_history, task_id=task_id)
|
||||
|
||||
return _A(model="test-model", api_key="test-key", provider=provider, api_mode=api_mode)
|
||||
|
||||
|
||||
def _anthropic_resp(input_tok, output_tok, cache_read=0, cache_creation=0):
|
||||
usage_fields = {"input_tokens": input_tok, "output_tokens": output_tok}
|
||||
if cache_read:
|
||||
usage_fields["cache_read_input_tokens"] = cache_read
|
||||
if cache_creation:
|
||||
usage_fields["cache_creation_input_tokens"] = cache_creation
|
||||
return SimpleNamespace(
|
||||
content=[SimpleNamespace(type="text", text="ok")],
|
||||
stop_reason="end_turn",
|
||||
usage=SimpleNamespace(**usage_fields),
|
||||
model="claude-sonnet-4-6",
|
||||
)
|
||||
|
||||
|
||||
# -- Anthropic: cached tokens must be included --
|
||||
|
||||
def test_anthropic_cache_read_and_creation_added(monkeypatch):
|
||||
agent = _make_agent(monkeypatch, "anthropic_messages", "anthropic",
|
||||
lambda: _anthropic_resp(3, 10, cache_read=15000, cache_creation=2000))
|
||||
agent.run_conversation("hi")
|
||||
assert agent.context_compressor.last_prompt_tokens == 17003 # 3+15000+2000
|
||||
assert agent.session_prompt_tokens == 17003
|
||||
|
||||
|
||||
def test_anthropic_no_cache_fields(monkeypatch):
|
||||
agent = _make_agent(monkeypatch, "anthropic_messages", "anthropic",
|
||||
lambda: _anthropic_resp(500, 20))
|
||||
agent.run_conversation("hi")
|
||||
assert agent.context_compressor.last_prompt_tokens == 500
|
||||
|
||||
|
||||
def test_anthropic_cache_read_only(monkeypatch):
|
||||
agent = _make_agent(monkeypatch, "anthropic_messages", "anthropic",
|
||||
lambda: _anthropic_resp(5, 15, cache_read=17666, cache_creation=15))
|
||||
agent.run_conversation("hi")
|
||||
assert agent.context_compressor.last_prompt_tokens == 17686 # 5+17666+15
|
||||
|
||||
|
||||
# -- OpenAI: prompt_tokens already total --
|
||||
|
||||
def test_openai_prompt_tokens_unchanged(monkeypatch):
|
||||
resp = lambda: SimpleNamespace(
|
||||
choices=[SimpleNamespace(index=0, message=SimpleNamespace(
|
||||
role="assistant", content="ok", tool_calls=None, reasoning_content=None,
|
||||
), finish_reason="stop")],
|
||||
usage=SimpleNamespace(prompt_tokens=5000, completion_tokens=100, total_tokens=5100),
|
||||
model="gpt-4o",
|
||||
)
|
||||
agent = _make_agent(monkeypatch, "chat_completions", "openrouter", resp)
|
||||
agent.run_conversation("hi")
|
||||
assert agent.context_compressor.last_prompt_tokens == 5000
|
||||
|
||||
|
||||
# -- Codex: no cache fields, getattr returns 0 --
|
||||
|
||||
def test_codex_no_cache_fields(monkeypatch):
|
||||
resp = lambda: SimpleNamespace(
|
||||
output=[SimpleNamespace(type="message", content=[SimpleNamespace(type="output_text", text="ok")])],
|
||||
usage=SimpleNamespace(input_tokens=3000, output_tokens=50, total_tokens=3050),
|
||||
status="completed", model="gpt-5-codex",
|
||||
)
|
||||
agent = _make_agent(monkeypatch, "codex_responses", "openai-codex", resp)
|
||||
agent.run_conversation("hi")
|
||||
assert agent.context_compressor.last_prompt_tokens == 3000
|
||||
Reference in New Issue
Block a user