diff --git a/agent/model_metadata.py b/agent/model_metadata.py index 7486afb04..66e97c0f8 100644 --- a/agent/model_metadata.py +++ b/agent/model_metadata.py @@ -113,6 +113,8 @@ DEFAULT_CONTEXT_LENGTHS = { "glm": 202752, # Kimi "kimi": 262144, + # Arcee + "trinity": 262144, # Hugging Face Inference Providers — model IDs use org/name format "Qwen/Qwen3.5-397B-A17B": 131072, "Qwen/Qwen3.5-35B-A3B": 131072, diff --git a/hermes_cli/models.py b/hermes_cli/models.py index 0915d150d..1243d160d 100644 --- a/hermes_cli/models.py +++ b/hermes_cli/models.py @@ -51,6 +51,7 @@ OPENROUTER_MODELS: list[tuple[str, str]] = [ ("nvidia/nemotron-3-super-120b-a12b", ""), ("nvidia/nemotron-3-super-120b-a12b:free", "free"), ("arcee-ai/trinity-large-preview:free", "free"), + ("arcee-ai/trinity-large-thinking", ""), ("openai/gpt-5.4-pro", ""), ("openai/gpt-5.4-nano", ""), ] @@ -82,6 +83,7 @@ _PROVIDER_MODELS: dict[str, list[str]] = { "nvidia/nemotron-3-super-120b-a12b", "nvidia/nemotron-3-super-120b-a12b:free", "arcee-ai/trinity-large-preview:free", + "arcee-ai/trinity-large-thinking", "openai/gpt-5.4-pro", "openai/gpt-5.4-nano", ],