From 9742f11fda2afbd79b7d3206112276f3e5949f83 Mon Sep 17 00:00:00 2001 From: teknium1 Date: Fri, 6 Mar 2026 19:01:38 -0800 Subject: [PATCH] chore: add context lengths for Kimi and MiniMax models Adds DEFAULT_CONTEXT_LENGTHS entries for kimi-k2.5 (262144), kimi-k2-thinking (262144), kimi-k2-turbo-preview (262144), kimi-k2-0905-preview (131072), MiniMax-M2.5/M2.5-highspeed/M2.1 (204800), and glm-4.5/4.5-flash (131072). Avoids unnecessary 2M-token probe on first use with direct providers. --- agent/model_metadata.py | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/agent/model_metadata.py b/agent/model_metadata.py index f61dc2dda..3b2ab9d0f 100644 --- a/agent/model_metadata.py +++ b/agent/model_metadata.py @@ -51,6 +51,15 @@ DEFAULT_CONTEXT_LENGTHS = { "qwen/qwen-2.5-72b-instruct": 32768, "glm-4.7": 202752, "glm-5": 202752, + "glm-4.5": 131072, + "glm-4.5-flash": 131072, + "kimi-k2.5": 262144, + "kimi-k2-thinking": 262144, + "kimi-k2-turbo-preview": 262144, + "kimi-k2-0905-preview": 131072, + "MiniMax-M2.5": 204800, + "MiniMax-M2.5-highspeed": 204800, + "MiniMax-M2.1": 204800, }