From f41031af3a68bbb3672af2393484f99f4ecba29f Mon Sep 17 00:00:00 2001 From: iborazzi Date: Tue, 14 Apr 2026 13:50:18 +0300 Subject: [PATCH] fix: increase max_tokens for GLM 5.1 reasoning headroom --- agent/title_generator.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/agent/title_generator.py b/agent/title_generator.py index d6ed9200a..99c771cb5 100644 --- a/agent/title_generator.py +++ b/agent/title_generator.py @@ -38,7 +38,7 @@ def generate_title(user_message: str, assistant_response: str, timeout: float = response = call_llm( task="title_generation", messages=messages, - max_tokens=30, + max_tokens=500, temperature=0.3, timeout=timeout, )