fix: increase max_tokens for GLM 5.1 reasoning headroom

This commit is contained in:
iborazzi 2026-04-14 13:50:18 +03:00 committed by Teknium
parent c78a188ddd
commit f41031af3a

View file

@ -38,7 +38,7 @@ def generate_title(user_message: str, assistant_response: str, timeout: float =
response = call_llm( response = call_llm(
task="title_generation", task="title_generation",
messages=messages, messages=messages,
max_tokens=30, max_tokens=500,
temperature=0.3, temperature=0.3,
timeout=timeout, timeout=timeout,
) )