mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-05-01 01:51:44 +00:00
48 lines
2 KiB
Python
48 lines
2 KiB
Python
"""LM Studio reasoning-effort resolution shared by the chat-completions
|
|
transport and run_agent's iteration-limit summary path.
|
|
|
|
LM Studio publishes per-model ``capabilities.reasoning.allowed_options`` (e.g.
|
|
``["off","on"]`` for toggle-style models, ``["off","minimal","low"]`` for
|
|
graduated models). We map the user's ``reasoning_config`` onto LM Studio's
|
|
OpenAI-compatible vocabulary, then clamp against the model's allowed set so
|
|
the server doesn't 400 on an unsupported effort.
|
|
"""
|
|
|
|
from __future__ import annotations
|
|
|
|
from typing import List, Optional
|
|
|
|
# LM Studio accepts these top-level reasoning_effort values via its
|
|
# OpenAI-compatible chat.completions endpoint.
|
|
_LM_VALID_EFFORTS = {"none", "minimal", "low", "medium", "high", "xhigh"}
|
|
|
|
# Toggle-style models publish allowed_options as ["off","on"] in /api/v1/models.
|
|
# Map them onto the OpenAI-compatible request vocabulary.
|
|
_LM_EFFORT_ALIASES = {"off": "none", "on": "medium"}
|
|
|
|
|
|
def resolve_lmstudio_effort(
|
|
reasoning_config: Optional[dict],
|
|
allowed_options: Optional[List[str]],
|
|
) -> Optional[str]:
|
|
"""Return the ``reasoning_effort`` string to send to LM Studio, or ``None``.
|
|
|
|
``None`` means "omit the field": the user picked a level the model can't
|
|
honor, so let LM Studio fall back to the model's declared default rather
|
|
than silently substituting a different effort. When ``allowed_options`` is
|
|
falsy (probe failed), skip clamping and send the resolved effort anyway.
|
|
"""
|
|
effort = "medium"
|
|
if reasoning_config and isinstance(reasoning_config, dict):
|
|
if reasoning_config.get("enabled") is False:
|
|
effort = "none"
|
|
else:
|
|
raw = (reasoning_config.get("effort") or "").strip().lower()
|
|
raw = _LM_EFFORT_ALIASES.get(raw, raw)
|
|
if raw in _LM_VALID_EFFORTS:
|
|
effort = raw
|
|
if allowed_options:
|
|
allowed = {_LM_EFFORT_ALIASES.get(opt, opt) for opt in allowed_options}
|
|
if effort not in allowed:
|
|
return None
|
|
return effort
|