diff --git a/hermes_cli/main.py b/hermes_cli/main.py index e522c0716..53f59a210 100644 --- a/hermes_cli/main.py +++ b/hermes_cli/main.py @@ -2472,10 +2472,10 @@ def _model_flow_kimi(config, current_model=""): # Step 3: Model selection — show appropriate models for the endpoint if is_coding_plan: - # Coding Plan models (kimi-for-coding first) + # Coding Plan models (kimi-k2.5 first) model_list = [ - "kimi-for-coding", "kimi-k2.5", + "kimi-for-coding", "kimi-k2-thinking", "kimi-k2-thinking-turbo", ] diff --git a/hermes_cli/models.py b/hermes_cli/models.py index 0208bc9fe..5b998ddc6 100644 --- a/hermes_cli/models.py +++ b/hermes_cli/models.py @@ -26,7 +26,8 @@ COPILOT_REASONING_EFFORTS_O_SERIES = ["low", "medium", "high"] # Fallback OpenRouter snapshot used when the live catalog is unavailable. # (model_id, display description shown in menus) OPENROUTER_MODELS: list[tuple[str, str]] = [ - ("anthropic/claude-opus-4.7", "recommended"), + ("moonshotai/kimi-k2.5", "recommended"), + ("anthropic/claude-opus-4.7", ""), ("anthropic/claude-opus-4.6", ""), ("anthropic/claude-sonnet-4.6", ""), ("qwen/qwen3.6-plus", ""), @@ -49,7 +50,6 @@ OPENROUTER_MODELS: list[tuple[str, str]] = [ ("z-ai/glm-5.1", ""), ("z-ai/glm-5v-turbo", ""), ("z-ai/glm-5-turbo", ""), - ("moonshotai/kimi-k2.5", ""), ("x-ai/grok-4.20", ""), ("nvidia/nemotron-3-super-120b-a12b", ""), ("nvidia/nemotron-3-super-120b-a12b:free", "free"), @@ -75,6 +75,7 @@ def _codex_curated_models() -> list[str]: _PROVIDER_MODELS: dict[str, list[str]] = { "nous": [ + "moonshotai/kimi-k2.5", "xiaomi/mimo-v2-pro", "anthropic/claude-opus-4.7", "anthropic/claude-opus-4.6", @@ -96,7 +97,6 @@ _PROVIDER_MODELS: dict[str, list[str]] = { "z-ai/glm-5.1", "z-ai/glm-5v-turbo", "z-ai/glm-5-turbo", - "moonshotai/kimi-k2.5", "x-ai/grok-4.20-beta", "nvidia/nemotron-3-super-120b-a12b", "nvidia/nemotron-3-super-120b-a12b:free", @@ -156,8 +156,8 @@ _PROVIDER_MODELS: dict[str, list[str]] = { "grok-4-1-fast-reasoning", ], "kimi-coding": [ - "kimi-for-coding", "kimi-k2.5", + "kimi-for-coding", "kimi-k2-thinking", "kimi-k2-thinking-turbo", "kimi-k2-turbo-preview", @@ -212,6 +212,7 @@ _PROVIDER_MODELS: dict[str, list[str]] = { "trinity-mini", ], "opencode-zen": [ + "kimi-k2.5", "gpt-5.4-pro", "gpt-5.4", "gpt-5.3-codex", @@ -243,16 +244,15 @@ _PROVIDER_MODELS: dict[str, list[str]] = { "glm-5", "glm-4.7", "glm-4.6", - "kimi-k2.5", "kimi-k2-thinking", "kimi-k2", "qwen3-coder", "big-pickle", ], "opencode-go": [ + "kimi-k2.5", "glm-5.1", "glm-5", - "kimi-k2.5", "mimo-v2-pro", "mimo-v2-omni", "minimax-m2.7", @@ -285,21 +285,21 @@ _PROVIDER_MODELS: dict[str, list[str]] = { # to https://dashscope-intl.aliyuncs.com/compatible-mode/v1 (OpenAI-compat) # or https://dashscope-intl.aliyuncs.com/apps/anthropic (Anthropic-compat). "alibaba": [ + "kimi-k2.5", "qwen3.5-plus", "qwen3-coder-plus", "qwen3-coder-next", # Third-party models available on coding-intl "glm-5", "glm-4.7", - "kimi-k2.5", "MiniMax-M2.5", ], # Curated HF model list — only agentic models that map to OpenRouter defaults. "huggingface": [ + "moonshotai/Kimi-K2.5", "Qwen/Qwen3.5-397B-A17B", "Qwen/Qwen3.5-35B-A3B", "deepseek-ai/DeepSeek-V3.2", - "moonshotai/Kimi-K2.5", "MiniMaxAI/MiniMax-M2.5", "zai-org/GLM-5", "XiaomiMiMo/MiMo-V2-Flash",