mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-04-25 00:51:20 +00:00
Custom Claude proxies fronted by Cloudflare with Browser Integrity Check
enabled (e.g. `packyapi.com`) reject requests with the default
`Python-urllib/*` signature, returning HTTP 403 "error code: 1010".
`probe_api_models` swallowed that in its blanket `except Exception:
continue`, so `validate_requested_model` returned the misleading
"Could not reach the <provider> API to validate `<model>`" error even
though the endpoint is reachable and lists the requested model.
Advertise the probe request as `hermes-cli/<version>` so Cloudflare
treats it as a first-party client. This mirrors the pattern already used
by `agent/gemini_native_adapter.py` and `agent/anthropic_adapter.py`,
which set a descriptive UA for the same reason.
Reproduction (pre-fix):
python3 -c "
import urllib.request
req = urllib.request.Request(
'https://www.packyapi.com/v1/models',
headers={'Authorization': 'Bearer sk-...'})
urllib.request.urlopen(req).read()
"
urllib.error.HTTPError: HTTP Error 403: Forbidden
(body: b'error code: 1010')
Any non-urllib UA (Mozilla, curl, reqwest) returns 200 with the
OpenAI-compatible models listing.
Tested on macOS (Python 3.11). No cross-platform concerns — the change
is a single header addition to an existing `urllib.request.Request`.
602 lines
26 KiB
Python
602 lines
26 KiB
Python
"""Tests for provider-aware `/model` validation in hermes_cli.models."""
|
|
|
|
from unittest.mock import patch
|
|
|
|
from hermes_cli.models import (
|
|
copilot_model_api_mode,
|
|
fetch_github_model_catalog,
|
|
curated_models_for_provider,
|
|
fetch_api_models,
|
|
github_model_reasoning_efforts,
|
|
normalize_copilot_model_id,
|
|
normalize_opencode_model_id,
|
|
normalize_provider,
|
|
opencode_model_api_mode,
|
|
parse_model_input,
|
|
probe_api_models,
|
|
provider_label,
|
|
provider_model_ids,
|
|
validate_requested_model,
|
|
)
|
|
|
|
|
|
# -- helpers -----------------------------------------------------------------
|
|
|
|
FAKE_API_MODELS = [
|
|
"anthropic/claude-opus-4.6",
|
|
"anthropic/claude-sonnet-4.5",
|
|
"openai/gpt-5.4-pro",
|
|
"openai/gpt-5.4",
|
|
"google/gemini-3-pro-preview",
|
|
]
|
|
|
|
|
|
def _validate(model, provider="openrouter", api_models=FAKE_API_MODELS, **kw):
|
|
"""Shortcut: call validate_requested_model with mocked API."""
|
|
probe_payload = {
|
|
"models": api_models,
|
|
"probed_url": "http://localhost:11434/v1/models",
|
|
"resolved_base_url": kw.get("base_url", "") or "http://localhost:11434/v1",
|
|
"suggested_base_url": None,
|
|
"used_fallback": False,
|
|
}
|
|
with patch("hermes_cli.models.fetch_api_models", return_value=api_models), \
|
|
patch("hermes_cli.models.probe_api_models", return_value=probe_payload):
|
|
return validate_requested_model(model, provider, **kw)
|
|
|
|
|
|
# -- parse_model_input -------------------------------------------------------
|
|
|
|
class TestParseModelInput:
|
|
def test_plain_model_keeps_current_provider(self):
|
|
provider, model = parse_model_input("anthropic/claude-sonnet-4.5", "openrouter")
|
|
assert provider == "openrouter"
|
|
assert model == "anthropic/claude-sonnet-4.5"
|
|
|
|
def test_provider_colon_model_switches_provider(self):
|
|
provider, model = parse_model_input("openrouter:anthropic/claude-sonnet-4.5", "nous")
|
|
assert provider == "openrouter"
|
|
assert model == "anthropic/claude-sonnet-4.5"
|
|
|
|
def test_provider_alias_resolved(self):
|
|
provider, model = parse_model_input("glm:glm-5", "openrouter")
|
|
assert provider == "zai"
|
|
assert model == "glm-5"
|
|
|
|
def test_no_slash_no_colon_keeps_provider(self):
|
|
provider, model = parse_model_input("gpt-5.4", "openrouter")
|
|
assert provider == "openrouter"
|
|
assert model == "gpt-5.4"
|
|
|
|
def test_nous_provider_switch(self):
|
|
provider, model = parse_model_input("nous:hermes-3", "openrouter")
|
|
assert provider == "nous"
|
|
assert model == "hermes-3"
|
|
|
|
def test_empty_model_after_colon_keeps_current(self):
|
|
provider, model = parse_model_input("openrouter:", "nous")
|
|
assert provider == "nous"
|
|
assert model == "openrouter:"
|
|
|
|
def test_colon_at_start_keeps_current(self):
|
|
provider, model = parse_model_input(":something", "openrouter")
|
|
assert provider == "openrouter"
|
|
assert model == ":something"
|
|
|
|
def test_unknown_prefix_colon_not_treated_as_provider(self):
|
|
"""Colons are only provider delimiters if the left side is a known provider."""
|
|
provider, model = parse_model_input("anthropic/claude-3.5-sonnet:beta", "openrouter")
|
|
assert provider == "openrouter"
|
|
assert model == "anthropic/claude-3.5-sonnet:beta"
|
|
|
|
def test_http_url_not_treated_as_provider(self):
|
|
provider, model = parse_model_input("http://localhost:8080/model", "openrouter")
|
|
assert provider == "openrouter"
|
|
assert model == "http://localhost:8080/model"
|
|
|
|
def test_custom_colon_model_single(self):
|
|
"""custom:model-name → anonymous custom provider."""
|
|
provider, model = parse_model_input("custom:qwen-2.5", "openrouter")
|
|
assert provider == "custom"
|
|
assert model == "qwen-2.5"
|
|
|
|
def test_custom_triple_syntax(self):
|
|
"""custom:name:model → named custom provider."""
|
|
provider, model = parse_model_input("custom:local-server:qwen-2.5", "openrouter")
|
|
assert provider == "custom:local-server"
|
|
assert model == "qwen-2.5"
|
|
|
|
def test_custom_triple_spaces(self):
|
|
"""Triple syntax should handle whitespace."""
|
|
provider, model = parse_model_input("custom: my-server : my-model ", "openrouter")
|
|
assert provider == "custom:my-server"
|
|
assert model == "my-model"
|
|
|
|
def test_custom_triple_empty_model_falls_back(self):
|
|
"""custom:name: with no model → treated as custom:name (bare)."""
|
|
provider, model = parse_model_input("custom:name:", "openrouter")
|
|
# Empty model after second colon → no triple match, falls through
|
|
assert provider == "custom"
|
|
assert model == "name:"
|
|
|
|
|
|
# -- curated_models_for_provider ---------------------------------------------
|
|
|
|
class TestCuratedModelsForProvider:
|
|
def test_openrouter_returns_curated_list(self):
|
|
with patch(
|
|
"hermes_cli.models.fetch_openrouter_models",
|
|
return_value=[
|
|
("anthropic/claude-opus-4.6", "recommended"),
|
|
("qwen/qwen3.6-plus", ""),
|
|
],
|
|
):
|
|
models = curated_models_for_provider("openrouter")
|
|
assert len(models) > 0
|
|
assert any("claude" in m[0] for m in models)
|
|
|
|
def test_zai_returns_glm_models(self):
|
|
models = curated_models_for_provider("zai")
|
|
assert any("glm" in m[0] for m in models)
|
|
|
|
def test_unknown_provider_returns_empty(self):
|
|
assert curated_models_for_provider("totally-unknown") == []
|
|
|
|
|
|
# -- normalize_provider ------------------------------------------------------
|
|
|
|
class TestNormalizeProvider:
|
|
def test_defaults_to_openrouter(self):
|
|
assert normalize_provider(None) == "openrouter"
|
|
assert normalize_provider("") == "openrouter"
|
|
|
|
def test_known_aliases(self):
|
|
assert normalize_provider("glm") == "zai"
|
|
assert normalize_provider("kimi") == "kimi-coding"
|
|
assert normalize_provider("moonshot") == "kimi-coding"
|
|
assert normalize_provider("github-copilot") == "copilot"
|
|
|
|
def test_case_insensitive(self):
|
|
assert normalize_provider("OpenRouter") == "openrouter"
|
|
|
|
|
|
class TestProviderLabel:
|
|
def test_known_labels_and_auto(self):
|
|
assert provider_label("anthropic") == "Anthropic"
|
|
assert provider_label("kimi") == "Kimi / Kimi Coding Plan"
|
|
assert provider_label("copilot") == "GitHub Copilot"
|
|
assert provider_label("copilot-acp") == "GitHub Copilot ACP"
|
|
assert provider_label("auto") == "Auto"
|
|
|
|
def test_unknown_provider_preserves_original_name(self):
|
|
assert provider_label("my-custom-provider") == "my-custom-provider"
|
|
|
|
|
|
# -- provider_model_ids ------------------------------------------------------
|
|
|
|
class TestProviderModelIds:
|
|
def test_openrouter_returns_curated_list(self):
|
|
with patch(
|
|
"hermes_cli.models.fetch_openrouter_models",
|
|
return_value=[
|
|
("anthropic/claude-opus-4.6", "recommended"),
|
|
("qwen/qwen3.6-plus", ""),
|
|
],
|
|
):
|
|
ids = provider_model_ids("openrouter")
|
|
assert len(ids) > 0
|
|
assert all("/" in mid for mid in ids)
|
|
|
|
def test_unknown_provider_returns_empty(self):
|
|
assert provider_model_ids("some-unknown-provider") == []
|
|
|
|
def test_zai_returns_glm_models(self):
|
|
assert "glm-5" in provider_model_ids("zai")
|
|
|
|
def test_copilot_prefers_live_catalog(self):
|
|
with patch("hermes_cli.auth.resolve_api_key_provider_credentials", return_value={"api_key": "gh-token"}), \
|
|
patch("hermes_cli.models._fetch_github_models", return_value=["gpt-5.4", "claude-sonnet-4.6"]):
|
|
assert provider_model_ids("copilot") == ["gpt-5.4", "claude-sonnet-4.6"]
|
|
|
|
def test_copilot_acp_reuses_copilot_catalog(self):
|
|
with patch("hermes_cli.auth.resolve_api_key_provider_credentials", return_value={"api_key": "gh-token"}), \
|
|
patch("hermes_cli.models._fetch_github_models", return_value=["gpt-5.4", "claude-sonnet-4.6"]):
|
|
assert provider_model_ids("copilot-acp") == ["gpt-5.4", "claude-sonnet-4.6"]
|
|
|
|
def test_copilot_acp_falls_back_to_copilot_defaults(self):
|
|
with patch("hermes_cli.auth.resolve_api_key_provider_credentials", side_effect=Exception("no token")), \
|
|
patch("hermes_cli.models._fetch_github_models", return_value=None):
|
|
ids = provider_model_ids("copilot-acp")
|
|
|
|
assert "gpt-5.4" in ids
|
|
assert "copilot-acp" not in ids
|
|
|
|
|
|
# -- fetch_api_models --------------------------------------------------------
|
|
|
|
class TestFetchApiModels:
|
|
def test_returns_none_when_no_base_url(self):
|
|
assert fetch_api_models("key", None) is None
|
|
|
|
def test_returns_none_on_network_error(self):
|
|
with patch("hermes_cli.models.urllib.request.urlopen", side_effect=Exception("timeout")):
|
|
assert fetch_api_models("key", "https://example.com/v1") is None
|
|
|
|
def test_probe_api_models_tries_v1_fallback(self):
|
|
class _Resp:
|
|
def __enter__(self):
|
|
return self
|
|
|
|
def __exit__(self, exc_type, exc, tb):
|
|
return False
|
|
|
|
def read(self):
|
|
return b'{"data": [{"id": "local-model"}]}'
|
|
|
|
calls = []
|
|
|
|
def _fake_urlopen(req, timeout=5.0):
|
|
calls.append(req.full_url)
|
|
if req.full_url.endswith("/v1/models"):
|
|
return _Resp()
|
|
raise Exception("404")
|
|
|
|
with patch("hermes_cli.models.urllib.request.urlopen", side_effect=_fake_urlopen):
|
|
probe = probe_api_models("key", "http://localhost:8000")
|
|
|
|
assert calls == ["http://localhost:8000/models", "http://localhost:8000/v1/models"]
|
|
assert probe["models"] == ["local-model"]
|
|
assert probe["resolved_base_url"] == "http://localhost:8000/v1"
|
|
assert probe["used_fallback"] is True
|
|
|
|
def test_probe_api_models_uses_copilot_catalog(self):
|
|
class _Resp:
|
|
def __enter__(self):
|
|
return self
|
|
|
|
def __exit__(self, exc_type, exc, tb):
|
|
return False
|
|
|
|
def read(self):
|
|
return b'{"data": [{"id": "gpt-5.4", "model_picker_enabled": true, "supported_endpoints": ["/responses"], "capabilities": {"type": "chat", "supports": {"reasoning_effort": ["low", "medium", "high"]}}}, {"id": "claude-sonnet-4.6", "model_picker_enabled": true, "supported_endpoints": ["/chat/completions"], "capabilities": {"type": "chat", "supports": {"reasoning_effort": ["low", "medium", "high"]}}}, {"id": "text-embedding-3-small", "model_picker_enabled": true, "capabilities": {"type": "embedding"}}]}'
|
|
|
|
with patch("hermes_cli.models.urllib.request.urlopen", return_value=_Resp()) as mock_urlopen:
|
|
probe = probe_api_models("gh-token", "https://api.githubcopilot.com")
|
|
|
|
assert mock_urlopen.call_args[0][0].full_url == "https://api.githubcopilot.com/models"
|
|
assert probe["models"] == ["gpt-5.4", "claude-sonnet-4.6"]
|
|
assert probe["resolved_base_url"] == "https://api.githubcopilot.com"
|
|
assert probe["used_fallback"] is False
|
|
|
|
def test_fetch_github_model_catalog_filters_non_chat_models(self):
|
|
class _Resp:
|
|
def __enter__(self):
|
|
return self
|
|
|
|
def __exit__(self, exc_type, exc, tb):
|
|
return False
|
|
|
|
def read(self):
|
|
return b'{"data": [{"id": "gpt-5.4", "model_picker_enabled": true, "supported_endpoints": ["/responses"], "capabilities": {"type": "chat", "supports": {"reasoning_effort": ["low", "medium", "high"]}}}, {"id": "text-embedding-3-small", "model_picker_enabled": true, "capabilities": {"type": "embedding"}}]}'
|
|
|
|
with patch("hermes_cli.models.urllib.request.urlopen", return_value=_Resp()):
|
|
catalog = fetch_github_model_catalog("gh-token")
|
|
|
|
assert catalog is not None
|
|
assert [item["id"] for item in catalog] == ["gpt-5.4"]
|
|
|
|
|
|
class TestGithubReasoningEfforts:
|
|
def test_gpt5_supports_minimal_to_high(self):
|
|
catalog = [{
|
|
"id": "gpt-5.4",
|
|
"capabilities": {"type": "chat", "supports": {"reasoning_effort": ["low", "medium", "high"]}},
|
|
"supported_endpoints": ["/responses"],
|
|
}]
|
|
assert github_model_reasoning_efforts("gpt-5.4", catalog=catalog) == [
|
|
"low",
|
|
"medium",
|
|
"high",
|
|
]
|
|
|
|
def test_legacy_catalog_reasoning_still_supported(self):
|
|
catalog = [{"id": "openai/o3", "capabilities": ["reasoning"]}]
|
|
assert github_model_reasoning_efforts("openai/o3", catalog=catalog) == [
|
|
"low",
|
|
"medium",
|
|
"high",
|
|
]
|
|
|
|
def test_non_reasoning_model_returns_empty(self):
|
|
catalog = [{"id": "gpt-4.1", "capabilities": {"type": "chat", "supports": {}}}]
|
|
assert github_model_reasoning_efforts("gpt-4.1", catalog=catalog) == []
|
|
|
|
|
|
class TestCopilotNormalization:
|
|
def test_normalize_old_github_models_slug(self):
|
|
catalog = [{"id": "gpt-4.1"}, {"id": "gpt-5.4"}]
|
|
assert normalize_copilot_model_id("openai/gpt-4.1-mini", catalog=catalog) == "gpt-4.1"
|
|
|
|
def test_copilot_api_mode_gpt5_uses_responses(self):
|
|
"""GPT-5+ models should use Responses API (matching opencode)."""
|
|
assert copilot_model_api_mode("gpt-5.4") == "codex_responses"
|
|
assert copilot_model_api_mode("gpt-5.4-mini") == "codex_responses"
|
|
assert copilot_model_api_mode("gpt-5.3-codex") == "codex_responses"
|
|
assert copilot_model_api_mode("gpt-5.2-codex") == "codex_responses"
|
|
assert copilot_model_api_mode("gpt-5.2") == "codex_responses"
|
|
|
|
def test_copilot_api_mode_gpt5_mini_uses_chat(self):
|
|
"""gpt-5-mini is the exception — uses Chat Completions."""
|
|
assert copilot_model_api_mode("gpt-5-mini") == "chat_completions"
|
|
|
|
def test_copilot_api_mode_non_gpt5_uses_chat(self):
|
|
"""Non-GPT-5 models use Chat Completions."""
|
|
assert copilot_model_api_mode("gpt-4.1") == "chat_completions"
|
|
assert copilot_model_api_mode("gpt-4o") == "chat_completions"
|
|
assert copilot_model_api_mode("gpt-4o-mini") == "chat_completions"
|
|
assert copilot_model_api_mode("claude-sonnet-4.6") == "chat_completions"
|
|
assert copilot_model_api_mode("claude-opus-4.6") == "chat_completions"
|
|
assert copilot_model_api_mode("gemini-2.5-pro") == "chat_completions"
|
|
|
|
def test_copilot_api_mode_with_catalog_both_endpoints(self):
|
|
"""When catalog shows both endpoints, model ID pattern wins."""
|
|
catalog = [{
|
|
"id": "gpt-5.4",
|
|
"supported_endpoints": ["/chat/completions", "/responses"],
|
|
}]
|
|
# GPT-5.4 should use responses even though chat/completions is listed
|
|
assert copilot_model_api_mode("gpt-5.4", catalog=catalog) == "codex_responses"
|
|
|
|
def test_copilot_api_mode_with_catalog_only_responses(self):
|
|
catalog = [{
|
|
"id": "gpt-5.4",
|
|
"supported_endpoints": ["/responses"],
|
|
"capabilities": {"type": "chat"},
|
|
}]
|
|
assert copilot_model_api_mode("gpt-5.4", catalog=catalog) == "codex_responses"
|
|
|
|
def test_normalize_opencode_model_id_strips_provider_prefix(self):
|
|
assert normalize_opencode_model_id("opencode-go", "opencode-go/kimi-k2.5") == "kimi-k2.5"
|
|
assert normalize_opencode_model_id("opencode-zen", "opencode-zen/claude-sonnet-4-6") == "claude-sonnet-4-6"
|
|
assert normalize_opencode_model_id("opencode-go", "glm-5") == "glm-5"
|
|
|
|
def test_opencode_zen_api_modes_match_docs(self):
|
|
assert opencode_model_api_mode("opencode-zen", "gpt-5.4") == "codex_responses"
|
|
assert opencode_model_api_mode("opencode-zen", "gpt-5.3-codex") == "codex_responses"
|
|
assert opencode_model_api_mode("opencode-zen", "opencode-zen/gpt-5.4") == "codex_responses"
|
|
assert opencode_model_api_mode("opencode-zen", "claude-sonnet-4-6") == "anthropic_messages"
|
|
assert opencode_model_api_mode("opencode-zen", "opencode-zen/claude-sonnet-4-6") == "anthropic_messages"
|
|
assert opencode_model_api_mode("opencode-zen", "gemini-3-flash") == "chat_completions"
|
|
assert opencode_model_api_mode("opencode-zen", "minimax-m2.5") == "chat_completions"
|
|
|
|
def test_opencode_go_api_modes_match_docs(self):
|
|
assert opencode_model_api_mode("opencode-go", "glm-5.1") == "chat_completions"
|
|
assert opencode_model_api_mode("opencode-go", "opencode-go/glm-5.1") == "chat_completions"
|
|
assert opencode_model_api_mode("opencode-go", "glm-5") == "chat_completions"
|
|
assert opencode_model_api_mode("opencode-go", "opencode-go/glm-5") == "chat_completions"
|
|
assert opencode_model_api_mode("opencode-go", "kimi-k2.5") == "chat_completions"
|
|
assert opencode_model_api_mode("opencode-go", "opencode-go/kimi-k2.5") == "chat_completions"
|
|
assert opencode_model_api_mode("opencode-go", "minimax-m2.5") == "anthropic_messages"
|
|
assert opencode_model_api_mode("opencode-go", "opencode-go/minimax-m2.5") == "anthropic_messages"
|
|
|
|
|
|
# -- validate — format checks -----------------------------------------------
|
|
|
|
class TestValidateFormatChecks:
|
|
def test_empty_model_rejected(self):
|
|
result = _validate("")
|
|
assert result["accepted"] is False
|
|
assert "empty" in result["message"]
|
|
|
|
def test_whitespace_only_rejected(self):
|
|
result = _validate(" ")
|
|
assert result["accepted"] is False
|
|
|
|
def test_model_with_spaces_rejected(self):
|
|
result = _validate("anthropic/ claude-opus")
|
|
assert result["accepted"] is False
|
|
|
|
def test_no_slash_model_still_probes_api(self):
|
|
result = _validate("gpt-5.4", api_models=["gpt-5.4", "gpt-5.4-pro"])
|
|
assert result["accepted"] is True
|
|
assert result["persist"] is True
|
|
|
|
def test_no_slash_model_rejected_if_not_in_api(self):
|
|
result = _validate("gpt-5.4", api_models=["openai/gpt-5.4"])
|
|
assert result["accepted"] is False
|
|
assert result["persist"] is False
|
|
assert "not found" in result["message"]
|
|
|
|
|
|
# -- validate — API found ----------------------------------------------------
|
|
|
|
class TestValidateApiFound:
|
|
def test_model_found_in_api(self):
|
|
result = _validate("anthropic/claude-opus-4.6")
|
|
assert result["accepted"] is True
|
|
assert result["persist"] is True
|
|
assert result["recognized"] is True
|
|
|
|
def test_model_found_for_custom_endpoint(self):
|
|
result = _validate(
|
|
"my-model", provider="openrouter",
|
|
api_models=["my-model"], base_url="http://localhost:11434/v1",
|
|
)
|
|
assert result["accepted"] is True
|
|
assert result["persist"] is True
|
|
assert result["recognized"] is True
|
|
|
|
|
|
# -- validate — API not found ------------------------------------------------
|
|
|
|
class TestValidateApiNotFound:
|
|
def test_model_not_in_api_rejected_with_guidance(self):
|
|
result = _validate("anthropic/claude-nonexistent")
|
|
assert result["accepted"] is False
|
|
assert result["persist"] is False
|
|
assert "not found" in result["message"]
|
|
|
|
def test_warning_includes_suggestions(self):
|
|
result = _validate("anthropic/claude-opus-4.5")
|
|
assert result["accepted"] is True
|
|
# Close match auto-corrects; less similar inputs show suggestions
|
|
assert "Auto-corrected" in result["message"] or "Similar models" in result["message"]
|
|
|
|
def test_auto_correction_returns_corrected_model(self):
|
|
"""When a very close match exists, validate returns corrected_model."""
|
|
result = _validate("anthropic/claude-opus-4.5")
|
|
assert result["accepted"] is True
|
|
assert result.get("corrected_model") == "anthropic/claude-opus-4.6"
|
|
assert result["recognized"] is True
|
|
|
|
def test_dissimilar_model_shows_suggestions_not_autocorrect(self):
|
|
"""Models too different for auto-correction are rejected with suggestions."""
|
|
result = _validate("anthropic/claude-nonexistent")
|
|
assert result["accepted"] is False
|
|
assert result.get("corrected_model") is None
|
|
assert "not found" in result["message"]
|
|
|
|
|
|
# -- validate — API unreachable — reject with guidance ----------------
|
|
|
|
class TestValidateApiFallback:
|
|
def test_any_model_rejected_when_api_down(self):
|
|
result = _validate("anthropic/claude-opus-4.6", api_models=None)
|
|
assert result["accepted"] is False
|
|
assert result["persist"] is False
|
|
|
|
def test_unknown_model_also_rejected_when_api_down(self):
|
|
result = _validate("anthropic/claude-next-gen", api_models=None)
|
|
assert result["accepted"] is False
|
|
assert result["persist"] is False
|
|
assert "could not reach" in result["message"].lower()
|
|
|
|
def test_zai_model_rejected_when_api_down(self):
|
|
result = _validate("glm-5", provider="zai", api_models=None)
|
|
assert result["accepted"] is False
|
|
assert result["persist"] is False
|
|
|
|
def test_unknown_provider_rejected_when_api_down(self):
|
|
result = _validate("some-model", provider="totally-unknown", api_models=None)
|
|
assert result["accepted"] is False
|
|
assert result["persist"] is False
|
|
|
|
def test_custom_endpoint_warns_with_probed_url_and_v1_hint(self):
|
|
with patch(
|
|
"hermes_cli.models.probe_api_models",
|
|
return_value={
|
|
"models": None,
|
|
"probed_url": "http://localhost:8000/v1/models",
|
|
"resolved_base_url": "http://localhost:8000",
|
|
"suggested_base_url": "http://localhost:8000/v1",
|
|
"used_fallback": False,
|
|
},
|
|
):
|
|
result = validate_requested_model(
|
|
"qwen3",
|
|
"custom",
|
|
api_key="local-key",
|
|
base_url="http://localhost:8000",
|
|
)
|
|
|
|
assert result["accepted"] is False
|
|
assert result["persist"] is False
|
|
assert "http://localhost:8000/v1/models" in result["message"]
|
|
assert "http://localhost:8000/v1" in result["message"]
|
|
|
|
|
|
# -- validate — Codex auto-correction ------------------------------------------
|
|
|
|
class TestValidateCodexAutoCorrection:
|
|
"""Auto-correction for typos on openai-codex provider."""
|
|
|
|
def test_missing_dash_auto_corrects(self):
|
|
"""gpt5.3-codex (missing dash) auto-corrects to gpt-5.3-codex."""
|
|
codex_models = ["gpt-5.4-mini", "gpt-5.4", "gpt-5.3-codex",
|
|
"gpt-5.2-codex", "gpt-5.1-codex-max"]
|
|
with patch("hermes_cli.models.provider_model_ids", return_value=codex_models):
|
|
result = validate_requested_model("gpt5.3-codex", "openai-codex")
|
|
assert result["accepted"] is True
|
|
assert result["recognized"] is True
|
|
assert result["corrected_model"] == "gpt-5.3-codex"
|
|
assert "Auto-corrected" in result["message"]
|
|
|
|
def test_exact_match_no_correction(self):
|
|
"""Exact model name does not trigger auto-correction."""
|
|
codex_models = ["gpt-5.4-mini", "gpt-5.4", "gpt-5.3-codex"]
|
|
with patch("hermes_cli.models.provider_model_ids", return_value=codex_models):
|
|
result = validate_requested_model("gpt-5.3-codex", "openai-codex")
|
|
assert result["accepted"] is True
|
|
assert result["recognized"] is True
|
|
assert result.get("corrected_model") is None
|
|
assert result["message"] is None
|
|
|
|
def test_very_different_name_falls_to_suggestions(self):
|
|
"""Names too different for auto-correction are rejected with a suggestion list."""
|
|
codex_models = ["gpt-5.4-mini", "gpt-5.4", "gpt-5.3-codex"]
|
|
with patch("hermes_cli.models.provider_model_ids", return_value=codex_models):
|
|
result = validate_requested_model("totally-wrong", "openai-codex")
|
|
assert result["accepted"] is False
|
|
assert result["recognized"] is False
|
|
assert result.get("corrected_model") is None
|
|
assert "not found" in result["message"]
|
|
|
|
|
|
# -- probe_api_models — Cloudflare UA mitigation --------------------------------
|
|
|
|
class TestProbeApiModelsUserAgent:
|
|
"""Probing custom /v1/models must send a Hermes User-Agent.
|
|
|
|
Some custom Claude proxies (e.g. ``packyapi.com``) sit behind Cloudflare with
|
|
Browser Integrity Check enabled. The default ``Python-urllib/3.x`` signature
|
|
is rejected with HTTP 403 ``error code: 1010``, which ``probe_api_models``
|
|
swallowed into ``{"models": None}``, surfacing to users as a misleading
|
|
"Could not reach the ... API to validate ..." error — even though the
|
|
endpoint is reachable and the listing exists.
|
|
"""
|
|
|
|
def _make_mock_response(self, body: bytes):
|
|
from unittest.mock import MagicMock
|
|
mock_resp = MagicMock()
|
|
mock_resp.__enter__ = MagicMock(return_value=mock_resp)
|
|
mock_resp.__exit__ = MagicMock(return_value=False)
|
|
mock_resp.read = MagicMock(return_value=body)
|
|
return mock_resp
|
|
|
|
def test_probe_sends_hermes_user_agent(self):
|
|
from unittest.mock import patch
|
|
|
|
body = b'{"data":[{"id":"claude-opus-4.7"}]}'
|
|
with patch(
|
|
"hermes_cli.models.urllib.request.urlopen",
|
|
return_value=self._make_mock_response(body),
|
|
) as mock_urlopen:
|
|
result = probe_api_models("sk-test", "https://example.com/v1")
|
|
|
|
assert result["models"] == ["claude-opus-4.7"]
|
|
# The urlopen call receives a Request object as its first positional arg
|
|
req = mock_urlopen.call_args[0][0]
|
|
ua = req.get_header("User-agent") # urllib title-cases header names
|
|
assert ua, "probe_api_models must send a User-Agent header"
|
|
assert ua.startswith("hermes-cli/"), (
|
|
f"User-Agent must advertise hermes-cli, got {ua!r}"
|
|
)
|
|
# Must not fall back to urllib's default — that's what Cloudflare 1010 blocks.
|
|
assert not ua.startswith("Python-urllib")
|
|
|
|
def test_probe_user_agent_sent_without_api_key(self):
|
|
"""UA must be present even for endpoints that don't need auth."""
|
|
from unittest.mock import patch
|
|
|
|
body = b'{"data":[]}'
|
|
with patch(
|
|
"hermes_cli.models.urllib.request.urlopen",
|
|
return_value=self._make_mock_response(body),
|
|
) as mock_urlopen:
|
|
probe_api_models(None, "https://example.com/v1")
|
|
|
|
req = mock_urlopen.call_args[0][0]
|
|
ua = req.get_header("User-agent")
|
|
assert ua and ua.startswith("hermes-cli/")
|
|
# No Authorization was set, but UA must still be present.
|
|
assert req.get_header("Authorization") is None
|