mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-04-25 00:51:20 +00:00
test(run_agent): pin proxy-env forwarding through keepalive transport
Adds a regression guard for the #11277 → proxy-bypass regression fixed in
42b394c3. With HTTPS_PROXY / HTTP_PROXY / ALL_PROXY set, the custom httpx
transport used for TCP keepalives must still route requests through an
HTTPProxy pool; without proxy env, no HTTPProxy mount should exist.
Also maps zrc <zhurongcheng@rcrai.com> → heykb in scripts/release.py
AUTHOR_MAP so the salvage PR passes the author-attribution CI check.
This commit is contained in:
parent
023208b17a
commit
d48d6fadff
2 changed files with 138 additions and 0 deletions
|
|
@ -275,6 +275,7 @@ AUTHOR_MAP = {
|
|||
"junminliu@gmail.com": "JimLiu",
|
||||
"jarvischer@gmail.com": "maxchernin",
|
||||
"levantam.98.2324@gmail.com": "LVT382009",
|
||||
"zhurongcheng@rcrai.com": "heykb",
|
||||
}
|
||||
|
||||
|
||||
|
|
|
|||
137
tests/run_agent/test_create_openai_client_proxy_env.py
Normal file
137
tests/run_agent/test_create_openai_client_proxy_env.py
Normal file
|
|
@ -0,0 +1,137 @@
|
|||
"""Regression guard: _create_openai_client must honor HTTP(S)_PROXY env vars.
|
||||
|
||||
When #11277 re-landed TCP keepalives, ``_create_openai_client`` began passing
|
||||
a custom ``transport=httpx.HTTPTransport(...)`` to ``httpx.Client``. httpx only
|
||||
auto-reads ``HTTP_PROXY`` / ``HTTPS_PROXY`` / ``ALL_PROXY`` when
|
||||
``transport is None`` (see ``Client.__init__``:
|
||||
``allow_env_proxies = trust_env and transport is None``). As a result, proxy
|
||||
env vars were silently ignored for the primary chat client, causing requests
|
||||
to bypass local proxies (Clash, corporate egress, etc.) and hit upstream
|
||||
directly from the raw interface.
|
||||
|
||||
For users on WSL2 + Clash TUN this surfaced as Cloudflare ``cf-mitigated:
|
||||
challenge`` 403s against ``chatgpt.com/backend-api/codex`` once they upgraded
|
||||
past #11277. The fix forwards the proxy URL explicitly to ``httpx.Client``
|
||||
while keeping the keepalive-enabled transport in place.
|
||||
|
||||
This test pins that the constructed ``httpx.Client`` mounts an ``HTTPProxy``
|
||||
pool when a proxy env var is set, AND that the socket-level keepalive
|
||||
transport is still installed on the no-proxy default path.
|
||||
"""
|
||||
from unittest.mock import patch
|
||||
|
||||
import httpx
|
||||
|
||||
from run_agent import AIAgent, _get_proxy_from_env
|
||||
|
||||
|
||||
def _make_agent():
|
||||
return AIAgent(
|
||||
api_key="test-key",
|
||||
base_url="https://chatgpt.com/backend-api/codex",
|
||||
provider="openai-codex",
|
||||
model="gpt-5.4",
|
||||
quiet_mode=True,
|
||||
skip_context_files=True,
|
||||
skip_memory=True,
|
||||
)
|
||||
|
||||
|
||||
def _extract_http_client(client_kwargs: dict):
|
||||
"""_create_openai_client calls ``OpenAI(**client_kwargs)``; grab the injected client."""
|
||||
return client_kwargs.get("http_client")
|
||||
|
||||
|
||||
def test_get_proxy_from_env_prefers_https_then_http_then_all(monkeypatch):
|
||||
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
||||
"https_proxy", "http_proxy", "all_proxy"):
|
||||
monkeypatch.delenv(key, raising=False)
|
||||
assert _get_proxy_from_env() is None
|
||||
|
||||
monkeypatch.setenv("ALL_PROXY", "http://all:1")
|
||||
assert _get_proxy_from_env() == "http://all:1"
|
||||
|
||||
monkeypatch.setenv("HTTP_PROXY", "http://http:2")
|
||||
assert _get_proxy_from_env() == "http://http:2"
|
||||
|
||||
monkeypatch.setenv("HTTPS_PROXY", "http://https:3")
|
||||
assert _get_proxy_from_env() == "http://https:3"
|
||||
|
||||
|
||||
def test_get_proxy_from_env_ignores_blank_values(monkeypatch):
|
||||
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
||||
"https_proxy", "http_proxy", "all_proxy"):
|
||||
monkeypatch.delenv(key, raising=False)
|
||||
monkeypatch.setenv("HTTPS_PROXY", " ")
|
||||
monkeypatch.setenv("HTTP_PROXY", "http://real-proxy:8080")
|
||||
assert _get_proxy_from_env() == "http://real-proxy:8080"
|
||||
|
||||
|
||||
@patch("run_agent.OpenAI")
|
||||
def test_create_openai_client_routes_via_proxy_when_env_set(mock_openai, monkeypatch):
|
||||
"""With HTTPS_PROXY set, the custom httpx.Client must mount an HTTPProxy pool.
|
||||
|
||||
This is the WSL2 + Clash / corporate-egress case. Before the fix, the custom
|
||||
transport suppressed httpx's env-proxy auto-detection, so requests bypassed
|
||||
the proxy entirely.
|
||||
"""
|
||||
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
||||
"https_proxy", "http_proxy", "all_proxy"):
|
||||
monkeypatch.delenv(key, raising=False)
|
||||
monkeypatch.setenv("HTTPS_PROXY", "http://127.0.0.1:7897")
|
||||
|
||||
agent = _make_agent()
|
||||
kwargs = {
|
||||
"api_key": "test-key",
|
||||
"base_url": "https://chatgpt.com/backend-api/codex",
|
||||
}
|
||||
agent._create_openai_client(kwargs, reason="test", shared=False)
|
||||
|
||||
forwarded = mock_openai.call_args.kwargs
|
||||
http_client = _extract_http_client(forwarded)
|
||||
assert isinstance(http_client, httpx.Client), (
|
||||
"Expected _create_openai_client to inject a keepalive-enabled "
|
||||
"httpx.Client; got %r" % (http_client,)
|
||||
)
|
||||
# Verify a proxy mount exists. httpx Client(proxy=...) rewrites _mounts so
|
||||
# the proxied pool (HTTPProxy) sits alongside the base transport.
|
||||
proxied_pools = [
|
||||
type(mount._pool).__name__
|
||||
for mount in http_client._mounts.values()
|
||||
if mount is not None and hasattr(mount, "_pool")
|
||||
]
|
||||
assert "HTTPProxy" in proxied_pools, (
|
||||
"Expected httpx.Client to route through HTTPProxy when HTTPS_PROXY is "
|
||||
"set; found pools: %r" % (proxied_pools,)
|
||||
)
|
||||
http_client.close()
|
||||
|
||||
|
||||
@patch("run_agent.OpenAI")
|
||||
def test_create_openai_client_no_proxy_when_env_unset(mock_openai, monkeypatch):
|
||||
"""Without proxy env vars, the keepalive transport must still be installed
|
||||
and no HTTPProxy mount should exist."""
|
||||
for key in ("HTTPS_PROXY", "HTTP_PROXY", "ALL_PROXY",
|
||||
"https_proxy", "http_proxy", "all_proxy"):
|
||||
monkeypatch.delenv(key, raising=False)
|
||||
|
||||
agent = _make_agent()
|
||||
kwargs = {
|
||||
"api_key": "test-key",
|
||||
"base_url": "https://chatgpt.com/backend-api/codex",
|
||||
}
|
||||
agent._create_openai_client(kwargs, reason="test", shared=False)
|
||||
|
||||
forwarded = mock_openai.call_args.kwargs
|
||||
http_client = _extract_http_client(forwarded)
|
||||
assert isinstance(http_client, httpx.Client)
|
||||
pool_types = [
|
||||
type(mount._pool).__name__
|
||||
for mount in http_client._mounts.values()
|
||||
if mount is not None and hasattr(mount, "_pool")
|
||||
]
|
||||
assert "HTTPProxy" not in pool_types, (
|
||||
"No proxy env set but httpx.Client still mounted HTTPProxy; "
|
||||
"pools were %r" % (pool_types,)
|
||||
)
|
||||
http_client.close()
|
||||
Loading…
Add table
Add a link
Reference in a new issue