hermes-agent/environments/benchmarks
dmahan93 13f5459670 fix: use ManagedServer for vLLM in TBLite eval + local_vllm config
TBLite eval was bypassing ManagedServer and calling ServerManager
directly, which uses /v1/chat/completions — not available on the
atropos vllm_api_server (/generate only).

Now uses _use_managed_server() to detect vLLM/SGLang backends and
route through ManagedServer (Phase 2) with proper tool_parser and
/generate endpoint. Falls back to Phase 1 for OpenAI endpoints.

Also adds local_vllm.yaml config for running against a local vLLM
server with Docker sandboxes.
2026-03-11 06:52:55 -07:00
..
tblite fix: use ManagedServer for vLLM in TBLite eval + local_vllm config 2026-03-11 06:52:55 -07:00
terminalbench_2 fix: use ManagedServer for vLLM in TBLite eval + local_vllm config 2026-03-11 06:52:55 -07:00
yc_bench fix: update OpenRouter model names for yc-bench config 2026-03-06 19:58:56 -08:00
__init__.py Add new environments and enhance tool context functionality 2026-02-10 19:39:05 +00:00