mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-05-03 02:11:48 +00:00
When the curator consolidates skill X into umbrella Y, any cron job that listed X in its skills field would fail to load X at run time — the scheduler logs a warning and skips it, so the scheduled job runs without the instructions it was scheduled to follow. cron.jobs.rewrite_skill_refs(consolidated, pruned) now updates jobs in-place: consolidated names route to the umbrella target (dedup when umbrella is already present), pruned names are dropped. agent.curator._write_run_report calls it after classification, best-effort so a cron-side failure never breaks the curator itself. Results are recorded in run.json (counts.cron_jobs_rewritten + full cron_rewrites payload), a separate cron_rewrites.json for convenience when jobs were touched, and a section in REPORT.md. Reported by @tombielecki.
436 lines
16 KiB
Python
436 lines
16 KiB
Python
"""Tests for the curator per-run report writer (run.json + REPORT.md).
|
|
|
|
Reports live under ``~/.hermes/logs/curator/{YYYYMMDD-HHMMSS}/`` alongside
|
|
the standard log dir, not inside the user's ``skills/`` data directory.
|
|
"""
|
|
|
|
from __future__ import annotations
|
|
|
|
import json
|
|
import os
|
|
from datetime import datetime, timezone, timedelta
|
|
from pathlib import Path
|
|
|
|
import pytest
|
|
|
|
|
|
@pytest.fixture
|
|
def curator_env(tmp_path, monkeypatch):
|
|
"""Isolated HERMES_HOME with a skills/ dir + reset curator module state."""
|
|
home = tmp_path / ".hermes"
|
|
home.mkdir()
|
|
(home / "skills").mkdir()
|
|
(home / "logs").mkdir()
|
|
monkeypatch.setenv("HERMES_HOME", str(home))
|
|
monkeypatch.setattr(Path, "home", lambda: tmp_path)
|
|
|
|
import importlib
|
|
import hermes_constants
|
|
importlib.reload(hermes_constants)
|
|
from agent import curator
|
|
importlib.reload(curator)
|
|
from tools import skill_usage
|
|
importlib.reload(skill_usage)
|
|
yield {"home": home, "curator": curator, "skill_usage": skill_usage}
|
|
|
|
|
|
def _make_llm_meta(**overrides):
|
|
base = {
|
|
"final": "short summary of the pass",
|
|
"summary": "short summary",
|
|
"model": "test-model",
|
|
"provider": "test-provider",
|
|
"tool_calls": [],
|
|
"error": None,
|
|
}
|
|
base.update(overrides)
|
|
return base
|
|
|
|
|
|
def test_reports_root_is_under_logs_not_skills(curator_env):
|
|
"""Reports live in logs/curator/, not skills/ — operational telemetry
|
|
belongs with the logs, not with user-authored skill data."""
|
|
curator = curator_env["curator"]
|
|
root = curator._reports_root()
|
|
home = curator_env["home"]
|
|
# Must be under logs/
|
|
assert root == home / "logs" / "curator"
|
|
# Must NOT be under skills/
|
|
assert "skills" not in root.parts
|
|
|
|
|
|
def test_write_run_report_creates_both_files(curator_env):
|
|
"""Each run writes both a run.json (machine) and a REPORT.md (human)."""
|
|
curator = curator_env["curator"]
|
|
start = datetime.now(timezone.utc)
|
|
|
|
run_dir = curator._write_run_report(
|
|
started_at=start,
|
|
elapsed_seconds=12.345,
|
|
auto_counts={"checked": 5, "marked_stale": 1, "archived": 0, "reactivated": 0},
|
|
auto_summary="1 marked stale",
|
|
before_report=[],
|
|
before_names=set(),
|
|
after_report=[],
|
|
llm_meta=_make_llm_meta(),
|
|
)
|
|
assert run_dir is not None
|
|
assert run_dir.is_dir()
|
|
assert (run_dir / "run.json").exists()
|
|
assert (run_dir / "REPORT.md").exists()
|
|
|
|
# The directory name is a timestamp under logs/curator/
|
|
assert run_dir.parent == curator._reports_root()
|
|
|
|
|
|
def test_run_json_has_expected_shape(curator_env):
|
|
"""run.json must carry the machine-readable fields downstream tooling needs."""
|
|
curator = curator_env["curator"]
|
|
start = datetime.now(timezone.utc)
|
|
|
|
before_report = [
|
|
{"name": "old-thing", "state": "active", "pinned": False},
|
|
{"name": "keeper", "state": "active", "pinned": True},
|
|
]
|
|
after_report = [
|
|
{"name": "keeper", "state": "active", "pinned": True},
|
|
{"name": "new-umbrella", "state": "active", "pinned": False},
|
|
]
|
|
|
|
run_dir = curator._write_run_report(
|
|
started_at=start,
|
|
elapsed_seconds=42.0,
|
|
auto_counts={"checked": 2, "marked_stale": 0, "archived": 0, "reactivated": 0},
|
|
auto_summary="no changes",
|
|
before_report=before_report,
|
|
before_names={r["name"] for r in before_report},
|
|
after_report=after_report,
|
|
llm_meta=_make_llm_meta(
|
|
final="I consolidated the whole universe.",
|
|
tool_calls=[
|
|
{"name": "skills_list", "arguments": "{}"},
|
|
{"name": "skill_manage", "arguments": '{"action":"create"}'},
|
|
{"name": "terminal", "arguments": "mv ..."},
|
|
],
|
|
),
|
|
)
|
|
payload = json.loads((run_dir / "run.json").read_text())
|
|
|
|
# top-level shape
|
|
for k in (
|
|
"started_at", "duration_seconds", "model", "provider",
|
|
"auto_transitions", "counts", "tool_call_counts",
|
|
"archived", "added", "state_transitions",
|
|
"llm_final", "llm_summary", "llm_error", "tool_calls",
|
|
):
|
|
assert k in payload, f"missing key: {k}"
|
|
|
|
# Diff logic
|
|
assert payload["archived"] == ["old-thing"]
|
|
assert payload["added"] == ["new-umbrella"]
|
|
# Counts reflect the diff
|
|
assert payload["counts"]["before"] == 2
|
|
assert payload["counts"]["after"] == 2
|
|
assert payload["counts"]["archived_this_run"] == 1
|
|
assert payload["counts"]["added_this_run"] == 1
|
|
# Tool call counts are aggregated
|
|
assert payload["tool_call_counts"]["skills_list"] == 1
|
|
assert payload["tool_call_counts"]["skill_manage"] == 1
|
|
assert payload["tool_call_counts"]["terminal"] == 1
|
|
assert payload["counts"]["tool_calls_total"] == 3
|
|
|
|
|
|
def test_report_md_is_human_readable(curator_env):
|
|
"""REPORT.md should be a valid markdown doc with the key sections visible."""
|
|
curator = curator_env["curator"]
|
|
start = datetime.now(timezone.utc)
|
|
|
|
run_dir = curator._write_run_report(
|
|
started_at=start,
|
|
elapsed_seconds=75.0,
|
|
auto_counts={"checked": 10, "marked_stale": 2, "archived": 1, "reactivated": 0},
|
|
auto_summary="2 marked stale, 1 archived",
|
|
before_report=[{"name": "foo", "state": "active", "pinned": False}],
|
|
before_names={"foo"},
|
|
after_report=[{"name": "foo-umbrella", "state": "active", "pinned": False}],
|
|
llm_meta=_make_llm_meta(
|
|
final="Consolidated foo-like skills into foo-umbrella.",
|
|
model="claude-opus-4.7",
|
|
provider="openrouter",
|
|
tool_calls=[
|
|
# Evidence that `foo` was absorbed into `foo-umbrella`:
|
|
# write_file under foo-umbrella referencing foo.
|
|
{
|
|
"name": "skill_manage",
|
|
"arguments": json.dumps({
|
|
"action": "write_file",
|
|
"name": "foo-umbrella",
|
|
"file_path": "references/foo.md",
|
|
"file_content": "# foo\nContent absorbed from the old foo skill.\n",
|
|
}),
|
|
},
|
|
],
|
|
),
|
|
)
|
|
md = (run_dir / "REPORT.md").read_text()
|
|
|
|
# Structural checks
|
|
assert "# Curator run" in md
|
|
assert "Auto-transitions" in md
|
|
assert "LLM consolidation pass" in md
|
|
assert "Recovery" in md
|
|
|
|
# The model / provider we passed in show up
|
|
assert "claude-opus-4.7" in md
|
|
assert "openrouter" in md
|
|
|
|
# The consolidated/added lists are present with clear language
|
|
assert "Consolidated into umbrella skills" in md
|
|
assert "`foo`" in md
|
|
assert "merged into" in md
|
|
assert "`foo-umbrella`" in md
|
|
assert "New skills this run" in md
|
|
|
|
# The full LLM final response is included verbatim (no 240-char truncation)
|
|
assert "Consolidated foo-like skills into foo-umbrella." in md
|
|
|
|
|
|
def test_same_second_reruns_get_unique_dirs(curator_env):
|
|
"""If the curator somehow runs twice in the same second, the second
|
|
report still gets its own directory rather than overwriting the first."""
|
|
curator = curator_env["curator"]
|
|
start = datetime(2026, 4, 29, 5, 33, 34, tzinfo=timezone.utc)
|
|
|
|
kwargs = dict(
|
|
started_at=start,
|
|
elapsed_seconds=1.0,
|
|
auto_counts={"checked": 0, "marked_stale": 0, "archived": 0, "reactivated": 0},
|
|
auto_summary="no changes",
|
|
before_report=[],
|
|
before_names=set(),
|
|
after_report=[],
|
|
llm_meta=_make_llm_meta(),
|
|
)
|
|
a = curator._write_run_report(**kwargs)
|
|
b = curator._write_run_report(**kwargs)
|
|
assert a != b
|
|
assert a is not None and b is not None
|
|
# Second dir has a numeric disambiguator suffix
|
|
assert b.name.startswith(a.name)
|
|
|
|
|
|
def test_report_captures_llm_error_and_continues(curator_env):
|
|
"""If the LLM pass recorded an error, the report still writes and
|
|
surfaces the error prominently."""
|
|
curator = curator_env["curator"]
|
|
run_dir = curator._write_run_report(
|
|
started_at=datetime.now(timezone.utc),
|
|
elapsed_seconds=2.0,
|
|
auto_counts={"checked": 0, "marked_stale": 0, "archived": 0, "reactivated": 0},
|
|
auto_summary="no changes",
|
|
before_report=[],
|
|
before_names=set(),
|
|
after_report=[],
|
|
llm_meta=_make_llm_meta(
|
|
error="HTTP 400: No models provided",
|
|
final="",
|
|
summary="error",
|
|
),
|
|
)
|
|
md = (run_dir / "REPORT.md").read_text()
|
|
assert "HTTP 400" in md
|
|
payload = json.loads((run_dir / "run.json").read_text())
|
|
assert payload["llm_error"] == "HTTP 400: No models provided"
|
|
|
|
|
|
def test_state_transitions_captured_in_report(curator_env):
|
|
"""When a skill moves active → stale or stale → archived between
|
|
before/after snapshots, the report records it."""
|
|
curator = curator_env["curator"]
|
|
start = datetime.now(timezone.utc)
|
|
|
|
before = [{"name": "getting-old", "state": "active", "pinned": False}]
|
|
after = [{"name": "getting-old", "state": "stale", "pinned": False}]
|
|
|
|
run_dir = curator._write_run_report(
|
|
started_at=start,
|
|
elapsed_seconds=1.0,
|
|
auto_counts={"checked": 1, "marked_stale": 1, "archived": 0, "reactivated": 0},
|
|
auto_summary="1 marked stale",
|
|
before_report=before,
|
|
before_names={r["name"] for r in before},
|
|
after_report=after,
|
|
llm_meta=_make_llm_meta(),
|
|
)
|
|
payload = json.loads((run_dir / "run.json").read_text())
|
|
assert payload["state_transitions"] == [
|
|
{"name": "getting-old", "from": "active", "to": "stale"}
|
|
]
|
|
md = (run_dir / "REPORT.md").read_text()
|
|
assert "State transitions" in md
|
|
assert "getting-old" in md
|
|
assert "active → stale" in md
|
|
|
|
|
|
# ---------------------------------------------------------------------------
|
|
# Cron job skill reference rewriting (curator ↔ cron integration)
|
|
# ---------------------------------------------------------------------------
|
|
#
|
|
# When the curator consolidates skill X into umbrella Y during a run, any
|
|
# cron job that listed X in its ``skills`` field would fail to load X at
|
|
# run time — the scheduler logs a warning and skips it, so the scheduled
|
|
# job runs without the instructions it was scheduled to follow. These
|
|
# tests verify that _write_run_report calls into cron.jobs to repair
|
|
# those references and records what it did in both run.json and
|
|
# cron_rewrites.json.
|
|
|
|
|
|
@pytest.fixture
|
|
def curator_env_with_cron(curator_env, monkeypatch):
|
|
"""Extend curator_env with an initialized + repointed cron.jobs module."""
|
|
home = curator_env["home"]
|
|
(home / "cron").mkdir(exist_ok=True)
|
|
(home / "cron" / "output").mkdir(exist_ok=True)
|
|
|
|
import importlib
|
|
import cron.jobs as jobs_mod
|
|
importlib.reload(jobs_mod)
|
|
monkeypatch.setattr(jobs_mod, "HERMES_DIR", home)
|
|
monkeypatch.setattr(jobs_mod, "CRON_DIR", home / "cron")
|
|
monkeypatch.setattr(jobs_mod, "JOBS_FILE", home / "cron" / "jobs.json")
|
|
monkeypatch.setattr(jobs_mod, "OUTPUT_DIR", home / "cron" / "output")
|
|
|
|
return {**curator_env, "jobs": jobs_mod}
|
|
|
|
|
|
def test_curator_rewrites_cron_skills_when_skill_consolidated(curator_env_with_cron):
|
|
"""A skill consolidated into an umbrella should be rewritten in any
|
|
cron job's skills list; the rewrite should be visible in run.json
|
|
and cron_rewrites.json."""
|
|
curator = curator_env_with_cron["curator"]
|
|
jobs = curator_env_with_cron["jobs"]
|
|
|
|
# Create a cron job that depends on a soon-to-be-consolidated skill
|
|
job = jobs.create_job(
|
|
prompt="",
|
|
schedule="every 1h",
|
|
skills=["foo"],
|
|
name="foo-watcher",
|
|
)
|
|
|
|
# Simulate a curator pass that consolidated `foo` → `foo-umbrella`
|
|
before = [{"name": "foo", "state": "active", "pinned": False}]
|
|
after = [{"name": "foo-umbrella", "state": "active", "pinned": False}]
|
|
|
|
run_dir = curator._write_run_report(
|
|
started_at=datetime.now(timezone.utc),
|
|
elapsed_seconds=3.0,
|
|
auto_counts={"checked": 1, "marked_stale": 0, "archived": 0, "reactivated": 0},
|
|
auto_summary="no changes",
|
|
before_report=before,
|
|
before_names={"foo"},
|
|
after_report=after,
|
|
llm_meta=_make_llm_meta(
|
|
final="Consolidated foo into foo-umbrella.",
|
|
tool_calls=[
|
|
{
|
|
"name": "skill_manage",
|
|
"arguments": json.dumps({
|
|
"action": "write_file",
|
|
"name": "foo-umbrella",
|
|
"file_path": "references/foo.md",
|
|
"file_content": "from foo",
|
|
}),
|
|
},
|
|
],
|
|
),
|
|
)
|
|
|
|
# Cron job is rewritten on disk
|
|
loaded = jobs.get_job(job["id"])
|
|
assert loaded["skills"] == ["foo-umbrella"]
|
|
assert loaded["skill"] == "foo-umbrella"
|
|
|
|
# Rewrite is recorded in run.json
|
|
payload = json.loads((run_dir / "run.json").read_text())
|
|
assert payload["cron_rewrites"]["jobs_updated"] == 1
|
|
assert payload["counts"]["cron_jobs_rewritten"] == 1
|
|
rewrites = payload["cron_rewrites"]["rewrites"]
|
|
assert len(rewrites) == 1
|
|
assert rewrites[0]["mapped"] == {"foo": "foo-umbrella"}
|
|
|
|
# Separate cron_rewrites.json is written for convenience
|
|
cron_file = run_dir / "cron_rewrites.json"
|
|
assert cron_file.exists()
|
|
detail = json.loads(cron_file.read_text())
|
|
assert detail["jobs_updated"] == 1
|
|
|
|
# Markdown surfaces the change
|
|
md = (run_dir / "REPORT.md").read_text()
|
|
assert "Cron job skill references rewritten" in md
|
|
assert "foo-watcher" in md
|
|
assert "foo-umbrella" in md
|
|
|
|
|
|
def test_curator_drops_pruned_skill_from_cron_job(curator_env_with_cron):
|
|
"""A pruned (no-umbrella) skill should be dropped from the cron
|
|
job's skill list entirely — there's no forwarding target."""
|
|
curator = curator_env_with_cron["curator"]
|
|
jobs = curator_env_with_cron["jobs"]
|
|
|
|
job = jobs.create_job(
|
|
prompt="",
|
|
schedule="every 1h",
|
|
skills=["keep", "stale-one"],
|
|
)
|
|
|
|
before = [{"name": "stale-one", "state": "active", "pinned": False}]
|
|
after: list = [] # stale-one was archived with no target
|
|
|
|
run_dir = curator._write_run_report(
|
|
started_at=datetime.now(timezone.utc),
|
|
elapsed_seconds=1.0,
|
|
auto_counts={"checked": 1, "marked_stale": 0, "archived": 1, "reactivated": 0},
|
|
auto_summary="1 archived",
|
|
before_report=before,
|
|
before_names={"stale-one"},
|
|
after_report=after,
|
|
llm_meta=_make_llm_meta(), # no tool calls → classifier marks it pruned
|
|
)
|
|
|
|
loaded = jobs.get_job(job["id"])
|
|
assert loaded["skills"] == ["keep"]
|
|
|
|
payload = json.loads((run_dir / "run.json").read_text())
|
|
assert payload["cron_rewrites"]["jobs_updated"] == 1
|
|
rewrites = payload["cron_rewrites"]["rewrites"]
|
|
assert rewrites[0]["dropped"] == ["stale-one"]
|
|
|
|
|
|
def test_curator_report_has_no_cron_section_when_nothing_changes(curator_env_with_cron):
|
|
"""When the curator run doesn't touch any skills, cron jobs are
|
|
untouched and cron_rewrites.json is not even written."""
|
|
curator = curator_env_with_cron["curator"]
|
|
jobs = curator_env_with_cron["jobs"]
|
|
|
|
jobs.create_job(prompt="", schedule="every 1h", skills=["foo"])
|
|
|
|
run_dir = curator._write_run_report(
|
|
started_at=datetime.now(timezone.utc),
|
|
elapsed_seconds=1.0,
|
|
auto_counts={"checked": 0, "marked_stale": 0, "archived": 0, "reactivated": 0},
|
|
auto_summary="no changes",
|
|
before_report=[{"name": "foo", "state": "active", "pinned": False}],
|
|
before_names={"foo"},
|
|
after_report=[{"name": "foo", "state": "active", "pinned": False}],
|
|
llm_meta=_make_llm_meta(),
|
|
)
|
|
|
|
# No rewrites → no separate file, no section in md
|
|
assert not (run_dir / "cron_rewrites.json").exists()
|
|
md = (run_dir / "REPORT.md").read_text()
|
|
assert "Cron job skill references rewritten" not in md
|
|
|
|
payload = json.loads((run_dir / "run.json").read_text())
|
|
assert payload["cron_rewrites"]["jobs_updated"] == 0
|
|
assert payload["counts"]["cron_jobs_rewritten"] == 0
|