mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-04-28 01:21:43 +00:00
* fix(terminal): three-layer defense against watch_patterns notification spam Background processes that stack notify_on_complete=True with watch_patterns can flood the user with duplicate, delayed notifications — matches deliver asynchronously via the completion queue and continue arriving minutes after the process has exited. The docstring warning against this (PR #12113) has proven insufficient; agents still misuse the combination. Three layered defenses, each sufficient on its own: 1. Mutual exclusion (terminal_tool.py): When both flags are set on a background process, drop watch_patterns with a warning. notify_on_complete wins because 'let me know when it's done' is the more useful signal and fires exactly once. Extracted as _resolve_notification_flag_conflict() so the rule is testable in isolation. 2. Suppress-after-exit (process_registry.py): _check_watch_patterns() now bails the moment session.exited is True. Post-exit chunks (buffered reads draining after the process is gone) no longer produce notifications. This is the fix flagged as future work in session 20260418_020302_79881c. 3. Global circuit breaker (process_registry.py): Per-session rate limits don't catch the sibling-flood case — N concurrent processes can each stay under 8/10s and still collectively spam. New WATCH_GLOBAL_MAX_PER_WINDOW=15 cap trips a 30-second cooldown across ALL sessions, emits a single watch_overflow_tripped event, silently counts dropped events, and emits a watch_overflow_released summary when the cooldown ends. Also updates the tool schema + docstring to document the new behavior. Tests: 8 new tests covering all three fixes (suppress-after-exit x2, mutual-exclusion resolver x4, global breaker trip/cooldown/release x2). All 60 tests across test_watch_patterns.py, test_notify_on_complete.py, test_terminal_tool.py pass. Real-world trigger: self-inflicted in session 20260425_051924 — three concurrent hermes-sweeper review subprocesses each set watch_patterns= ['failed validation', 'errored'] AND notify_on_complete=True, then iterated over multiple items, producing enough matches per process to defeat the per-session cap while staying under the global cap that didn't yet exist. * fix(terminal): aggressive 1-per-15s watch_patterns rate limit + strike-3 promotion Per Teknium's direction, the watch_patterns rate limit is now much more aggressive and self-healing. ## New rule — per session - HARD cap: 1 watch-match notification per 15 seconds per process. - Any match arriving inside the cooldown window is dropped and counts as ONE strike for that window (many drops in the same window still = 1 strike). - After 3 consecutive strike windows, watch_patterns is permanently disabled for the session and the session is auto-promoted to notify_on_complete semantics — exactly one notification when the process actually exits. - A cooldown window that expires with zero drops resets the consecutive strike counter — healthy cadence is forgiven. ## Schema + docstring rewritten The tool schema description now gives the model explicit guidance: - notify_on_complete is 'the right choice for almost every long-running task' - watch_patterns is for RARE one-shot signals on LONG-LIVED processes - Do NOT use watch_patterns with loops/batch jobs — error patterns fire every iteration and will hit the strike limit fast - Mutual exclusion is stated on both parameter descriptions - 1/15s cooldown and 3-strike promotion are stated in the watch_patterns description so the model sees the contract every turn ## Removed - WATCH_MAX_PER_WINDOW (8/10s) and WATCH_OVERLOAD_KILL_SECONDS (45) — the new 1/15s limit subsumes both; keeping them would double-count. - _watch_window_hits / _watch_window_start / _watch_overload_since fields on ProcessSession. Replaced by _watch_last_emit_at / _watch_cooldown_until / _watch_strike_candidate / _watch_consecutive_strikes. ## Kept - Global circuit breaker across all sessions (15/10s → 30s cooldown) as a secondary safety net for concurrent siblings. Still valuable when 20 short-lived processes each fire once — none individually violates the per-session limit. - Suppress-after-exit guard. - Mutual exclusion resolver at the tool entry point. ## Tests - 6 new tests in TestPerSessionRateLimit covering: first match delivers, second in cooldown suppressed, multi-drop = single strike, 3 strikes disables + promotes, clean window resets counter, suppressed count carried to next emit. - Global circuit breaker tests rewritten to use fresh sessions instead of hacking removed per-window fields. - 50/50 watch_patterns + notify_on_complete tests pass. - 60/60 including test_terminal_tool.py pass.
491 lines
21 KiB
Python
491 lines
21 KiB
Python
"""Tests for watch_patterns background process monitoring feature.
|
|
|
|
Covers:
|
|
- ProcessSession.watch_patterns field
|
|
- ProcessRegistry._check_watch_patterns() matching + notification
|
|
- Rate limiting (WATCH_MAX_PER_WINDOW) and overload kill switch
|
|
- watch_queue population
|
|
- Checkpoint persistence of watch_patterns
|
|
- Terminal tool schema includes watch_patterns
|
|
- Terminal tool handler passes watch_patterns through
|
|
"""
|
|
|
|
import json
|
|
import queue
|
|
import time
|
|
import pytest
|
|
from unittest.mock import patch
|
|
|
|
from tools.process_registry import (
|
|
ProcessRegistry,
|
|
ProcessSession,
|
|
WATCH_MIN_INTERVAL_SECONDS,
|
|
WATCH_STRIKE_LIMIT,
|
|
WATCH_GLOBAL_MAX_PER_WINDOW,
|
|
WATCH_GLOBAL_WINDOW_SECONDS,
|
|
WATCH_GLOBAL_COOLDOWN_SECONDS,
|
|
)
|
|
|
|
|
|
@pytest.fixture()
|
|
def registry():
|
|
"""Create a fresh ProcessRegistry."""
|
|
return ProcessRegistry()
|
|
|
|
|
|
def _make_session(
|
|
sid="proc_test_watch",
|
|
command="tail -f app.log",
|
|
task_id="t1",
|
|
watch_patterns=None,
|
|
) -> ProcessSession:
|
|
s = ProcessSession(
|
|
id=sid,
|
|
command=command,
|
|
task_id=task_id,
|
|
started_at=time.time(),
|
|
watch_patterns=watch_patterns or [],
|
|
)
|
|
return s
|
|
|
|
|
|
# =========================================================================
|
|
# ProcessSession field defaults
|
|
# =========================================================================
|
|
|
|
class TestProcessSessionField:
|
|
def test_default_empty(self):
|
|
s = ProcessSession(id="proc_1", command="echo hi")
|
|
assert s.watch_patterns == []
|
|
assert s._watch_disabled is False
|
|
assert s._watch_hits == 0
|
|
assert s._watch_suppressed == 0
|
|
|
|
def test_can_set_patterns(self):
|
|
s = _make_session(watch_patterns=["ERROR", "WARN"])
|
|
assert s.watch_patterns == ["ERROR", "WARN"]
|
|
|
|
|
|
# =========================================================================
|
|
# Pattern matching + queue population
|
|
# =========================================================================
|
|
|
|
class TestCheckWatchPatterns:
|
|
def test_no_patterns_no_notification(self, registry):
|
|
"""No watch_patterns → no notifications."""
|
|
session = _make_session(watch_patterns=[])
|
|
registry._check_watch_patterns(session, "ERROR: something broke\n")
|
|
assert registry.completion_queue.empty()
|
|
|
|
def test_no_match_no_notification(self, registry):
|
|
"""Output that doesn't match any pattern → no notification."""
|
|
session = _make_session(watch_patterns=["ERROR", "FAIL"])
|
|
registry._check_watch_patterns(session, "INFO: all good\nDEBUG: fine\n")
|
|
assert registry.completion_queue.empty()
|
|
|
|
def test_basic_match(self, registry):
|
|
"""Single matching line triggers a notification."""
|
|
session = _make_session(watch_patterns=["ERROR"])
|
|
registry._check_watch_patterns(session, "INFO: ok\nERROR: disk full\n")
|
|
assert not registry.completion_queue.empty()
|
|
evt = registry.completion_queue.get_nowait()
|
|
assert evt["type"] == "watch_match"
|
|
assert evt["pattern"] == "ERROR"
|
|
assert "disk full" in evt["output"]
|
|
assert evt["session_id"] == "proc_test_watch"
|
|
|
|
def test_match_carries_session_key_and_watcher_routing_metadata(self, registry):
|
|
session = _make_session(watch_patterns=["ERROR"])
|
|
session.session_key = "agent:main:telegram:group:-100:42"
|
|
session.watcher_platform = "telegram"
|
|
session.watcher_chat_id = "-100"
|
|
session.watcher_user_id = "u123"
|
|
session.watcher_user_name = "alice"
|
|
session.watcher_thread_id = "42"
|
|
|
|
registry._check_watch_patterns(session, "ERROR: disk full\n")
|
|
evt = registry.completion_queue.get_nowait()
|
|
|
|
assert evt["session_key"] == "agent:main:telegram:group:-100:42"
|
|
assert evt["platform"] == "telegram"
|
|
assert evt["chat_id"] == "-100"
|
|
assert evt["user_id"] == "u123"
|
|
assert evt["user_name"] == "alice"
|
|
assert evt["thread_id"] == "42"
|
|
|
|
def test_multiple_patterns(self, registry):
|
|
"""First matching pattern is reported."""
|
|
session = _make_session(watch_patterns=["WARN", "ERROR"])
|
|
registry._check_watch_patterns(session, "ERROR: bad\nWARN: hmm\n")
|
|
evt = registry.completion_queue.get_nowait()
|
|
# ERROR appears first in the output, and we check patterns in order
|
|
# so "WARN" won't match "ERROR: bad" but "ERROR" will
|
|
assert evt["pattern"] == "ERROR"
|
|
assert "bad" in evt["output"]
|
|
|
|
def test_disabled_skips(self, registry):
|
|
"""Disabled watch produces no notifications."""
|
|
session = _make_session(watch_patterns=["ERROR"])
|
|
session._watch_disabled = True
|
|
registry._check_watch_patterns(session, "ERROR: boom\n")
|
|
assert registry.completion_queue.empty()
|
|
|
|
def test_hit_counter_increments(self, registry):
|
|
"""Each delivered notification increments _watch_hits.
|
|
|
|
With 1/15s rate limit, we need to reset cooldown between calls.
|
|
"""
|
|
session = _make_session(watch_patterns=["X"])
|
|
registry._check_watch_patterns(session, "X\n")
|
|
assert session._watch_hits == 1
|
|
# Reset cooldown so the second match gets delivered.
|
|
session._watch_cooldown_until = 0.0
|
|
registry._check_watch_patterns(session, "X\n")
|
|
assert session._watch_hits == 2
|
|
|
|
def test_output_truncation(self, registry):
|
|
"""Very long matched output is truncated."""
|
|
session = _make_session(watch_patterns=["X"])
|
|
# Generate 30 matching lines (more than the 20-line cap)
|
|
text = "\n".join(f"X line {i}" for i in range(30)) + "\n"
|
|
registry._check_watch_patterns(session, text)
|
|
evt = registry.completion_queue.get_nowait()
|
|
# Should only have 20 lines max
|
|
assert evt["output"].count("\n") <= 20
|
|
|
|
|
|
# =========================================================================
|
|
# Per-session rate limiting: 1 notification per 15s, 3 strikes → disable
|
|
# =========================================================================
|
|
|
|
class TestPerSessionRateLimit:
|
|
def test_first_match_delivers(self, registry):
|
|
"""A fresh session with no prior cooldown delivers the first match."""
|
|
session = _make_session(watch_patterns=["E"])
|
|
registry._check_watch_patterns(session, "E first\n")
|
|
assert registry.completion_queue.qsize() == 1
|
|
evt = registry.completion_queue.get_nowait()
|
|
assert evt["type"] == "watch_match"
|
|
assert session._watch_hits == 1
|
|
# Cooldown is now armed.
|
|
assert session._watch_cooldown_until > 0
|
|
|
|
def test_second_match_within_cooldown_is_suppressed(self, registry):
|
|
"""A second match inside the 15s cooldown is dropped and counted."""
|
|
session = _make_session(watch_patterns=["E"])
|
|
registry._check_watch_patterns(session, "E first\n")
|
|
assert registry.completion_queue.qsize() == 1
|
|
# Immediately trigger another match — well inside cooldown.
|
|
registry._check_watch_patterns(session, "E second\n")
|
|
# Still only one notification.
|
|
assert registry.completion_queue.qsize() == 1
|
|
assert session._watch_suppressed == 1
|
|
assert session._watch_consecutive_strikes == 1
|
|
|
|
def test_many_drops_inside_window_count_as_ONE_strike(self, registry):
|
|
"""Multiple suppressions inside the same cooldown window = 1 strike."""
|
|
session = _make_session(watch_patterns=["E"])
|
|
registry._check_watch_patterns(session, "E\n")
|
|
for _ in range(10):
|
|
registry._check_watch_patterns(session, "E\n")
|
|
assert session._watch_consecutive_strikes == 1
|
|
assert session._watch_suppressed == 10
|
|
|
|
def test_three_strikes_disables_watch_and_promotes_to_notify(self, registry):
|
|
"""Three consecutive strike windows → watch_disabled + notify_on_complete."""
|
|
session = _make_session(watch_patterns=["E"])
|
|
session.notify_on_complete = False
|
|
|
|
for strike in range(WATCH_STRIKE_LIMIT):
|
|
# Emit → arms cooldown.
|
|
registry._check_watch_patterns(session, f"E emit {strike}\n")
|
|
# Attempt while inside cooldown → one strike, dropped.
|
|
registry._check_watch_patterns(session, f"E drop {strike}\n")
|
|
# Fast-forward past the cooldown for the NEXT iteration, BUT leave
|
|
# the strike candidate set so the cooldown-expiry branch sees
|
|
# "this was a strike window" and doesn't reset the counter.
|
|
session._watch_cooldown_until = time.time() - 0.01
|
|
|
|
# After WATCH_STRIKE_LIMIT strikes, the next attempt should find
|
|
# the session disabled.
|
|
assert session._watch_disabled is True
|
|
assert session.notify_on_complete is True
|
|
# One watch_disabled summary event should be in the queue.
|
|
disabled_evts = []
|
|
matches = 0
|
|
while not registry.completion_queue.empty():
|
|
evt = registry.completion_queue.get_nowait()
|
|
if evt.get("type") == "watch_disabled":
|
|
disabled_evts.append(evt)
|
|
elif evt.get("type") == "watch_match":
|
|
matches += 1
|
|
assert len(disabled_evts) == 1
|
|
assert "notify_on_complete" in disabled_evts[0]["message"]
|
|
# We should have had exactly WATCH_STRIKE_LIMIT emissions before disable.
|
|
assert matches == WATCH_STRIKE_LIMIT
|
|
|
|
def test_clean_window_resets_strike_counter(self, registry):
|
|
"""A cooldown that expires with zero drops resets the consecutive counter."""
|
|
session = _make_session(watch_patterns=["E"])
|
|
# Emit + drop inside window → 1 strike.
|
|
registry._check_watch_patterns(session, "E emit\n")
|
|
registry._check_watch_patterns(session, "E drop\n")
|
|
assert session._watch_consecutive_strikes == 1
|
|
|
|
# Fast-forward past cooldown. No match arrived during the window —
|
|
# strike_candidate stays False from the prior window's reset, but
|
|
# it was True during that window. On the NEXT emission, the
|
|
# cooldown-expiry branch checks strike_candidate. Since we emitted
|
|
# at the start of this new window and no drop has happened, the
|
|
# reset branch should fire.
|
|
session._watch_cooldown_until = time.time() - 0.01
|
|
# Clear strike candidate to simulate "this cooldown had no drops".
|
|
session._watch_strike_candidate = False
|
|
registry._check_watch_patterns(session, "E clean\n")
|
|
assert session._watch_consecutive_strikes == 0
|
|
|
|
def test_suppressed_count_in_next_delivery(self, registry):
|
|
"""Suppressed count from a strike window is reported in the next emit."""
|
|
session = _make_session(watch_patterns=["E"])
|
|
registry._check_watch_patterns(session, "E emit\n")
|
|
for _ in range(4):
|
|
registry._check_watch_patterns(session, "E drop\n")
|
|
assert session._watch_suppressed == 4
|
|
|
|
# Fast-forward past cooldown.
|
|
session._watch_cooldown_until = time.time() - 0.01
|
|
# Drain the queue so we can inspect the next emission.
|
|
while not registry.completion_queue.empty():
|
|
registry.completion_queue.get_nowait()
|
|
|
|
registry._check_watch_patterns(session, "E back\n")
|
|
evt = registry.completion_queue.get_nowait()
|
|
assert evt["type"] == "watch_match"
|
|
assert evt["suppressed"] == 4
|
|
assert session._watch_suppressed == 0 # reset after delivery
|
|
|
|
|
|
# =========================================================================
|
|
# Checkpoint persistence
|
|
# =========================================================================
|
|
|
|
class TestCheckpointPersistence:
|
|
def test_watch_patterns_in_checkpoint(self, registry):
|
|
"""watch_patterns is included in checkpoint data."""
|
|
session = _make_session(watch_patterns=["ERROR", "FAIL"])
|
|
with registry._lock:
|
|
registry._running[session.id] = session
|
|
|
|
with patch("utils.atomic_json_write") as mock_write:
|
|
registry._write_checkpoint()
|
|
args = mock_write.call_args
|
|
entries = args[0][1] # second positional arg
|
|
assert len(entries) == 1
|
|
assert entries[0]["watch_patterns"] == ["ERROR", "FAIL"]
|
|
|
|
def test_watch_patterns_recovery(self, registry, tmp_path, monkeypatch):
|
|
"""watch_patterns survives checkpoint recovery."""
|
|
import tools.process_registry as pr_mod
|
|
checkpoint = tmp_path / "processes.json"
|
|
checkpoint.write_text(json.dumps([{
|
|
"session_id": "proc_recovered",
|
|
"command": "tail -f log",
|
|
"pid": 99999999, # non-existent
|
|
"pid_scope": "host",
|
|
"started_at": time.time(),
|
|
"task_id": "",
|
|
"session_key": "",
|
|
"watcher_platform": "",
|
|
"watcher_chat_id": "",
|
|
"watcher_thread_id": "",
|
|
"watcher_interval": 0,
|
|
"notify_on_complete": False,
|
|
"watch_patterns": ["PANIC", "OOM"],
|
|
}]))
|
|
monkeypatch.setattr(pr_mod, "CHECKPOINT_PATH", checkpoint)
|
|
# PID doesn't exist, so nothing will be recovered
|
|
count = registry.recover_from_checkpoint()
|
|
# Won't recover since PID is fake, but verify the code path doesn't crash
|
|
assert count == 0
|
|
|
|
|
|
# =========================================================================
|
|
# Terminal tool schema + handler
|
|
# =========================================================================
|
|
|
|
class TestTerminalToolSchema:
|
|
def test_schema_includes_watch_patterns(self):
|
|
from tools.terminal_tool import TERMINAL_SCHEMA
|
|
props = TERMINAL_SCHEMA["parameters"]["properties"]
|
|
assert "watch_patterns" in props
|
|
assert props["watch_patterns"]["type"] == "array"
|
|
assert props["watch_patterns"]["items"] == {"type": "string"}
|
|
|
|
def test_handler_passes_watch_patterns(self):
|
|
"""_handle_terminal passes watch_patterns to terminal_tool."""
|
|
from tools.terminal_tool import _handle_terminal
|
|
with patch("tools.terminal_tool.terminal_tool") as mock_tt:
|
|
mock_tt.return_value = json.dumps({"output": "ok", "exit_code": 0})
|
|
_handle_terminal(
|
|
{"command": "echo hi", "watch_patterns": ["ERR"]},
|
|
task_id="t1",
|
|
)
|
|
_, kwargs = mock_tt.call_args
|
|
assert kwargs.get("watch_patterns") == ["ERR"]
|
|
|
|
|
|
# =========================================================================
|
|
# Code execution tool blocked params
|
|
# =========================================================================
|
|
|
|
class TestCodeExecutionBlocked:
|
|
def test_watch_patterns_blocked(self):
|
|
from tools.code_execution_tool import _TERMINAL_BLOCKED_PARAMS
|
|
assert "watch_patterns" in _TERMINAL_BLOCKED_PARAMS
|
|
|
|
|
|
# =========================================================================
|
|
# Suppress-after-exit (anti-spam fix)
|
|
# =========================================================================
|
|
|
|
class TestSuppressAfterExit:
|
|
def test_match_dropped_once_session_exited(self, registry):
|
|
"""watch_patterns notifications stop the moment session.exited is set."""
|
|
session = _make_session(watch_patterns=["ERROR"])
|
|
# Mark the process as exited BEFORE the late chunk arrives.
|
|
session.exited = True
|
|
registry._check_watch_patterns(session, "ERROR: late buffer\n")
|
|
assert registry.completion_queue.empty()
|
|
assert session._watch_hits == 0
|
|
|
|
def test_match_still_delivered_while_session_running(self, registry):
|
|
"""Sanity: while the process is still running, matches still deliver."""
|
|
session = _make_session(watch_patterns=["ERROR"])
|
|
session.exited = False
|
|
registry._check_watch_patterns(session, "ERROR: oh no\n")
|
|
assert not registry.completion_queue.empty()
|
|
evt = registry.completion_queue.get_nowait()
|
|
assert evt["type"] == "watch_match"
|
|
|
|
|
|
# =========================================================================
|
|
# Mutual exclusion: notify_on_complete wins over watch_patterns
|
|
# =========================================================================
|
|
|
|
class TestMutualExclusion:
|
|
def test_resolver_drops_watch_when_notify_set(self):
|
|
"""Both flags set → watch_patterns dropped with a note."""
|
|
from tools.terminal_tool import _resolve_notification_flag_conflict
|
|
|
|
resolved, note = _resolve_notification_flag_conflict(
|
|
notify_on_complete=True,
|
|
watch_patterns=["ERROR", "DONE"],
|
|
background=True,
|
|
)
|
|
assert resolved is None
|
|
assert "notify_on_complete" in note
|
|
assert "duplicate notifications" in note
|
|
|
|
def test_resolver_keeps_watch_when_notify_off(self):
|
|
"""notify_on_complete=False → watch_patterns kept intact."""
|
|
from tools.terminal_tool import _resolve_notification_flag_conflict
|
|
|
|
resolved, note = _resolve_notification_flag_conflict(
|
|
notify_on_complete=False,
|
|
watch_patterns=["ERROR"],
|
|
background=True,
|
|
)
|
|
assert resolved == ["ERROR"]
|
|
assert note == ""
|
|
|
|
def test_resolver_keeps_notify_when_no_watch(self):
|
|
"""Only notify_on_complete set → no conflict."""
|
|
from tools.terminal_tool import _resolve_notification_flag_conflict
|
|
|
|
resolved, note = _resolve_notification_flag_conflict(
|
|
notify_on_complete=True,
|
|
watch_patterns=None,
|
|
background=True,
|
|
)
|
|
assert resolved is None
|
|
assert note == ""
|
|
|
|
def test_resolver_inert_when_not_background(self):
|
|
"""Without background=True, the whole thing is a no-op."""
|
|
from tools.terminal_tool import _resolve_notification_flag_conflict
|
|
|
|
resolved, note = _resolve_notification_flag_conflict(
|
|
notify_on_complete=True,
|
|
watch_patterns=["ERROR"],
|
|
background=False,
|
|
)
|
|
assert resolved == ["ERROR"]
|
|
assert note == ""
|
|
|
|
|
|
# =========================================================================
|
|
# Global circuit breaker (cross-session overflow blocker)
|
|
# =========================================================================
|
|
|
|
class TestGlobalCircuitBreaker:
|
|
def test_trips_after_global_threshold(self, registry):
|
|
"""When >N matches fire across sessions in the window, breaker trips."""
|
|
sessions = [
|
|
_make_session(sid=f"proc_s{i}", watch_patterns=["E"])
|
|
for i in range(WATCH_GLOBAL_MAX_PER_WINDOW + 3)
|
|
]
|
|
# Each session fires exactly one match — individually well under the
|
|
# per-session cap. But collectively they should trip the global cap.
|
|
for s in sessions:
|
|
registry._check_watch_patterns(s, "E hit\n")
|
|
|
|
# Drain the queue and count event types.
|
|
watch_matches = 0
|
|
overflow_tripped = 0
|
|
while not registry.completion_queue.empty():
|
|
evt = registry.completion_queue.get_nowait()
|
|
if evt.get("type") == "watch_match":
|
|
watch_matches += 1
|
|
elif evt.get("type") == "watch_overflow_tripped":
|
|
overflow_tripped += 1
|
|
assert watch_matches == WATCH_GLOBAL_MAX_PER_WINDOW
|
|
assert overflow_tripped == 1
|
|
assert registry._global_watch_tripped_until > 0
|
|
|
|
def test_cooldown_suppresses_and_then_releases(self, registry):
|
|
"""After trip, further events are suppressed; cooldown expiry emits release."""
|
|
# Spawn enough fresh sessions to trip the global breaker.
|
|
sessions = [
|
|
_make_session(sid=f"proc_t{i}", watch_patterns=["E"])
|
|
for i in range(WATCH_GLOBAL_MAX_PER_WINDOW + 1)
|
|
]
|
|
for s in sessions:
|
|
registry._check_watch_patterns(s, "E hit\n")
|
|
assert registry._global_watch_tripped_until > 0
|
|
|
|
# Further matches from BRAND-NEW sessions during cooldown are dropped.
|
|
q_size_before = registry.completion_queue.qsize()
|
|
extra1 = _make_session(sid="proc_extra1", watch_patterns=["E"])
|
|
extra2 = _make_session(sid="proc_extra2", watch_patterns=["E"])
|
|
registry._check_watch_patterns(extra1, "E hit\n")
|
|
registry._check_watch_patterns(extra2, "E hit\n")
|
|
assert registry.completion_queue.qsize() == q_size_before # no new events
|
|
assert registry._global_watch_suppressed_during_trip >= 2
|
|
|
|
# Simulate cooldown expiry.
|
|
registry._global_watch_tripped_until = time.time() - 1
|
|
|
|
# Next call admits AND emits the release summary.
|
|
released_session = _make_session(sid="proc_after", watch_patterns=["E"])
|
|
registry._check_watch_patterns(released_session, "E hit\n")
|
|
released = False
|
|
admitted = False
|
|
while not registry.completion_queue.empty():
|
|
evt = registry.completion_queue.get_nowait()
|
|
if evt.get("type") == "watch_overflow_released":
|
|
released = True
|
|
assert evt["suppressed"] >= 2
|
|
elif evt.get("type") == "watch_match":
|
|
admitted = True
|
|
assert released
|
|
assert admitted
|