Comprehensive cleanup across 80 files based on automated (ruff, pyflakes, vulture)
and manual analysis of the entire codebase.
Changes by category:
Unused imports removed (~95 across 55 files):
- Removed genuinely unused imports from all major subsystems
- agent/, hermes_cli/, tools/, gateway/, plugins/, cron/
- Includes imports in try/except blocks that were truly unused
(vs availability checks which were left alone)
Unused variables removed (~25):
- Removed dead variables: connected, inner, channels, last_exc,
source, new_server_names, verify, pconfig, default_terminal,
result, pending_handled, temperature, loop
- Dropped unused argparse subparser assignments in hermes_cli/main.py
(12 instances of add_parser() where result was never used)
Dead code removed:
- run_agent.py: Removed dead ternary (None if False else None) and
surrounding unreachable branch in identity fallback
- run_agent.py: Removed write-only attribute _last_reported_tool
- hermes_cli/providers.py: Removed dead @property decorator on
module-level function (decorator has no effect outside a class)
- gateway/run.py: Removed unused MCP config load before reconnect
- gateway/platforms/slack.py: Removed dead SessionSource construction
Undefined name bugs fixed (would cause NameError at runtime):
- batch_runner.py: Added missing logger = logging.getLogger(__name__)
- tools/environments/daytona.py: Added missing Dict and Path imports
Unnecessary global statements removed (14):
- tools/terminal_tool.py: 5 functions declared global for dicts
they only mutated via .pop()/[key]=value (no rebinding)
- tools/browser_tool.py: cleanup thread loop only reads flag
- tools/rl_training_tool.py: 4 functions only do dict mutations
- tools/mcp_oauth.py: only reads the global
- hermes_time.py: only reads cached values
Inefficient patterns fixed:
- startswith/endswith tuple form: 15 instances of
x.startswith('a') or x.startswith('b') consolidated to
x.startswith(('a', 'b'))
- len(x)==0 / len(x)>0: 13 instances replaced with pythonic
truthiness checks (not x / bool(x))
- in dict.keys(): 5 instances simplified to in dict
- Redefined unused name: removed duplicate _strip_mdv2 import in
send_message_tool.py
Other fixes:
- hermes_cli/doctor.py: Replaced undefined logger.debug() with pass
- hermes_cli/config.py: Consolidated chained .endswith() calls
Test results: 3934 passed, 17 failed (all pre-existing on main),
19 skipped. Zero regressions.
336 lines
11 KiB
Python
336 lines
11 KiB
Python
"""``hermes logs`` — view and filter Hermes log files.
|
|
|
|
Supports tailing, following, session filtering, level filtering, and
|
|
relative time ranges. All log files live under ``~/.hermes/logs/``.
|
|
|
|
Usage examples::
|
|
|
|
hermes logs # last 50 lines of agent.log
|
|
hermes logs -f # follow agent.log in real time
|
|
hermes logs errors # last 50 lines of errors.log
|
|
hermes logs gateway -n 100 # last 100 lines of gateway.log
|
|
hermes logs --level WARNING # only WARNING+ lines
|
|
hermes logs --session abc123 # filter by session ID substring
|
|
hermes logs --since 1h # lines from the last hour
|
|
hermes logs --since 30m -f # follow, starting 30 min ago
|
|
"""
|
|
|
|
import re
|
|
import sys
|
|
import time
|
|
from datetime import datetime, timedelta
|
|
from pathlib import Path
|
|
from typing import Optional
|
|
|
|
from hermes_constants import get_hermes_home, display_hermes_home
|
|
|
|
# Known log files (name → filename)
|
|
LOG_FILES = {
|
|
"agent": "agent.log",
|
|
"errors": "errors.log",
|
|
"gateway": "gateway.log",
|
|
}
|
|
|
|
# Log line timestamp regex — matches "2026-04-05 22:35:00,123" or
|
|
# "2026-04-05 22:35:00" at the start of a line.
|
|
_TS_RE = re.compile(r"^(\d{4}-\d{2}-\d{2}\s+\d{2}:\d{2}:\d{2})")
|
|
|
|
# Level extraction — matches " INFO ", " WARNING ", " ERROR ", " DEBUG ", " CRITICAL "
|
|
_LEVEL_RE = re.compile(r"\s(DEBUG|INFO|WARNING|ERROR|CRITICAL)\s")
|
|
|
|
# Level ordering for >= filtering
|
|
_LEVEL_ORDER = {"DEBUG": 0, "INFO": 1, "WARNING": 2, "ERROR": 3, "CRITICAL": 4}
|
|
|
|
|
|
def _parse_since(since_str: str) -> Optional[datetime]:
|
|
"""Parse a relative time string like '1h', '30m', '2d' into a datetime cutoff.
|
|
|
|
Returns None if the string can't be parsed.
|
|
"""
|
|
since_str = since_str.strip().lower()
|
|
match = re.match(r"^(\d+)\s*([smhd])$", since_str)
|
|
if not match:
|
|
return None
|
|
value = int(match.group(1))
|
|
unit = match.group(2)
|
|
delta = {
|
|
"s": timedelta(seconds=value),
|
|
"m": timedelta(minutes=value),
|
|
"h": timedelta(hours=value),
|
|
"d": timedelta(days=value),
|
|
}[unit]
|
|
return datetime.now() - delta
|
|
|
|
|
|
def _parse_line_timestamp(line: str) -> Optional[datetime]:
|
|
"""Extract timestamp from a log line. Returns None if not parseable."""
|
|
m = _TS_RE.match(line)
|
|
if not m:
|
|
return None
|
|
try:
|
|
return datetime.strptime(m.group(1), "%Y-%m-%d %H:%M:%S")
|
|
except ValueError:
|
|
return None
|
|
|
|
|
|
def _extract_level(line: str) -> Optional[str]:
|
|
"""Extract the log level from a line."""
|
|
m = _LEVEL_RE.search(line)
|
|
return m.group(1) if m else None
|
|
|
|
|
|
def _matches_filters(
|
|
line: str,
|
|
*,
|
|
min_level: Optional[str] = None,
|
|
session_filter: Optional[str] = None,
|
|
since: Optional[datetime] = None,
|
|
) -> bool:
|
|
"""Check if a log line passes all active filters."""
|
|
if since is not None:
|
|
ts = _parse_line_timestamp(line)
|
|
if ts is not None and ts < since:
|
|
return False
|
|
|
|
if min_level is not None:
|
|
level = _extract_level(line)
|
|
if level is not None:
|
|
if _LEVEL_ORDER.get(level, 0) < _LEVEL_ORDER.get(min_level, 0):
|
|
return False
|
|
|
|
if session_filter is not None:
|
|
if session_filter not in line:
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
def tail_log(
|
|
log_name: str = "agent",
|
|
*,
|
|
num_lines: int = 50,
|
|
follow: bool = False,
|
|
level: Optional[str] = None,
|
|
session: Optional[str] = None,
|
|
since: Optional[str] = None,
|
|
) -> None:
|
|
"""Read and display log lines, optionally following in real time.
|
|
|
|
Parameters
|
|
----------
|
|
log_name
|
|
Which log to read: ``"agent"``, ``"errors"``, ``"gateway"``.
|
|
num_lines
|
|
Number of recent lines to show (before follow starts).
|
|
follow
|
|
If True, keep watching for new lines (Ctrl+C to stop).
|
|
level
|
|
Minimum log level to show (e.g. ``"WARNING"``).
|
|
session
|
|
Session ID substring to filter on.
|
|
since
|
|
Relative time string (e.g. ``"1h"``, ``"30m"``).
|
|
"""
|
|
filename = LOG_FILES.get(log_name)
|
|
if filename is None:
|
|
print(f"Unknown log: {log_name!r}. Available: {', '.join(sorted(LOG_FILES))}")
|
|
sys.exit(1)
|
|
|
|
log_path = get_hermes_home() / "logs" / filename
|
|
if not log_path.exists():
|
|
print(f"Log file not found: {log_path}")
|
|
print(f"(Logs are created when Hermes runs — try 'hermes chat' first)")
|
|
sys.exit(1)
|
|
|
|
# Parse --since into a datetime cutoff
|
|
since_dt = None
|
|
if since:
|
|
since_dt = _parse_since(since)
|
|
if since_dt is None:
|
|
print(f"Invalid --since value: {since!r}. Use format like '1h', '30m', '2d'.")
|
|
sys.exit(1)
|
|
|
|
min_level = level.upper() if level else None
|
|
if min_level and min_level not in _LEVEL_ORDER:
|
|
print(f"Invalid --level: {level!r}. Use DEBUG, INFO, WARNING, ERROR, or CRITICAL.")
|
|
sys.exit(1)
|
|
|
|
has_filters = min_level is not None or session is not None or since_dt is not None
|
|
|
|
# Read and display the tail
|
|
try:
|
|
lines = _read_tail(log_path, num_lines, has_filters=has_filters,
|
|
min_level=min_level, session_filter=session,
|
|
since=since_dt)
|
|
except PermissionError:
|
|
print(f"Permission denied: {log_path}")
|
|
sys.exit(1)
|
|
|
|
# Print header
|
|
filter_parts = []
|
|
if min_level:
|
|
filter_parts.append(f"level>={min_level}")
|
|
if session:
|
|
filter_parts.append(f"session={session}")
|
|
if since:
|
|
filter_parts.append(f"since={since}")
|
|
filter_desc = f" [{', '.join(filter_parts)}]" if filter_parts else ""
|
|
|
|
if follow:
|
|
print(f"--- {display_hermes_home()}/logs/{filename}{filter_desc} (Ctrl+C to stop) ---")
|
|
else:
|
|
print(f"--- {display_hermes_home()}/logs/{filename}{filter_desc} (last {num_lines}) ---")
|
|
|
|
for line in lines:
|
|
print(line, end="")
|
|
|
|
if not follow:
|
|
return
|
|
|
|
# Follow mode — poll for new content
|
|
try:
|
|
_follow_log(log_path, min_level=min_level, session_filter=session,
|
|
since=since_dt)
|
|
except KeyboardInterrupt:
|
|
print("\n--- stopped ---")
|
|
|
|
|
|
def _read_tail(
|
|
path: Path,
|
|
num_lines: int,
|
|
*,
|
|
has_filters: bool = False,
|
|
min_level: Optional[str] = None,
|
|
session_filter: Optional[str] = None,
|
|
since: Optional[datetime] = None,
|
|
) -> list:
|
|
"""Read the last *num_lines* matching lines from a log file.
|
|
|
|
When filters are active, we read more raw lines to find enough matches.
|
|
"""
|
|
if has_filters:
|
|
# Read more lines to ensure we get enough after filtering.
|
|
# For large files, read last 10K lines and filter down.
|
|
raw_lines = _read_last_n_lines(path, max(num_lines * 20, 2000))
|
|
filtered = [
|
|
l for l in raw_lines
|
|
if _matches_filters(l, min_level=min_level,
|
|
session_filter=session_filter, since=since)
|
|
]
|
|
return filtered[-num_lines:]
|
|
else:
|
|
return _read_last_n_lines(path, num_lines)
|
|
|
|
|
|
def _read_last_n_lines(path: Path, n: int) -> list:
|
|
"""Efficiently read the last N lines from a file.
|
|
|
|
For files under 1MB, reads the whole file (fast, simple).
|
|
For larger files, reads chunks from the end.
|
|
"""
|
|
try:
|
|
size = path.stat().st_size
|
|
if size == 0:
|
|
return []
|
|
|
|
# For files up to 1MB, just read the whole thing — simple and correct.
|
|
if size <= 1_048_576:
|
|
with open(path, "r", encoding="utf-8", errors="replace") as f:
|
|
all_lines = f.readlines()
|
|
return all_lines[-n:]
|
|
|
|
# For large files, read chunks from the end.
|
|
with open(path, "rb") as f:
|
|
chunk_size = 8192
|
|
lines = []
|
|
pos = size
|
|
|
|
while pos > 0 and len(lines) <= n + 1:
|
|
read_size = min(chunk_size, pos)
|
|
pos -= read_size
|
|
f.seek(pos)
|
|
chunk = f.read(read_size)
|
|
chunk_lines = chunk.split(b"\n")
|
|
if lines:
|
|
# Merge the last partial line of the new chunk with the
|
|
# first partial line of what we already have.
|
|
lines[0] = chunk_lines[-1] + lines[0]
|
|
lines = chunk_lines[:-1] + lines
|
|
else:
|
|
lines = chunk_lines
|
|
chunk_size = min(chunk_size * 2, 65536)
|
|
|
|
# Decode and return last N non-empty lines.
|
|
decoded = []
|
|
for raw in lines:
|
|
if not raw.strip():
|
|
continue
|
|
try:
|
|
decoded.append(raw.decode("utf-8", errors="replace") + "\n")
|
|
except Exception:
|
|
decoded.append(raw.decode("latin-1") + "\n")
|
|
return decoded[-n:]
|
|
|
|
except Exception:
|
|
# Fallback: read entire file
|
|
with open(path, "r", encoding="utf-8", errors="replace") as f:
|
|
all_lines = f.readlines()
|
|
return all_lines[-n:]
|
|
|
|
|
|
def _follow_log(
|
|
path: Path,
|
|
*,
|
|
min_level: Optional[str] = None,
|
|
session_filter: Optional[str] = None,
|
|
since: Optional[datetime] = None,
|
|
) -> None:
|
|
"""Poll a log file for new content and print matching lines."""
|
|
with open(path, "r", encoding="utf-8", errors="replace") as f:
|
|
# Seek to end
|
|
f.seek(0, 2)
|
|
while True:
|
|
line = f.readline()
|
|
if line:
|
|
if _matches_filters(line, min_level=min_level,
|
|
session_filter=session_filter, since=since):
|
|
print(line, end="")
|
|
sys.stdout.flush()
|
|
else:
|
|
time.sleep(0.3)
|
|
|
|
|
|
def list_logs() -> None:
|
|
"""Print available log files with sizes."""
|
|
log_dir = get_hermes_home() / "logs"
|
|
if not log_dir.exists():
|
|
print(f"No logs directory at {display_hermes_home()}/logs/")
|
|
return
|
|
|
|
print(f"Log files in {display_hermes_home()}/logs/:\n")
|
|
found = False
|
|
for entry in sorted(log_dir.iterdir()):
|
|
if entry.is_file() and entry.suffix == ".log":
|
|
size = entry.stat().st_size
|
|
mtime = datetime.fromtimestamp(entry.stat().st_mtime)
|
|
if size < 1024:
|
|
size_str = f"{size}B"
|
|
elif size < 1024 * 1024:
|
|
size_str = f"{size / 1024:.1f}KB"
|
|
else:
|
|
size_str = f"{size / (1024 * 1024):.1f}MB"
|
|
age = datetime.now() - mtime
|
|
if age.total_seconds() < 60:
|
|
age_str = "just now"
|
|
elif age.total_seconds() < 3600:
|
|
age_str = f"{int(age.total_seconds() / 60)}m ago"
|
|
elif age.total_seconds() < 86400:
|
|
age_str = f"{int(age.total_seconds() / 3600)}h ago"
|
|
else:
|
|
age_str = mtime.strftime("%Y-%m-%d")
|
|
print(f" {entry.name:<25} {size_str:>8} {age_str}")
|
|
found = True
|
|
|
|
if not found:
|
|
print(" (no log files yet — run 'hermes chat' to generate logs)")
|