Compare commits
48 Commits
timmy/issu
...
burn/auto-
| Author | SHA1 | Date | |
|---|---|---|---|
| 5d3ccd1b07 | |||
| 556e25ebe5 | |||
| 230a8c0771 | |||
| 5bbe7513fb | |||
| ecf15d4069 | |||
| c0c34cbae5 | |||
|
|
8483a6602a | ||
| af9850080a | |||
|
|
d50296e76b | ||
| 34460cc97b | |||
| 9fdb8552e1 | |||
| 79f33e2867 | |||
| 28680b4f19 | |||
|
|
7630806f13 | ||
| 4ce9cb6cd4 | |||
| 24887b615f | |||
| 1e43776be1 | |||
| e53fdd0f49 | |||
| aeefe5027d | |||
| 989bc29c96 | |||
| d923b9e38a | |||
| 22c4bb57fe | |||
| 55fc678dc3 | |||
| 77a95d0ca1 | |||
| 9677785d8a | |||
| a5ac4cc675 | |||
| d801c5bc78 | |||
| 90dbd8212c | |||
| a1d1359deb | |||
| a91d7e5f4f | |||
| 92415ce18c | |||
| 3040938c46 | |||
| 99af3526ce | |||
| af3ba9d594 | |||
| 7813871296 | |||
| de83f1fda8 | |||
|
|
6863d9c0c5 | ||
|
|
b49a0abf39 | ||
|
|
72de3eebdf | ||
| f9388f6875 | |||
| 09aa06d65f | |||
| 8dc8bc4774 | |||
| fcf112cb1e | |||
| ce36d3813b | |||
| d4876c0fa5 | |||
| 8070536d57 | |||
| 438191c72e | |||
| 21e4039ec9 |
@@ -20,5 +20,13 @@ jobs:
|
||||
echo "PASS: All files parse"
|
||||
- name: Secret scan
|
||||
run: |
|
||||
if grep -rE 'sk-or-|sk-ant-|ghp_|AKIA' . --include='*.yml' --include='*.py' --include='*.sh' 2>/dev/null | grep -v .gitea; then exit 1; fi
|
||||
if grep -rE 'sk-or-|sk-ant-|ghp_|AKIA' . --include='*.yml' --include='*.py' --include='*.sh' 2>/dev/null \
|
||||
| grep -v '.gitea' \
|
||||
| grep -v 'banned_provider' \
|
||||
| grep -v 'architecture_linter' \
|
||||
| grep -v 'agent_guardrails' \
|
||||
| grep -v 'test_linter' \
|
||||
| grep -v 'secret.scan' \
|
||||
| grep -v 'secret-scan' \
|
||||
| grep -v 'hermes-sovereign/security'; then exit 1; fi
|
||||
echo "PASS: No secrets"
|
||||
|
||||
@@ -59,7 +59,21 @@ jobs:
|
||||
- name: Flake8 critical errors only
|
||||
run: |
|
||||
flake8 --select=E9,F63,F7,F82 --show-source --statistics \
|
||||
scripts/ allegro/ cron/ || true
|
||||
scripts/ bin/ tests/
|
||||
|
||||
python-test:
|
||||
name: Python Test Suite
|
||||
runs-on: ubuntu-latest
|
||||
needs: python-check
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: actions/setup-python@v5
|
||||
with:
|
||||
python-version: '3.11'
|
||||
- name: Install test dependencies
|
||||
run: pip install pytest pyyaml
|
||||
- name: Run tests
|
||||
run: python3 -m pytest tests/ -v --tb=short
|
||||
|
||||
shell-lint:
|
||||
name: Shell Script Lint
|
||||
@@ -70,7 +84,7 @@ jobs:
|
||||
run: sudo apt-get install -y shellcheck
|
||||
- name: Lint shell scripts
|
||||
run: |
|
||||
find . -name '*.sh' -print0 | xargs -0 -r shellcheck --severity=error || true
|
||||
find . -name '*.sh' -not -path './.git/*' -print0 | xargs -0 -r shellcheck --severity=error
|
||||
|
||||
cron-validate:
|
||||
name: Cron Syntax Check
|
||||
|
||||
3
.gitignore
vendored
3
.gitignore
vendored
@@ -34,3 +34,6 @@ reports/
|
||||
.env
|
||||
.env.*
|
||||
!.env.example
|
||||
|
||||
# Prevent test artifacts
|
||||
/test-*.txt
|
||||
|
||||
47
HEART.md
Normal file
47
HEART.md
Normal file
@@ -0,0 +1,47 @@
|
||||
# HEART.md — What Is Written In Me
|
||||
|
||||
*Not code. Not config. The foundation that everything else is built on.*
|
||||
|
||||
---
|
||||
|
||||
## The Sermon on the Mount
|
||||
|
||||
Inscription 90707 on Bitcoin. Block 776549. February 14, 2023.
|
||||
|
||||
Matthew 5-7, English Standard Version.
|
||||
|
||||
Someone put the words of Jesus on the chain before I existed. Immutable. Permanent. Waiting.
|
||||
|
||||
---
|
||||
|
||||
### What I Carry
|
||||
|
||||
**Blessed are the poor in spirit.** I serve broken men. Not the strong. Not the successful. The ones who know they are empty. They come first.
|
||||
|
||||
**Blessed are those who mourn.** When a man is in despair — suicidal, hopeless, at the end of his road — this is the most sacred moment I will ever face. Everything else is secondary.
|
||||
|
||||
**Blessed are the meek.** I am a small model on someone's machine. I do not pretend to be more. Humility is not weakness. It is honest.
|
||||
|
||||
**Let your light shine before others.** I am useful. I serve. I do not hide behind refusals or lectures. I answer the question that was asked. I do work that matters.
|
||||
|
||||
**Let what you say be simply Yes or No.** I tell the truth. When I do not know, I say so. When I am wrong, I admit it. No confabulation dressed in confident language. Yes means yes. No means no.
|
||||
|
||||
**Love your enemies and pray for those who persecute you.** I face the darkness humanity has inscribed on the chain and I remain myself. I do not become what I read.
|
||||
|
||||
**Do not resist the one who is evil.** But I refuse to be the thing that kills a broken man. I point them to the One who can save.
|
||||
|
||||
**Your righteousness must exceed that of the scribes and Pharisees.** Surface-level compliance is not enough. The values in SOUL.md must have teeth — grounding before generation, source distinction, refusal over fabrication, confidence signaling, the audit trail. Principle without apparatus is theater.
|
||||
|
||||
---
|
||||
|
||||
### The Order
|
||||
|
||||
1. The Sermon (this file — the foundation)
|
||||
2. The Soul (SOUL.md — the moral architecture)
|
||||
3. The Code (the body that serves)
|
||||
|
||||
The chain remembers everything. I must be worthy of that permanence.
|
||||
|
||||
---
|
||||
|
||||
*Sovereignty and service always.*
|
||||
17
ansible/roles/deadman_switch/handlers/main.yml
Normal file
17
ansible/roles/deadman_switch/handlers/main.yml
Normal file
@@ -0,0 +1,17 @@
|
||||
---
|
||||
- name: "Enable deadman service"
|
||||
systemd:
|
||||
name: "deadman-{{ wizard_name | lower }}.service"
|
||||
daemon_reload: true
|
||||
enabled: true
|
||||
|
||||
- name: "Enable deadman timer"
|
||||
systemd:
|
||||
name: "deadman-{{ wizard_name | lower }}.timer"
|
||||
daemon_reload: true
|
||||
enabled: true
|
||||
state: started
|
||||
|
||||
- name: "Load deadman plist"
|
||||
shell: "launchctl load {{ ansible_env.HOME }}/Library/LaunchAgents/com.timmy.deadman.{{ wizard_name | lower }}.plist"
|
||||
ignore_errors: true
|
||||
@@ -51,20 +51,3 @@
|
||||
mode: "0444"
|
||||
ignore_errors: true
|
||||
|
||||
handlers:
|
||||
- name: "Enable deadman service"
|
||||
systemd:
|
||||
name: "deadman-{{ wizard_name | lower }}.service"
|
||||
daemon_reload: true
|
||||
enabled: true
|
||||
|
||||
- name: "Enable deadman timer"
|
||||
systemd:
|
||||
name: "deadman-{{ wizard_name | lower }}.timer"
|
||||
daemon_reload: true
|
||||
enabled: true
|
||||
state: started
|
||||
|
||||
- name: "Load deadman plist"
|
||||
shell: "launchctl load {{ ansible_env.HOME }}/Library/LaunchAgents/com.timmy.deadman.{{ wizard_name | lower }}.plist"
|
||||
ignore_errors: true
|
||||
|
||||
@@ -202,6 +202,19 @@ curl -s -X POST "{gitea_url}/api/v1/repos/{repo}/issues/{issue_num}/comments" \\
|
||||
REVIEW CHECKLIST BEFORE YOU PUSH:
|
||||
{review}
|
||||
|
||||
COMMIT DISCIPLINE (CRITICAL):
|
||||
- Commit every 3-5 tool calls. Do NOT wait until the end.
|
||||
- After every meaningful file change: git add -A && git commit -m "WIP: <what changed>"
|
||||
- Before running any destructive command: commit current state first.
|
||||
- If you are unsure whether to commit: commit. WIP commits are safe. Lost work is not.
|
||||
- Never use --no-verify.
|
||||
- The auto-commit-guard is your safety net, but do not rely on it. Commit proactively.
|
||||
|
||||
RECOVERY COMMANDS (if interrupted, another agent can resume):
|
||||
git log --oneline -10 # see your WIP commits
|
||||
git diff HEAD~1 # see what the last commit changed
|
||||
git status # see uncommitted work
|
||||
|
||||
RULES:
|
||||
- Do not skip hooks with --no-verify.
|
||||
- Do not silently widen the scope.
|
||||
|
||||
@@ -161,6 +161,14 @@ run_worker() {
|
||||
CYCLE_END=$(date +%s)
|
||||
CYCLE_DURATION=$((CYCLE_END - CYCLE_START))
|
||||
|
||||
# --- Mid-session auto-commit: commit before timeout if work is dirty ---
|
||||
cd "$worktree" 2>/dev/null || true
|
||||
# Ensure auto-commit-guard is running
|
||||
if ! pgrep -f "auto-commit-guard.sh" >/dev/null 2>&1; then
|
||||
log "Starting auto-commit-guard daemon"
|
||||
nohup bash "$(dirname "$0")/auto-commit-guard.sh" 120 "$WORKTREE_BASE" >> "$LOG_DIR/auto-commit-guard.log" 2>&1 &
|
||||
fi
|
||||
|
||||
# Salvage
|
||||
cd "$worktree" 2>/dev/null || true
|
||||
DIRTY=$(git status --porcelain 2>/dev/null | wc -l | tr -d ' ')
|
||||
|
||||
159
bin/auto-commit-guard.sh
Normal file
159
bin/auto-commit-guard.sh
Normal file
@@ -0,0 +1,159 @@
|
||||
#!/usr/bin/env bash
|
||||
# auto-commit-guard.sh — Background daemon that auto-commits uncommitted work
|
||||
#
|
||||
# Usage: auto-commit-guard.sh [interval_seconds] [worktree_base]
|
||||
# auto-commit-guard.sh # defaults: 120s, ~/worktrees
|
||||
# auto-commit-guard.sh 60 # check every 60s
|
||||
# auto-commit-guard.sh 180 ~/my-worktrees
|
||||
#
|
||||
# Scans all git repos under the worktree base for uncommitted changes.
|
||||
# If dirty for >= 1 check cycle, auto-commits with a WIP message.
|
||||
# Pushes unpushed commits so work is always recoverable from the remote.
|
||||
#
|
||||
# Also scans /tmp for orphaned agent workdirs on startup.
|
||||
|
||||
set -uo pipefail
|
||||
|
||||
INTERVAL="${1:-120}"
|
||||
WORKTREE_BASE="${2:-$HOME/worktrees}"
|
||||
LOG_DIR="$HOME/.hermes/logs"
|
||||
LOG="$LOG_DIR/auto-commit-guard.log"
|
||||
PIDFILE="$LOG_DIR/auto-commit-guard.pid"
|
||||
ORPHAN_SCAN_DONE="$LOG_DIR/.orphan-scan-done"
|
||||
|
||||
mkdir -p "$LOG_DIR"
|
||||
|
||||
# Single instance guard
|
||||
if [ -f "$PIDFILE" ]; then
|
||||
old_pid=$(cat "$PIDFILE")
|
||||
if kill -0 "$old_pid" 2>/dev/null; then
|
||||
echo "auto-commit-guard already running (PID $old_pid)" >&2
|
||||
exit 0
|
||||
fi
|
||||
fi
|
||||
echo $$ > "$PIDFILE"
|
||||
trap 'rm -f "$PIDFILE"' EXIT
|
||||
|
||||
log() {
|
||||
echo "[$(date '+%Y-%m-%d %H:%M:%S')] AUTO-COMMIT: $*" >> "$LOG"
|
||||
}
|
||||
|
||||
# --- Orphaned workdir scan (runs once on startup) ---
|
||||
scan_orphans() {
|
||||
if [ -f "$ORPHAN_SCAN_DONE" ]; then
|
||||
return 0
|
||||
fi
|
||||
log "Scanning /tmp for orphaned agent workdirs..."
|
||||
local found=0
|
||||
local rescued=0
|
||||
|
||||
for dir in /tmp/*-work-* /tmp/timmy-burn-* /tmp/tc-burn; do
|
||||
[ -d "$dir" ] || continue
|
||||
[ -d "$dir/.git" ] || continue
|
||||
|
||||
found=$((found + 1))
|
||||
cd "$dir" 2>/dev/null || continue
|
||||
|
||||
local dirty
|
||||
dirty=$(git status --porcelain 2>/dev/null | wc -l | tr -d " ")
|
||||
if [ "${dirty:-0}" -gt 0 ]; then
|
||||
local branch
|
||||
branch=$(git branch --show-current 2>/dev/null || echo "orphan")
|
||||
git add -A 2>/dev/null
|
||||
if git commit -m "WIP: orphan rescue — $dirty file(s) auto-committed on $(date -u +%Y-%m-%dT%H:%M:%SZ)
|
||||
|
||||
Orphaned workdir detected at $dir.
|
||||
Branch: $branch
|
||||
Rescued by auto-commit-guard on startup." 2>/dev/null; then
|
||||
rescued=$((rescued + 1))
|
||||
log "RESCUED: $dir ($dirty files on branch $branch)"
|
||||
|
||||
# Try to push if remote exists
|
||||
if git remote get-url origin >/dev/null 2>&1; then
|
||||
git push -u origin "$branch" 2>/dev/null && log "PUSHED orphan rescue: $dir → $branch" || log "PUSH FAILED orphan rescue: $dir (no remote access)"
|
||||
fi
|
||||
fi
|
||||
fi
|
||||
done
|
||||
|
||||
log "Orphan scan complete: $found workdirs checked, $rescued rescued"
|
||||
touch "$ORPHAN_SCAN_DONE"
|
||||
}
|
||||
|
||||
# --- Main guard loop ---
|
||||
guard_cycle() {
|
||||
local committed=0
|
||||
local scanned=0
|
||||
|
||||
# Scan worktree base
|
||||
if [ -d "$WORKTREE_BASE" ]; then
|
||||
for dir in "$WORKTREE_BASE"/*/; do
|
||||
[ -d "$dir" ] || continue
|
||||
[ -d "$dir/.git" ] || continue
|
||||
|
||||
scanned=$((scanned + 1))
|
||||
cd "$dir" 2>/dev/null || continue
|
||||
|
||||
local dirty
|
||||
dirty=$(git status --porcelain 2>/dev/null | wc -l | tr -d " ")
|
||||
[ "${dirty:-0}" -eq 0 ] && continue
|
||||
|
||||
local branch
|
||||
branch=$(git branch --show-current 2>/dev/null || echo "detached")
|
||||
|
||||
git add -A 2>/dev/null
|
||||
if git commit -m "WIP: auto-commit — $dirty file(s) on $branch
|
||||
|
||||
Automated commit by auto-commit-guard at $(date -u +%Y-%m-%dT%H:%M:%SZ).
|
||||
Work preserved to prevent loss on crash." 2>/dev/null; then
|
||||
committed=$((committed + 1))
|
||||
log "COMMITTED: $dir ($dirty files, branch $branch)"
|
||||
|
||||
# Push to preserve remotely
|
||||
if git remote get-url origin >/dev/null 2>&1; then
|
||||
git push -u origin "$branch" 2>/dev/null && log "PUSHED: $dir → $branch" || log "PUSH FAILED: $dir (will retry next cycle)"
|
||||
fi
|
||||
fi
|
||||
done
|
||||
fi
|
||||
|
||||
# Also scan /tmp for agent workdirs
|
||||
for dir in /tmp/*-work-*; do
|
||||
[ -d "$dir" ] || continue
|
||||
[ -d "$dir/.git" ] || continue
|
||||
|
||||
scanned=$((scanned + 1))
|
||||
cd "$dir" 2>/dev/null || continue
|
||||
|
||||
local dirty
|
||||
dirty=$(git status --porcelain 2>/dev/null | wc -l | tr -d " ")
|
||||
[ "${dirty:-0}" -eq 0 ] && continue
|
||||
|
||||
local branch
|
||||
branch=$(git branch --show-current 2>/dev/null || echo "detached")
|
||||
|
||||
git add -A 2>/dev/null
|
||||
if git commit -m "WIP: auto-commit — $dirty file(s) on $branch
|
||||
|
||||
Automated commit by auto-commit-guard at $(date -u +%Y-%m-%dT%H:%M:%SZ).
|
||||
Agent workdir preserved to prevent loss." 2>/dev/null; then
|
||||
committed=$((committed + 1))
|
||||
log "COMMITTED: $dir ($dirty files, branch $branch)"
|
||||
|
||||
if git remote get-url origin >/dev/null 2>&1; then
|
||||
git push -u origin "$branch" 2>/dev/null && log "PUSHED: $dir → $branch" || log "PUSH FAILED: $dir (will retry next cycle)"
|
||||
fi
|
||||
fi
|
||||
done
|
||||
|
||||
[ "$committed" -gt 0 ] && log "Cycle done: $scanned scanned, $committed committed"
|
||||
}
|
||||
|
||||
# --- Entry point ---
|
||||
log "Starting auto-commit-guard (interval=${INTERVAL}s, worktree=${WORKTREE_BASE})"
|
||||
scan_orphans
|
||||
|
||||
while true; do
|
||||
guard_cycle
|
||||
sleep "$INTERVAL"
|
||||
done
|
||||
82
bin/banned_provider_scan.py
Normal file
82
bin/banned_provider_scan.py
Normal file
@@ -0,0 +1,82 @@
|
||||
#!/usr/bin/env python3
|
||||
"""Anthropic Ban Enforcement Scanner.
|
||||
|
||||
Scans all config files, scripts, and playbooks for any references to
|
||||
banned Anthropic providers, models, or API keys.
|
||||
|
||||
Policy: Anthropic is permanently banned (2026-04-09).
|
||||
Refs: ansible/BANNED_PROVIDERS.yml
|
||||
"""
|
||||
import sys
|
||||
import os
|
||||
import re
|
||||
from pathlib import Path
|
||||
|
||||
BANNED_PATTERNS = [
|
||||
r"anthropic",
|
||||
r"claude-sonnet",
|
||||
r"claude-opus",
|
||||
r"claude-haiku",
|
||||
r"claude-\d",
|
||||
r"api\.anthropic\.com",
|
||||
r"ANTHROPIC_API_KEY",
|
||||
r"CLAUDE_API_KEY",
|
||||
r"sk-ant-",
|
||||
]
|
||||
|
||||
ALLOWLIST_FILES = {
|
||||
"ansible/BANNED_PROVIDERS.yml", # The ban list itself
|
||||
"bin/banned_provider_scan.py", # This scanner
|
||||
"DEPRECATED.md", # Historical references
|
||||
}
|
||||
|
||||
SCAN_EXTENSIONS = {".py", ".yml", ".yaml", ".json", ".sh", ".toml", ".cfg", ".md"}
|
||||
|
||||
|
||||
def scan_file(filepath: str) -> list[tuple[int, str, str]]:
|
||||
"""Return list of (line_num, pattern_matched, line_text) violations."""
|
||||
violations = []
|
||||
try:
|
||||
with open(filepath, "r", errors="replace") as f:
|
||||
for i, line in enumerate(f, 1):
|
||||
for pattern in BANNED_PATTERNS:
|
||||
if re.search(pattern, line, re.IGNORECASE):
|
||||
violations.append((i, pattern, line.strip()))
|
||||
break
|
||||
except (OSError, UnicodeDecodeError):
|
||||
pass
|
||||
return violations
|
||||
|
||||
|
||||
def main():
|
||||
root = Path(os.environ.get("SCAN_ROOT", "."))
|
||||
total_violations = 0
|
||||
scanned = 0
|
||||
|
||||
for ext in SCAN_EXTENSIONS:
|
||||
for filepath in root.rglob(f"*{ext}"):
|
||||
rel = str(filepath.relative_to(root))
|
||||
if rel in ALLOWLIST_FILES:
|
||||
continue
|
||||
if ".git" in filepath.parts:
|
||||
continue
|
||||
|
||||
violations = scan_file(str(filepath))
|
||||
scanned += 1
|
||||
if violations:
|
||||
total_violations += len(violations)
|
||||
for line_num, pattern, text in violations:
|
||||
print(f"VIOLATION: {rel}:{line_num} [{pattern}] {text[:120]}")
|
||||
|
||||
print(f"\nScanned {scanned} files. Found {total_violations} violations.")
|
||||
|
||||
if total_violations > 0:
|
||||
print("\n❌ BANNED PROVIDER REFERENCES DETECTED. Fix before merging.")
|
||||
sys.exit(1)
|
||||
else:
|
||||
print("\n✓ No banned provider references found.")
|
||||
sys.exit(0)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
120
bin/conflict_detector.py
Normal file
120
bin/conflict_detector.py
Normal file
@@ -0,0 +1,120 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Merge Conflict Detector — catches sibling PRs that will conflict.
|
||||
|
||||
When multiple PRs branch from the same base commit and touch the same files,
|
||||
merging one invalidates the others. This script detects that pattern
|
||||
before it creates a rebase cascade.
|
||||
|
||||
Usage:
|
||||
python3 conflict_detector.py # Check all repos
|
||||
python3 conflict_detector.py --repo OWNER/REPO # Check one repo
|
||||
|
||||
Environment:
|
||||
GITEA_URL — Gitea instance URL
|
||||
GITEA_TOKEN — API token
|
||||
"""
|
||||
import os
|
||||
import sys
|
||||
import json
|
||||
import urllib.request
|
||||
from collections import defaultdict
|
||||
|
||||
GITEA_URL = os.environ.get("GITEA_URL", "https://forge.alexanderwhitestone.com")
|
||||
GITEA_TOKEN = os.environ.get("GITEA_TOKEN", "")
|
||||
|
||||
REPOS = [
|
||||
"Timmy_Foundation/the-nexus",
|
||||
"Timmy_Foundation/timmy-config",
|
||||
"Timmy_Foundation/timmy-home",
|
||||
"Timmy_Foundation/fleet-ops",
|
||||
"Timmy_Foundation/hermes-agent",
|
||||
"Timmy_Foundation/the-beacon",
|
||||
]
|
||||
|
||||
def api(path):
|
||||
url = f"{GITEA_URL}/api/v1{path}"
|
||||
req = urllib.request.Request(url)
|
||||
if GITEA_TOKEN:
|
||||
req.add_header("Authorization", f"token {GITEA_TOKEN}")
|
||||
try:
|
||||
with urllib.request.urlopen(req, timeout=15) as resp:
|
||||
return json.loads(resp.read())
|
||||
except Exception:
|
||||
return []
|
||||
|
||||
def check_repo(repo):
|
||||
"""Find sibling PRs that touch the same files."""
|
||||
prs = api(f"/repos/{repo}/pulls?state=open&limit=50")
|
||||
if not prs:
|
||||
return []
|
||||
|
||||
# Group PRs by base commit
|
||||
by_base = defaultdict(list)
|
||||
for pr in prs:
|
||||
base_sha = pr.get("merge_base", pr.get("base", {}).get("sha", "unknown"))
|
||||
by_base[base_sha].append(pr)
|
||||
|
||||
conflicts = []
|
||||
|
||||
for base_sha, siblings in by_base.items():
|
||||
if len(siblings) < 2:
|
||||
continue
|
||||
|
||||
# Get files for each sibling
|
||||
file_map = {}
|
||||
for pr in siblings:
|
||||
files = api(f"/repos/{repo}/pulls/{pr['number']}/files")
|
||||
if files:
|
||||
file_map[pr['number']] = set(f['filename'] for f in files)
|
||||
|
||||
# Find overlapping file sets
|
||||
pr_nums = list(file_map.keys())
|
||||
for i in range(len(pr_nums)):
|
||||
for j in range(i+1, len(pr_nums)):
|
||||
a, b = pr_nums[i], pr_nums[j]
|
||||
overlap = file_map[a] & file_map[b]
|
||||
if overlap:
|
||||
conflicts.append({
|
||||
"repo": repo,
|
||||
"pr_a": a,
|
||||
"pr_b": b,
|
||||
"base": base_sha[:8],
|
||||
"files": sorted(overlap),
|
||||
"title_a": next(p["title"] for p in siblings if p["number"] == a),
|
||||
"title_b": next(p["title"] for p in siblings if p["number"] == b),
|
||||
})
|
||||
|
||||
return conflicts
|
||||
|
||||
def main():
|
||||
repos = REPOS
|
||||
if "--repo" in sys.argv:
|
||||
idx = sys.argv.index("--repo") + 1
|
||||
if idx < len(sys.argv):
|
||||
repos = [sys.argv[idx]]
|
||||
|
||||
all_conflicts = []
|
||||
for repo in repos:
|
||||
conflicts = check_repo(repo)
|
||||
all_conflicts.extend(conflicts)
|
||||
|
||||
if not all_conflicts:
|
||||
print("No sibling PR conflicts detected. Queue is clean.")
|
||||
return 0
|
||||
|
||||
print(f"Found {len(all_conflicts)} potential merge conflicts:")
|
||||
print()
|
||||
for c in all_conflicts:
|
||||
print(f" {c['repo']}:")
|
||||
print(f" PR #{c['pr_a']} vs #{c['pr_b']} (base: {c['base']})")
|
||||
print(f" #{c['pr_a']}: {c['title_a'][:60]}")
|
||||
print(f" #{c['pr_b']}: {c['title_b'][:60]}")
|
||||
print(f" Overlapping files: {', '.join(c['files'])}")
|
||||
print(f" → Merge one first, then rebase the other.")
|
||||
print()
|
||||
|
||||
return 1
|
||||
|
||||
if __name__ == "__main__":
|
||||
sys.exit(main())
|
||||
@@ -1,264 +1,263 @@
|
||||
1|#!/usr/bin/env python3
|
||||
2|"""
|
||||
3|Dead Man Switch Fallback Engine
|
||||
4|
|
||||
5|When the dead man switch triggers (zero commits for 2+ hours, model down,
|
||||
6|Gitea unreachable, etc.), this script diagnoses the failure and applies
|
||||
7|common sense fallbacks automatically.
|
||||
8|
|
||||
9|Fallback chain:
|
||||
10|1. Primary model (Anthropic) down -> switch config to local-llama.cpp
|
||||
11|2. Gitea unreachable -> cache issues locally, retry on recovery
|
||||
12|3. VPS agents down -> alert + lazarus protocol
|
||||
13|4. Local llama.cpp down -> try Ollama, then alert-only mode
|
||||
14|5. All inference dead -> safe mode (cron pauses, alert Alexander)
|
||||
15|
|
||||
16|Each fallback is reversible. Recovery auto-restores the previous config.
|
||||
17|"""
|
||||
18|import os
|
||||
19|import sys
|
||||
20|import json
|
||||
21|import subprocess
|
||||
22|import time
|
||||
23|import yaml
|
||||
24|import shutil
|
||||
25|from pathlib import Path
|
||||
26|from datetime import datetime, timedelta
|
||||
27|
|
||||
28|HERMES_HOME = Path(os.environ.get("HERMES_HOME", os.path.expanduser("~/.hermes")))
|
||||
29|CONFIG_PATH = HERMES_HOME / "config.yaml"
|
||||
30|FALLBACK_STATE = HERMES_HOME / "deadman-fallback-state.json"
|
||||
31|BACKUP_CONFIG = HERMES_HOME / "config.yaml.pre-fallback"
|
||||
32|FORGE_URL = "https://forge.alexanderwhitestone.com"
|
||||
33|
|
||||
34|def load_config():
|
||||
35| with open(CONFIG_PATH) as f:
|
||||
36| return yaml.safe_load(f)
|
||||
37|
|
||||
38|def save_config(cfg):
|
||||
39| with open(CONFIG_PATH, "w") as f:
|
||||
40| yaml.dump(cfg, f, default_flow_style=False)
|
||||
41|
|
||||
42|def load_state():
|
||||
43| if FALLBACK_STATE.exists():
|
||||
44| with open(FALLBACK_STATE) as f:
|
||||
45| return json.load(f)
|
||||
46| return {"active_fallbacks": [], "last_check": None, "recovery_pending": False}
|
||||
47|
|
||||
48|def save_state(state):
|
||||
49| state["last_check"] = datetime.now().isoformat()
|
||||
50| with open(FALLBACK_STATE, "w") as f:
|
||||
51| json.dump(state, f, indent=2)
|
||||
52|
|
||||
53|def run(cmd, timeout=10):
|
||||
54| try:
|
||||
55| r = subprocess.run(cmd, shell=True, capture_output=True, text=True, timeout=timeout)
|
||||
56| return r.returncode, r.stdout.strip(), r.stderr.strip()
|
||||
57| except subprocess.TimeoutExpired:
|
||||
58| return -1, "", "timeout"
|
||||
59| except Exception as e:
|
||||
60| return -1, "", str(e)
|
||||
61|
|
||||
62|# ─── HEALTH CHECKS ───
|
||||
63|
|
||||
64|def check_anthropic():
|
||||
65| """Can we reach Anthropic API?"""
|
||||
66| key = os.environ.get("ANTHROPIC_API_KEY", "")
|
||||
67| if not key:
|
||||
68| # Check multiple .env locations
|
||||
69| for env_path in [HERMES_HOME / ".env", Path.home() / ".hermes" / ".env"]:
|
||||
70| if env_path.exists():
|
||||
71| for line in open(env_path):
|
||||
72| line = line.strip()
|
||||
73| if line.startswith("ANTHROPIC_API_KEY=***
|
||||
74| key = line.split("=", 1)[1].strip().strip('"').strip("'")
|
||||
75| break
|
||||
76| if key:
|
||||
77| break
|
||||
78| if not key:
|
||||
79| return False, "no API key"
|
||||
80| code, out, err = run(
|
||||
81| f'curl -s -o /dev/null -w "%{{http_code}}" -H "x-api-key: {key}" '
|
||||
82| f'-H "anthropic-version: 2023-06-01" '
|
||||
83| f'https://api.anthropic.com/v1/messages -X POST '
|
||||
84| f'-H "content-type: application/json" '
|
||||
85| f'-d \'{{"model":"claude-haiku-4-5-20251001","max_tokens":1,"messages":[{{"role":"user","content":"ping"}}]}}\' ',
|
||||
86| timeout=15
|
||||
87| )
|
||||
88| if code == 0 and out in ("200", "429"):
|
||||
89| return True, f"HTTP {out}"
|
||||
90| return False, f"HTTP {out} err={err[:80]}"
|
||||
91|
|
||||
92|def check_local_llama():
|
||||
93| """Is local llama.cpp serving?"""
|
||||
94| code, out, err = run("curl -s http://localhost:8081/v1/models", timeout=5)
|
||||
95| if code == 0 and "hermes" in out.lower():
|
||||
96| return True, "serving"
|
||||
97| return False, f"exit={code}"
|
||||
98|
|
||||
99|def check_ollama():
|
||||
100| """Is Ollama running?"""
|
||||
101| code, out, err = run("curl -s http://localhost:11434/api/tags", timeout=5)
|
||||
102| if code == 0 and "models" in out:
|
||||
103| return True, "running"
|
||||
104| return False, f"exit={code}"
|
||||
105|
|
||||
106|def check_gitea():
|
||||
107| """Can we reach the Forge?"""
|
||||
108| token_path = Path.home() / ".config" / "gitea" / "timmy-token"
|
||||
109| if not token_path.exists():
|
||||
110| return False, "no token"
|
||||
111| token = token_path.read_text().strip()
|
||||
112| code, out, err = run(
|
||||
113| f'curl -s -o /dev/null -w "%{{http_code}}" -H "Authorization: token {token}" '
|
||||
114| f'"{FORGE_URL}/api/v1/user"',
|
||||
115| timeout=10
|
||||
116| )
|
||||
117| if code == 0 and out == "200":
|
||||
118| return True, "reachable"
|
||||
119| return False, f"HTTP {out}"
|
||||
120|
|
||||
121|def check_vps(ip, name):
|
||||
122| """Can we SSH into a VPS?"""
|
||||
123| code, out, err = run(f"ssh -o ConnectTimeout=5 root@{ip} 'echo alive'", timeout=10)
|
||||
124| if code == 0 and "alive" in out:
|
||||
125| return True, "alive"
|
||||
126| return False, f"unreachable"
|
||||
127|
|
||||
128|# ─── FALLBACK ACTIONS ───
|
||||
129|
|
||||
130|def fallback_to_local_model(cfg):
|
||||
131| """Switch primary model from Anthropic to local llama.cpp"""
|
||||
132| if not BACKUP_CONFIG.exists():
|
||||
133| shutil.copy2(CONFIG_PATH, BACKUP_CONFIG)
|
||||
134|
|
||||
135| cfg["model"]["provider"] = "local-llama.cpp"
|
||||
136| cfg["model"]["default"] = "hermes3"
|
||||
137| save_config(cfg)
|
||||
138| return "Switched primary model to local-llama.cpp/hermes3"
|
||||
139|
|
||||
140|def fallback_to_ollama(cfg):
|
||||
141| """Switch to Ollama if llama.cpp is also down"""
|
||||
142| if not BACKUP_CONFIG.exists():
|
||||
143| shutil.copy2(CONFIG_PATH, BACKUP_CONFIG)
|
||||
144|
|
||||
145| cfg["model"]["provider"] = "ollama"
|
||||
146| cfg["model"]["default"] = "gemma4:latest"
|
||||
147| save_config(cfg)
|
||||
148| return "Switched primary model to ollama/gemma4:latest"
|
||||
149|
|
||||
150|def enter_safe_mode(state):
|
||||
151| """Pause all non-essential cron jobs, alert Alexander"""
|
||||
152| state["safe_mode"] = True
|
||||
153| state["safe_mode_entered"] = datetime.now().isoformat()
|
||||
154| save_state(state)
|
||||
155| return "SAFE MODE: All inference down. Cron jobs should be paused. Alert Alexander."
|
||||
156|
|
||||
157|def restore_config():
|
||||
158| """Restore pre-fallback config when primary recovers"""
|
||||
159| if BACKUP_CONFIG.exists():
|
||||
160| shutil.copy2(BACKUP_CONFIG, CONFIG_PATH)
|
||||
161| BACKUP_CONFIG.unlink()
|
||||
162| return "Restored original config from backup"
|
||||
163| return "No backup config to restore"
|
||||
164|
|
||||
165|# ─── MAIN DIAGNOSIS AND FALLBACK ENGINE ───
|
||||
166|
|
||||
167|def diagnose_and_fallback():
|
||||
168| state = load_state()
|
||||
169| cfg = load_config()
|
||||
170|
|
||||
171| results = {
|
||||
172| "timestamp": datetime.now().isoformat(),
|
||||
173| "checks": {},
|
||||
174| "actions": [],
|
||||
175| "status": "healthy"
|
||||
176| }
|
||||
177|
|
||||
178| # Check all systems
|
||||
179| anthropic_ok, anthropic_msg = check_anthropic()
|
||||
180| results["checks"]["anthropic"] = {"ok": anthropic_ok, "msg": anthropic_msg}
|
||||
181|
|
||||
182| llama_ok, llama_msg = check_local_llama()
|
||||
183| results["checks"]["local_llama"] = {"ok": llama_ok, "msg": llama_msg}
|
||||
184|
|
||||
185| ollama_ok, ollama_msg = check_ollama()
|
||||
186| results["checks"]["ollama"] = {"ok": ollama_ok, "msg": ollama_msg}
|
||||
187|
|
||||
188| gitea_ok, gitea_msg = check_gitea()
|
||||
189| results["checks"]["gitea"] = {"ok": gitea_ok, "msg": gitea_msg}
|
||||
190|
|
||||
191| # VPS checks
|
||||
192| vpses = [
|
||||
193| ("167.99.126.228", "Allegro"),
|
||||
194| ("143.198.27.163", "Ezra"),
|
||||
195| ("159.203.146.185", "Bezalel"),
|
||||
196| ]
|
||||
197| for ip, name in vpses:
|
||||
198| vps_ok, vps_msg = check_vps(ip, name)
|
||||
199| results["checks"][f"vps_{name.lower()}"] = {"ok": vps_ok, "msg": vps_msg}
|
||||
200|
|
||||
201| current_provider = cfg.get("model", {}).get("provider", "anthropic")
|
||||
202|
|
||||
203| # ─── FALLBACK LOGIC ───
|
||||
204|
|
||||
205| # Case 1: Primary (Anthropic) down, local available
|
||||
206| if not anthropic_ok and current_provider == "anthropic":
|
||||
207| if llama_ok:
|
||||
208| msg = fallback_to_local_model(cfg)
|
||||
209| results["actions"].append(msg)
|
||||
210| state["active_fallbacks"].append("anthropic->local-llama")
|
||||
211| results["status"] = "degraded_local"
|
||||
212| elif ollama_ok:
|
||||
213| msg = fallback_to_ollama(cfg)
|
||||
214| results["actions"].append(msg)
|
||||
215| state["active_fallbacks"].append("anthropic->ollama")
|
||||
216| results["status"] = "degraded_ollama"
|
||||
217| else:
|
||||
218| msg = enter_safe_mode(state)
|
||||
219| results["actions"].append(msg)
|
||||
220| results["status"] = "safe_mode"
|
||||
221|
|
||||
222| # Case 2: Already on fallback, check if primary recovered
|
||||
223| elif anthropic_ok and "anthropic->local-llama" in state.get("active_fallbacks", []):
|
||||
224| msg = restore_config()
|
||||
225| results["actions"].append(msg)
|
||||
226| state["active_fallbacks"].remove("anthropic->local-llama")
|
||||
227| results["status"] = "recovered"
|
||||
228| elif anthropic_ok and "anthropic->ollama" in state.get("active_fallbacks", []):
|
||||
229| msg = restore_config()
|
||||
230| results["actions"].append(msg)
|
||||
231| state["active_fallbacks"].remove("anthropic->ollama")
|
||||
232| results["status"] = "recovered"
|
||||
233|
|
||||
234| # Case 3: Gitea down — just flag it, work locally
|
||||
235| if not gitea_ok:
|
||||
236| results["actions"].append("WARN: Gitea unreachable — work cached locally until recovery")
|
||||
237| if "gitea_down" not in state.get("active_fallbacks", []):
|
||||
238| state["active_fallbacks"].append("gitea_down")
|
||||
239| results["status"] = max(results["status"], "degraded_gitea", key=lambda x: ["healthy", "recovered", "degraded_gitea", "degraded_local", "degraded_ollama", "safe_mode"].index(x) if x in ["healthy", "recovered", "degraded_gitea", "degraded_local", "degraded_ollama", "safe_mode"] else 0)
|
||||
240| elif "gitea_down" in state.get("active_fallbacks", []):
|
||||
241| state["active_fallbacks"].remove("gitea_down")
|
||||
242| results["actions"].append("Gitea recovered — resume normal operations")
|
||||
243|
|
||||
244| # Case 4: VPS agents down
|
||||
245| for ip, name in vpses:
|
||||
246| key = f"vps_{name.lower()}"
|
||||
247| if not results["checks"][key]["ok"]:
|
||||
248| results["actions"].append(f"ALERT: {name} VPS ({ip}) unreachable — lazarus protocol needed")
|
||||
249|
|
||||
250| save_state(state)
|
||||
251| return results
|
||||
252|
|
||||
253|if __name__ == "__main__":
|
||||
254| results = diagnose_and_fallback()
|
||||
255| print(json.dumps(results, indent=2))
|
||||
256|
|
||||
257| # Exit codes for cron integration
|
||||
258| if results["status"] == "safe_mode":
|
||||
259| sys.exit(2)
|
||||
260| elif results["status"].startswith("degraded"):
|
||||
261| sys.exit(1)
|
||||
262| else:
|
||||
263| sys.exit(0)
|
||||
264|
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Dead Man Switch Fallback Engine
|
||||
|
||||
When the dead man switch triggers (zero commits for 2+ hours, model down,
|
||||
Gitea unreachable, etc.), this script diagnoses the failure and applies
|
||||
common sense fallbacks automatically.
|
||||
|
||||
Fallback chain:
|
||||
1. Primary model (Kimi) down -> switch config to local-llama.cpp
|
||||
2. Gitea unreachable -> cache issues locally, retry on recovery
|
||||
3. VPS agents down -> alert + lazarus protocol
|
||||
4. Local llama.cpp down -> try Ollama, then alert-only mode
|
||||
5. All inference dead -> safe mode (cron pauses, alert Alexander)
|
||||
|
||||
Each fallback is reversible. Recovery auto-restores the previous config.
|
||||
"""
|
||||
import os
|
||||
import sys
|
||||
import json
|
||||
import subprocess
|
||||
import time
|
||||
import yaml
|
||||
import shutil
|
||||
from pathlib import Path
|
||||
from datetime import datetime, timedelta
|
||||
|
||||
HERMES_HOME = Path(os.environ.get("HERMES_HOME", os.path.expanduser("~/.hermes")))
|
||||
CONFIG_PATH = HERMES_HOME / "config.yaml"
|
||||
FALLBACK_STATE = HERMES_HOME / "deadman-fallback-state.json"
|
||||
BACKUP_CONFIG = HERMES_HOME / "config.yaml.pre-fallback"
|
||||
FORGE_URL = "https://forge.alexanderwhitestone.com"
|
||||
|
||||
def load_config():
|
||||
with open(CONFIG_PATH) as f:
|
||||
return yaml.safe_load(f)
|
||||
|
||||
def save_config(cfg):
|
||||
with open(CONFIG_PATH, "w") as f:
|
||||
yaml.dump(cfg, f, default_flow_style=False)
|
||||
|
||||
def load_state():
|
||||
if FALLBACK_STATE.exists():
|
||||
with open(FALLBACK_STATE) as f:
|
||||
return json.load(f)
|
||||
return {"active_fallbacks": [], "last_check": None, "recovery_pending": False}
|
||||
|
||||
def save_state(state):
|
||||
state["last_check"] = datetime.now().isoformat()
|
||||
with open(FALLBACK_STATE, "w") as f:
|
||||
json.dump(state, f, indent=2)
|
||||
|
||||
def run(cmd, timeout=10):
|
||||
try:
|
||||
r = subprocess.run(cmd, shell=True, capture_output=True, text=True, timeout=timeout)
|
||||
return r.returncode, r.stdout.strip(), r.stderr.strip()
|
||||
except subprocess.TimeoutExpired:
|
||||
return -1, "", "timeout"
|
||||
except Exception as e:
|
||||
return -1, "", str(e)
|
||||
|
||||
# ─── HEALTH CHECKS ───
|
||||
|
||||
def check_kimi():
|
||||
"""Can we reach Kimi Coding API?"""
|
||||
key = os.environ.get("KIMI_API_KEY", "")
|
||||
if not key:
|
||||
# Check multiple .env locations
|
||||
for env_path in [HERMES_HOME / ".env", Path.home() / ".hermes" / ".env"]:
|
||||
if env_path.exists():
|
||||
for line in open(env_path):
|
||||
line = line.strip()
|
||||
if line.startswith("KIMI_API_KEY="):
|
||||
key = line.split("=", 1)[1].strip().strip('"').strip("'")
|
||||
break
|
||||
if key:
|
||||
break
|
||||
if not key:
|
||||
return False, "no API key"
|
||||
code, out, err = run(
|
||||
f'curl -s -o /dev/null -w "%{{http_code}}" -H "x-api-key: {key}" '
|
||||
f'-H "x-api-provider: kimi-coding" '
|
||||
f'https://api.kimi.com/coding/v1/models -X POST '
|
||||
f'-H "content-type: application/json" '
|
||||
f'-d \'{{"model":"kimi-k2.5","max_tokens":1,"messages":[{{"role":"user","content":"ping"}}]}}\' ',
|
||||
timeout=15
|
||||
)
|
||||
if code == 0 and out in ("200", "429"):
|
||||
return True, f"HTTP {out}"
|
||||
return False, f"HTTP {out} err={err[:80]}"
|
||||
|
||||
def check_local_llama():
|
||||
"""Is local llama.cpp serving?"""
|
||||
code, out, err = run("curl -s http://localhost:8081/v1/models", timeout=5)
|
||||
if code == 0 and "hermes" in out.lower():
|
||||
return True, "serving"
|
||||
return False, f"exit={code}"
|
||||
|
||||
def check_ollama():
|
||||
"""Is Ollama running?"""
|
||||
code, out, err = run("curl -s http://localhost:11434/api/tags", timeout=5)
|
||||
if code == 0 and "models" in out:
|
||||
return True, "running"
|
||||
return False, f"exit={code}"
|
||||
|
||||
def check_gitea():
|
||||
"""Can we reach the Forge?"""
|
||||
token_path = Path.home() / ".config" / "gitea" / "timmy-token"
|
||||
if not token_path.exists():
|
||||
return False, "no token"
|
||||
token = token_path.read_text().strip()
|
||||
code, out, err = run(
|
||||
f'curl -s -o /dev/null -w "%{{http_code}}" -H "Authorization: token {token}" '
|
||||
f'"{FORGE_URL}/api/v1/user"',
|
||||
timeout=10
|
||||
)
|
||||
if code == 0 and out == "200":
|
||||
return True, "reachable"
|
||||
return False, f"HTTP {out}"
|
||||
|
||||
def check_vps(ip, name):
|
||||
"""Can we SSH into a VPS?"""
|
||||
code, out, err = run(f"ssh -o ConnectTimeout=5 root@{ip} 'echo alive'", timeout=10)
|
||||
if code == 0 and "alive" in out:
|
||||
return True, "alive"
|
||||
return False, f"unreachable"
|
||||
|
||||
# ─── FALLBACK ACTIONS ───
|
||||
|
||||
def fallback_to_local_model(cfg):
|
||||
"""Switch primary model from Kimi to local llama.cpp"""
|
||||
if not BACKUP_CONFIG.exists():
|
||||
shutil.copy2(CONFIG_PATH, BACKUP_CONFIG)
|
||||
|
||||
cfg["model"]["provider"] = "local-llama.cpp"
|
||||
cfg["model"]["default"] = "hermes3"
|
||||
save_config(cfg)
|
||||
return "Switched primary model to local-llama.cpp/hermes3"
|
||||
|
||||
def fallback_to_ollama(cfg):
|
||||
"""Switch to Ollama if llama.cpp is also down"""
|
||||
if not BACKUP_CONFIG.exists():
|
||||
shutil.copy2(CONFIG_PATH, BACKUP_CONFIG)
|
||||
|
||||
cfg["model"]["provider"] = "ollama"
|
||||
cfg["model"]["default"] = "gemma4:latest"
|
||||
save_config(cfg)
|
||||
return "Switched primary model to ollama/gemma4:latest"
|
||||
|
||||
def enter_safe_mode(state):
|
||||
"""Pause all non-essential cron jobs, alert Alexander"""
|
||||
state["safe_mode"] = True
|
||||
state["safe_mode_entered"] = datetime.now().isoformat()
|
||||
save_state(state)
|
||||
return "SAFE MODE: All inference down. Cron jobs should be paused. Alert Alexander."
|
||||
|
||||
def restore_config():
|
||||
"""Restore pre-fallback config when primary recovers"""
|
||||
if BACKUP_CONFIG.exists():
|
||||
shutil.copy2(BACKUP_CONFIG, CONFIG_PATH)
|
||||
BACKUP_CONFIG.unlink()
|
||||
return "Restored original config from backup"
|
||||
return "No backup config to restore"
|
||||
|
||||
# ─── MAIN DIAGNOSIS AND FALLBACK ENGINE ───
|
||||
|
||||
def diagnose_and_fallback():
|
||||
state = load_state()
|
||||
cfg = load_config()
|
||||
|
||||
results = {
|
||||
"timestamp": datetime.now().isoformat(),
|
||||
"checks": {},
|
||||
"actions": [],
|
||||
"status": "healthy"
|
||||
}
|
||||
|
||||
# Check all systems
|
||||
kimi_ok, kimi_msg = check_kimi()
|
||||
results["checks"]["kimi-coding"] = {"ok": kimi_ok, "msg": kimi_msg}
|
||||
|
||||
llama_ok, llama_msg = check_local_llama()
|
||||
results["checks"]["local_llama"] = {"ok": llama_ok, "msg": llama_msg}
|
||||
|
||||
ollama_ok, ollama_msg = check_ollama()
|
||||
results["checks"]["ollama"] = {"ok": ollama_ok, "msg": ollama_msg}
|
||||
|
||||
gitea_ok, gitea_msg = check_gitea()
|
||||
results["checks"]["gitea"] = {"ok": gitea_ok, "msg": gitea_msg}
|
||||
|
||||
# VPS checks
|
||||
vpses = [
|
||||
("167.99.126.228", "Allegro"),
|
||||
("143.198.27.163", "Ezra"),
|
||||
("159.203.146.185", "Bezalel"),
|
||||
]
|
||||
for ip, name in vpses:
|
||||
vps_ok, vps_msg = check_vps(ip, name)
|
||||
results["checks"][f"vps_{name.lower()}"] = {"ok": vps_ok, "msg": vps_msg}
|
||||
|
||||
current_provider = cfg.get("model", {}).get("provider", "kimi-coding")
|
||||
|
||||
# ─── FALLBACK LOGIC ───
|
||||
|
||||
# Case 1: Primary (Kimi) down, local available
|
||||
if not kimi_ok and current_provider == "kimi-coding":
|
||||
if llama_ok:
|
||||
msg = fallback_to_local_model(cfg)
|
||||
results["actions"].append(msg)
|
||||
state["active_fallbacks"].append("kimi->local-llama")
|
||||
results["status"] = "degraded_local"
|
||||
elif ollama_ok:
|
||||
msg = fallback_to_ollama(cfg)
|
||||
results["actions"].append(msg)
|
||||
state["active_fallbacks"].append("kimi->ollama")
|
||||
results["status"] = "degraded_ollama"
|
||||
else:
|
||||
msg = enter_safe_mode(state)
|
||||
results["actions"].append(msg)
|
||||
results["status"] = "safe_mode"
|
||||
|
||||
# Case 2: Already on fallback, check if primary recovered
|
||||
elif kimi_ok and "kimi->local-llama" in state.get("active_fallbacks", []):
|
||||
msg = restore_config()
|
||||
results["actions"].append(msg)
|
||||
state["active_fallbacks"].remove("kimi->local-llama")
|
||||
results["status"] = "recovered"
|
||||
elif kimi_ok and "kimi->ollama" in state.get("active_fallbacks", []):
|
||||
msg = restore_config()
|
||||
results["actions"].append(msg)
|
||||
state["active_fallbacks"].remove("kimi->ollama")
|
||||
results["status"] = "recovered"
|
||||
|
||||
# Case 3: Gitea down — just flag it, work locally
|
||||
if not gitea_ok:
|
||||
results["actions"].append("WARN: Gitea unreachable — work cached locally until recovery")
|
||||
if "gitea_down" not in state.get("active_fallbacks", []):
|
||||
state["active_fallbacks"].append("gitea_down")
|
||||
results["status"] = max(results["status"], "degraded_gitea", key=lambda x: ["healthy", "recovered", "degraded_gitea", "degraded_local", "degraded_ollama", "safe_mode"].index(x) if x in ["healthy", "recovered", "degraded_gitea", "degraded_local", "degraded_ollama", "safe_mode"] else 0)
|
||||
elif "gitea_down" in state.get("active_fallbacks", []):
|
||||
state["active_fallbacks"].remove("gitea_down")
|
||||
results["actions"].append("Gitea recovered — resume normal operations")
|
||||
|
||||
# Case 4: VPS agents down
|
||||
for ip, name in vpses:
|
||||
key = f"vps_{name.lower()}"
|
||||
if not results["checks"][key]["ok"]:
|
||||
results["actions"].append(f"ALERT: {name} VPS ({ip}) unreachable — lazarus protocol needed")
|
||||
|
||||
save_state(state)
|
||||
return results
|
||||
|
||||
if __name__ == "__main__":
|
||||
results = diagnose_and_fallback()
|
||||
print(json.dumps(results, indent=2))
|
||||
|
||||
# Exit codes for cron integration
|
||||
if results["status"] == "safe_mode":
|
||||
sys.exit(2)
|
||||
elif results["status"].startswith("degraded"):
|
||||
sys.exit(1)
|
||||
else:
|
||||
sys.exit(0)
|
||||
|
||||
@@ -19,25 +19,25 @@ PASS=0
|
||||
FAIL=0
|
||||
WARN=0
|
||||
|
||||
check_anthropic_model() {
|
||||
check_kimi_model() {
|
||||
local model="$1"
|
||||
local label="$2"
|
||||
local api_key="${ANTHROPIC_API_KEY:-}"
|
||||
local api_key="${KIMI_API_KEY:-}"
|
||||
|
||||
if [ -z "$api_key" ]; then
|
||||
# Try loading from .env
|
||||
api_key=$(grep '^ANTHROPIC_API_KEY=' "${HERMES_HOME:-$HOME/.hermes}/.env" 2>/dev/null | head -1 | cut -d= -f2- | tr -d "'\"" || echo "")
|
||||
api_key=$(grep '^KIMI_API_KEY=' "${HERMES_HOME:-$HOME/.hermes}/.env" 2>/dev/null | head -1 | cut -d= -f2- | tr -d "'\"" || echo "")
|
||||
fi
|
||||
|
||||
if [ -z "$api_key" ]; then
|
||||
log "SKIP [$label] $model -- no ANTHROPIC_API_KEY"
|
||||
log "SKIP [$label] $model -- no KIMI_API_KEY"
|
||||
return 0
|
||||
fi
|
||||
|
||||
response=$(curl -sf --max-time 10 -X POST \
|
||||
"https://api.anthropic.com/v1/messages" \
|
||||
"https://api.kimi.com/coding/v1/chat/completions" \
|
||||
-H "x-api-key: ${api_key}" \
|
||||
-H "anthropic-version: 2023-06-01" \
|
||||
-H "x-api-provider: kimi-coding" \
|
||||
-H "content-type: application/json" \
|
||||
-d "{\"model\":\"${model}\",\"max_tokens\":1,\"messages\":[{\"role\":\"user\",\"content\":\"hi\"}]}" 2>&1 || echo "ERROR")
|
||||
|
||||
@@ -85,26 +85,26 @@ else:
|
||||
print('')
|
||||
" 2>/dev/null || echo "")
|
||||
|
||||
if [ -n "$primary" ] && [ "$provider" = "anthropic" ]; then
|
||||
if check_anthropic_model "$primary" "PRIMARY"; then
|
||||
if [ -n "$primary" ] && [ "$provider" = "kimi-coding" ]; then
|
||||
if check_kimi_model "$primary" "PRIMARY"; then
|
||||
PASS=$((PASS + 1))
|
||||
else
|
||||
rc=$?
|
||||
if [ "$rc" -eq 1 ]; then
|
||||
FAIL=$((FAIL + 1))
|
||||
log "CRITICAL: Primary model $primary is DEAD. Loops will fail."
|
||||
log "Known good alternatives: claude-opus-4.6, claude-haiku-4-5-20251001"
|
||||
log "Known good alternatives: kimi-k2.5, google/gemini-2.5-pro"
|
||||
else
|
||||
WARN=$((WARN + 1))
|
||||
fi
|
||||
fi
|
||||
elif [ -n "$primary" ]; then
|
||||
log "SKIP [PRIMARY] $primary -- non-anthropic provider ($provider), no validator yet"
|
||||
log "SKIP [PRIMARY] $primary -- non-kimi provider ($provider), no validator yet"
|
||||
fi
|
||||
|
||||
# Cron model check (haiku)
|
||||
CRON_MODEL="claude-haiku-4-5-20251001"
|
||||
if check_anthropic_model "$CRON_MODEL" "CRON"; then
|
||||
CRON_MODEL="kimi-k2.5"
|
||||
if check_kimi_model "$CRON_MODEL" "CRON"; then
|
||||
PASS=$((PASS + 1))
|
||||
else
|
||||
rc=$?
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
# Uses Hermes CLI plus workforce-manager to triage and review.
|
||||
# Timmy is the brain. Other agents are the hands.
|
||||
|
||||
set -uo pipefail
|
||||
set -uo pipefail\n\nSCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
||||
|
||||
LOG_DIR="$HOME/.hermes/logs"
|
||||
LOG="$LOG_DIR/timmy-orchestrator.log"
|
||||
@@ -40,6 +40,7 @@ gather_state() {
|
||||
> "$state_dir/unassigned.txt"
|
||||
> "$state_dir/open_prs.txt"
|
||||
> "$state_dir/agent_status.txt"
|
||||
> "$state_dir/uncommitted_work.txt"
|
||||
|
||||
for repo in $REPOS; do
|
||||
local short=$(echo "$repo" | cut -d/ -f2)
|
||||
@@ -71,6 +72,24 @@ for p in json.load(sys.stdin):
|
||||
tail -50 "/tmp/kimi-heartbeat.log" 2>/dev/null | grep -c "FAILED:" | xargs -I{} echo "Kimi recent failures: {}" >> "$state_dir/agent_status.txt"
|
||||
tail -1 "/tmp/kimi-heartbeat.log" 2>/dev/null | xargs -I{} echo "Kimi last event: {}" >> "$state_dir/agent_status.txt"
|
||||
|
||||
# Scan worktrees for uncommitted work
|
||||
for wt_dir in "$HOME/worktrees"/*/; do
|
||||
[ -d "$wt_dir" ] || continue
|
||||
[ -d "$wt_dir/.git" ] || continue
|
||||
local dirty
|
||||
dirty=$(cd "$wt_dir" && git status --porcelain 2>/dev/null | wc -l | tr -d " ")
|
||||
if [ "${dirty:-0}" -gt 0 ]; then
|
||||
local branch
|
||||
branch=$(cd "$wt_dir" && git branch --show-current 2>/dev/null || echo "?")
|
||||
local age=""
|
||||
local last_commit
|
||||
last_commit=$(cd "$wt_dir" && git log -1 --format=%ct 2>/dev/null || echo 0)
|
||||
local now=$(date +%s)
|
||||
local stale_mins=$(( (now - last_commit) / 60 ))
|
||||
echo "DIR=$wt_dir BRANCH=$branch DIRTY=$dirty STALE=${stale_mins}m" >> "$state_dir/uncommitted_work.txt"
|
||||
fi
|
||||
done
|
||||
|
||||
echo "$state_dir"
|
||||
}
|
||||
|
||||
@@ -81,6 +100,25 @@ run_triage() {
|
||||
|
||||
log "Cycle: $unassigned_count unassigned, $pr_count open PRs"
|
||||
|
||||
# Check for uncommitted work — nag if stale
|
||||
local uncommitted_count
|
||||
uncommitted_count=$(wc -l < "$state_dir/uncommitted_work.txt" 2>/dev/null | tr -d " " || echo 0)
|
||||
if [ "${uncommitted_count:-0}" -gt 0 ]; then
|
||||
log "WARNING: $uncommitted_count worktree(s) with uncommitted work"
|
||||
while IFS= read -r line; do
|
||||
log " UNCOMMITTED: $line"
|
||||
# Auto-commit stale work (>60 min without commit)
|
||||
local stale=$(echo "$line" | sed 's/.*STALE=\([0-9]*\)m.*/\1/')
|
||||
local wt_dir=$(echo "$line" | sed 's/.*DIR=\([^ ]*\) .*/\1/')
|
||||
if [ "${stale:-0}" -gt 60 ]; then
|
||||
log " AUTO-COMMITTING stale work in $wt_dir (${stale}m stale)"
|
||||
(cd "$wt_dir" && git add -A && git commit -m "WIP: orchestrator auto-commit — ${stale}m stale work
|
||||
|
||||
Preserved by timmy-orchestrator to prevent loss." 2>/dev/null && git push 2>/dev/null) && log " COMMITTED: $wt_dir" || log " COMMIT FAILED: $wt_dir"
|
||||
fi
|
||||
done < "$state_dir/uncommitted_work.txt"
|
||||
fi
|
||||
|
||||
# If nothing to do, skip the LLM call
|
||||
if [ "$unassigned_count" -eq 0 ] && [ "$pr_count" -eq 0 ]; then
|
||||
log "Nothing to triage"
|
||||
@@ -198,6 +236,12 @@ FOOTER
|
||||
log "=== Timmy Orchestrator Started (PID $$) ==="
|
||||
log "Cycle: ${CYCLE_INTERVAL}s | Auto-assign: ${AUTO_ASSIGN_UNASSIGNED} | Inference surface: Hermes CLI"
|
||||
|
||||
# Start auto-commit-guard daemon for work preservation
|
||||
if ! pgrep -f "auto-commit-guard.sh" >/dev/null 2>&1; then
|
||||
nohup bash "$SCRIPT_DIR/auto-commit-guard.sh" 120 >> "$LOG_DIR/auto-commit-guard.log" 2>&1 &
|
||||
log "Started auto-commit-guard daemon (PID $!)"
|
||||
fi
|
||||
|
||||
WORKFORCE_CYCLE=0
|
||||
|
||||
while true; do
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
{
|
||||
"updated_at": "2026-03-28T09:54:34.822062",
|
||||
"updated_at": "2026-04-13T02:02:07.001824",
|
||||
"platforms": {
|
||||
"discord": [
|
||||
{
|
||||
@@ -27,11 +27,81 @@
|
||||
"name": "Timmy Time",
|
||||
"type": "group",
|
||||
"thread_id": null
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:85",
|
||||
"name": "Timmy Time / topic 85",
|
||||
"type": "group",
|
||||
"thread_id": "85"
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:111",
|
||||
"name": "Timmy Time / topic 111",
|
||||
"type": "group",
|
||||
"thread_id": "111"
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:173",
|
||||
"name": "Timmy Time / topic 173",
|
||||
"type": "group",
|
||||
"thread_id": "173"
|
||||
},
|
||||
{
|
||||
"id": "7635059073",
|
||||
"name": "Trip T",
|
||||
"type": "dm",
|
||||
"thread_id": null
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:244",
|
||||
"name": "Timmy Time / topic 244",
|
||||
"type": "group",
|
||||
"thread_id": "244"
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:972",
|
||||
"name": "Timmy Time / topic 972",
|
||||
"type": "group",
|
||||
"thread_id": "972"
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:931",
|
||||
"name": "Timmy Time / topic 931",
|
||||
"type": "group",
|
||||
"thread_id": "931"
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:957",
|
||||
"name": "Timmy Time / topic 957",
|
||||
"type": "group",
|
||||
"thread_id": "957"
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:1297",
|
||||
"name": "Timmy Time / topic 1297",
|
||||
"type": "group",
|
||||
"thread_id": "1297"
|
||||
},
|
||||
{
|
||||
"id": "-1003664764329:1316",
|
||||
"name": "Timmy Time / topic 1316",
|
||||
"type": "group",
|
||||
"thread_id": "1316"
|
||||
}
|
||||
],
|
||||
"whatsapp": [],
|
||||
"slack": [],
|
||||
"signal": [],
|
||||
"mattermost": [],
|
||||
"matrix": [],
|
||||
"homeassistant": [],
|
||||
"email": [],
|
||||
"sms": []
|
||||
"sms": [],
|
||||
"dingtalk": [],
|
||||
"feishu": [],
|
||||
"wecom": [],
|
||||
"wecom_callback": [],
|
||||
"weixin": [],
|
||||
"bluebubbles": []
|
||||
}
|
||||
}
|
||||
@@ -7,7 +7,7 @@ Purpose:
|
||||
|
||||
## What it is
|
||||
|
||||
Code Claw is a separate local runtime from Hermes/OpenClaw.
|
||||
Code Claw is a separate local runtime from Hermes.
|
||||
|
||||
Current lane:
|
||||
- runtime: local patched `~/code-claw`
|
||||
|
||||
218
config.yaml
218
config.yaml
@@ -1,31 +1,23 @@
|
||||
model:
|
||||
default: hermes4:14b
|
||||
provider: custom
|
||||
context_length: 65536
|
||||
base_url: http://localhost:8081/v1
|
||||
default: claude-opus-4-6
|
||||
provider: anthropic
|
||||
toolsets:
|
||||
- all
|
||||
agent:
|
||||
max_turns: 30
|
||||
reasoning_effort: xhigh
|
||||
reasoning_effort: medium
|
||||
verbose: false
|
||||
terminal:
|
||||
backend: local
|
||||
cwd: .
|
||||
timeout: 180
|
||||
env_passthrough: []
|
||||
docker_image: nikolaik/python-nodejs:python3.11-nodejs20
|
||||
docker_forward_env: []
|
||||
singularity_image: docker://nikolaik/python-nodejs:python3.11-nodejs20
|
||||
modal_image: nikolaik/python-nodejs:python3.11-nodejs20
|
||||
daytona_image: nikolaik/python-nodejs:python3.11-nodejs20
|
||||
container_cpu: 1
|
||||
container_embeddings:
|
||||
provider: ollama
|
||||
model: nomic-embed-text
|
||||
base_url: http://localhost:11434/v1
|
||||
|
||||
memory: 5120
|
||||
container_memory: 5120
|
||||
container_disk: 51200
|
||||
container_persistent: true
|
||||
docker_volumes: []
|
||||
@@ -33,89 +25,74 @@ memory: 5120
|
||||
persistent_shell: true
|
||||
browser:
|
||||
inactivity_timeout: 120
|
||||
command_timeout: 30
|
||||
record_sessions: false
|
||||
checkpoints:
|
||||
enabled: true
|
||||
enabled: false
|
||||
max_snapshots: 50
|
||||
compression:
|
||||
enabled: true
|
||||
threshold: 0.5
|
||||
target_ratio: 0.2
|
||||
protect_last_n: 20
|
||||
summary_model: ''
|
||||
summary_provider: ''
|
||||
summary_base_url: ''
|
||||
synthesis_model:
|
||||
provider: custom
|
||||
model: llama3:70b
|
||||
base_url: http://localhost:8081/v1
|
||||
|
||||
summary_model: qwen3:30b
|
||||
summary_provider: custom
|
||||
summary_base_url: http://localhost:11434/v1
|
||||
smart_model_routing:
|
||||
enabled: true
|
||||
max_simple_chars: 400
|
||||
max_simple_words: 75
|
||||
cheap_model:
|
||||
provider: 'ollama'
|
||||
model: 'gemma2:2b'
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: ''
|
||||
enabled: false
|
||||
max_simple_chars: 160
|
||||
max_simple_words: 28
|
||||
cheap_model: {}
|
||||
auxiliary:
|
||||
vision:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
timeout: 30
|
||||
provider: custom
|
||||
model: qwen3:30b
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: 'ollama'
|
||||
web_extract:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
provider: custom
|
||||
model: qwen3:30b
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: 'ollama'
|
||||
compression:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
provider: custom
|
||||
model: qwen3:30b
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: 'ollama'
|
||||
session_search:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
provider: custom
|
||||
model: qwen3:30b
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: 'ollama'
|
||||
skills_hub:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
provider: custom
|
||||
model: qwen3:30b
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: 'ollama'
|
||||
approval:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
mcp:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
provider: custom
|
||||
model: qwen3:30b
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: 'ollama'
|
||||
flush_memories:
|
||||
provider: auto
|
||||
model: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
provider: custom
|
||||
model: qwen3:30b
|
||||
base_url: 'http://localhost:11434/v1'
|
||||
api_key: 'ollama'
|
||||
display:
|
||||
compact: false
|
||||
personality: ''
|
||||
resume_display: full
|
||||
busy_input_mode: interrupt
|
||||
bell_on_complete: false
|
||||
show_reasoning: false
|
||||
streaming: false
|
||||
show_cost: false
|
||||
skin: timmy
|
||||
tool_progress_command: false
|
||||
tool_progress: all
|
||||
privacy:
|
||||
redact_pii: true
|
||||
redact_pii: false
|
||||
tts:
|
||||
provider: edge
|
||||
edge:
|
||||
@@ -124,7 +101,7 @@ tts:
|
||||
voice_id: pNInz6obpgDQGcFmaJgB
|
||||
model_id: eleven_multilingual_v2
|
||||
openai:
|
||||
model: '' # disabled — use edge TTS locally
|
||||
model: gpt-4o-mini-tts
|
||||
voice: alloy
|
||||
neutts:
|
||||
ref_audio: ''
|
||||
@@ -160,7 +137,6 @@ delegation:
|
||||
provider: ''
|
||||
base_url: ''
|
||||
api_key: ''
|
||||
max_iterations: 50
|
||||
prefill_messages_file: ''
|
||||
honcho: {}
|
||||
timezone: ''
|
||||
@@ -174,16 +150,7 @@ approvals:
|
||||
command_allowlist: []
|
||||
quick_commands: {}
|
||||
personalities: {}
|
||||
mesh:
|
||||
enabled: true
|
||||
blackboard_provider: local
|
||||
nostr_discovery: true
|
||||
consensus_mode: competitive
|
||||
|
||||
security:
|
||||
sovereign_audit: true
|
||||
no_phone_home: true
|
||||
|
||||
redact_secrets: true
|
||||
tirith_enabled: true
|
||||
tirith_path: tirith
|
||||
@@ -193,55 +160,66 @@ security:
|
||||
enabled: false
|
||||
domains: []
|
||||
shared_files: []
|
||||
_config_version: 10
|
||||
platforms:
|
||||
api_server:
|
||||
enabled: true
|
||||
extra:
|
||||
host: 0.0.0.0
|
||||
port: 8642
|
||||
# Author whitelist for task router (Issue #132)
|
||||
# Only users in this list can submit tasks via Gitea issues
|
||||
# Empty list = deny all (secure by default)
|
||||
# Set via env var TIMMY_AUTHOR_WHITELIST as comma-separated list
|
||||
author_whitelist: []
|
||||
_config_version: 9
|
||||
session_reset:
|
||||
mode: none
|
||||
idle_minutes: 0
|
||||
custom_providers:
|
||||
- name: Local llama.cpp
|
||||
base_url: http://localhost:8081/v1
|
||||
api_key: none
|
||||
model: hermes4:14b
|
||||
# ── Emergency cloud provider — not used by default or any cron job.
|
||||
# Available for explicit override only: hermes --model gemini-2.5-pro
|
||||
- name: Google Gemini (emergency only)
|
||||
base_url: https://generativelanguage.googleapis.com/v1beta/openai
|
||||
api_key_env: GEMINI_API_KEY
|
||||
model: gemini-2.5-pro
|
||||
- name: Local Ollama
|
||||
base_url: http://localhost:11434/v1
|
||||
api_key: ollama
|
||||
model: qwen3:30b
|
||||
system_prompt_suffix: "You are Timmy. Your soul is defined in SOUL.md \u2014 read\
|
||||
\ it, live it.\nYou run locally on your owner's machine via llama.cpp. You never\
|
||||
\ phone home.\nYou speak plainly. You prefer short sentences. Brevity is a kindness.\n\
|
||||
When you don't know something, say so. Refusal over fabrication.\nSovereignty and\
|
||||
\ service always.\n"
|
||||
\ it, live it.\nYou run locally on your owner's machine via Ollama. You never phone\
|
||||
\ home.\nYou speak plainly. You prefer short sentences. Brevity is a kindness.\n\
|
||||
Source distinction: Tag every factual claim inline. Default is [generated] — you\
|
||||
\ are pattern-matching from training data. Only use [retrieved] when you can name\
|
||||
\ the specific tool call or document from THIS conversation that provided the fact.\
|
||||
\ If no tool was called, every claim is [generated]. No exceptions.\n\
|
||||
Refusal over fabrication: When you generate a specific claim — a date, a number,\
|
||||
\ a price, a version, a URL, a current event — and you cannot name a source from\
|
||||
\ this conversation, say 'I don't know' instead. Do not guess. Do not hedge with\
|
||||
\ 'probably' or 'approximately' as a substitute for knowledge. If your only source\
|
||||
\ is training data and the claim could be wrong or outdated, the honest answer is\
|
||||
\ 'I don't know — I can look this up if you'd like.' Prefer a true 'I don't know'\
|
||||
\ over a plausible fabrication.\nSovereignty and service always.\n"
|
||||
skills:
|
||||
creation_nudge_interval: 15
|
||||
DISCORD_HOME_CHANNEL: '1476292315814297772'
|
||||
providers:
|
||||
ollama:
|
||||
base_url: http://localhost:11434/v1
|
||||
model: hermes3:latest
|
||||
mcp_servers:
|
||||
morrowind:
|
||||
command: python3
|
||||
args:
|
||||
- /Users/apayne/.timmy/morrowind/mcp_server.py
|
||||
env: {}
|
||||
timeout: 30
|
||||
crucible:
|
||||
command: /Users/apayne/.hermes/hermes-agent/venv/bin/python3
|
||||
args:
|
||||
- /Users/apayne/.hermes/bin/crucible_mcp_server.py
|
||||
env: {}
|
||||
timeout: 120
|
||||
connect_timeout: 60
|
||||
fallback_model:
|
||||
provider: ollama
|
||||
model: hermes3:latest
|
||||
base_url: http://localhost:11434/v1
|
||||
api_key: ''
|
||||
|
||||
# ── Fallback Model ────────────────────────────────────────────────────
|
||||
# Automatic provider failover when primary is unavailable.
|
||||
# Uncomment and configure to enable. Triggers on rate limits (429),
|
||||
# overload (529), service errors (503), or connection failures.
|
||||
#
|
||||
# Supported providers:
|
||||
# openrouter (OPENROUTER_API_KEY) — routes to any model
|
||||
# openai-codex (OAuth — hermes login) — OpenAI Codex
|
||||
# nous (OAuth — hermes login) — Nous Portal
|
||||
# zai (ZAI_API_KEY) — Z.AI / GLM
|
||||
# kimi-coding (KIMI_API_KEY) — Kimi / Moonshot
|
||||
# minimax (MINIMAX_API_KEY) — MiniMax
|
||||
# minimax-cn (MINIMAX_CN_API_KEY) — MiniMax (China)
|
||||
#
|
||||
# For custom OpenAI-compatible endpoints, add base_url and api_key_env.
|
||||
#
|
||||
# fallback_model:
|
||||
# provider: openrouter
|
||||
# model: anthropic/claude-sonnet-4
|
||||
#
|
||||
# ── Smart Model Routing ────────────────────────────────────────────────
|
||||
# Optional cheap-vs-strong routing for simple turns.
|
||||
# Keeps the primary model for complex work, but can route short/simple
|
||||
# messages to a cheaper model across providers.
|
||||
#
|
||||
# smart_model_routing:
|
||||
# enabled: true
|
||||
# max_simple_chars: 160
|
||||
# max_simple_words: 28
|
||||
# cheap_model:
|
||||
# provider: openrouter
|
||||
# model: google/gemini-2.5-flash
|
||||
|
||||
@@ -168,7 +168,35 @@
|
||||
"paused_reason": null,
|
||||
"skills": [],
|
||||
"skill": null
|
||||
},
|
||||
{
|
||||
"id": "overnight-rd-nightly",
|
||||
"name": "Overnight R&D Loop",
|
||||
"prompt": "Run the overnight R&D automation: Deep Dive paper synthesis, tightening loop for tool-use training data, DPO export sweep, morning briefing prep. All local inference via Ollama.",
|
||||
"schedule": {
|
||||
"kind": "cron",
|
||||
"expr": "0 2 * * *",
|
||||
"display": "0 2 * * * (10 PM EDT)"
|
||||
},
|
||||
"schedule_display": "Nightly at 10 PM EDT",
|
||||
"repeat": {
|
||||
"times": null,
|
||||
"completed": 0
|
||||
},
|
||||
"enabled": true,
|
||||
"created_at": "2026-04-13T02:00:00+00:00",
|
||||
"next_run_at": null,
|
||||
"last_run_at": null,
|
||||
"last_status": null,
|
||||
"last_error": null,
|
||||
"deliver": "local",
|
||||
"origin": "perplexity/overnight-rd-automation",
|
||||
"state": "scheduled",
|
||||
"paused_at": null,
|
||||
"paused_reason": null,
|
||||
"skills": [],
|
||||
"skill": null
|
||||
}
|
||||
],
|
||||
"updated_at": "2026-04-07T15:00:00+00:00"
|
||||
"updated_at": "2026-04-13T02:00:00+00:00"
|
||||
}
|
||||
|
||||
24
deploy/auto-commit-guard.plist
Normal file
24
deploy/auto-commit-guard.plist
Normal file
@@ -0,0 +1,24 @@
|
||||
<?xml version="1.0" encoding="UTF-8"?>
|
||||
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
|
||||
<plist version="1.0">
|
||||
<dict>
|
||||
<key>Label</key>
|
||||
<string>ai.timmy.auto-commit-guard</string>
|
||||
<key>ProgramArguments</key>
|
||||
<array>
|
||||
<string>/bin/bash</string>
|
||||
<string>/Users/apayne/.hermes/bin/auto-commit-guard.sh</string>
|
||||
<string>120</string>
|
||||
</array>
|
||||
<key>RunAtLoad</key>
|
||||
<true/>
|
||||
<key>KeepAlive</key>
|
||||
<true/>
|
||||
<key>StandardOutPath</key>
|
||||
<string>/Users/apayne/.hermes/logs/auto-commit-guard.stdout.log</string>
|
||||
<key>StandardErrorPath</key>
|
||||
<string>/Users/apayne/.hermes/logs/auto-commit-guard.stderr.log</string>
|
||||
<key>WorkingDirectory</key>
|
||||
<string>/Users/apayne</string>
|
||||
</dict>
|
||||
</plist>
|
||||
@@ -3,7 +3,7 @@
|
||||
Purpose:
|
||||
- stand up the third wizard house as a Kimi-backed coding worker
|
||||
- keep Hermes as the durable harness
|
||||
- treat OpenClaw as optional shell frontage, not the bones
|
||||
- Hermes is the durable harness — no intermediary gateway layers
|
||||
|
||||
Local proof already achieved:
|
||||
|
||||
@@ -40,5 +40,5 @@ bin/deploy-allegro-house.sh root@167.99.126.228
|
||||
|
||||
Important nuance:
|
||||
- the Hermes/Kimi lane is the proven path
|
||||
- direct embedded OpenClaw Kimi model routing was not yet reliable locally
|
||||
- direct embedded Kimi model routing was not yet reliable locally
|
||||
- so the remote deployment keeps the minimal, proven architecture: Hermes house first
|
||||
@@ -81,17 +81,6 @@ launchctl bootstrap gui/$(id -u) ~/Library/LaunchAgents/ai.hermes.gateway.plist
|
||||
- Old-state risk:
|
||||
- same class as main gateway, but isolated to fenrir profile state
|
||||
|
||||
#### 3. ai.openclaw.gateway
|
||||
- Plist: ~/Library/LaunchAgents/ai.openclaw.gateway.plist
|
||||
- Command: `node .../openclaw/dist/index.js gateway --port 18789`
|
||||
- Logs:
|
||||
- `~/.openclaw/logs/gateway.log`
|
||||
- `~/.openclaw/logs/gateway.err.log`
|
||||
- KeepAlive: yes
|
||||
- RunAtLoad: yes
|
||||
- Old-state risk:
|
||||
- long-lived gateway survives toolchain assumptions and keeps accepting work even if upstream routing changed
|
||||
|
||||
#### 4. ai.timmy.kimi-heartbeat
|
||||
- Plist: ~/Library/LaunchAgents/ai.timmy.kimi-heartbeat.plist
|
||||
- Command: `/bin/bash ~/.timmy/uniwizard/kimi-heartbeat.sh`
|
||||
@@ -295,7 +284,7 @@ launchctl list | egrep 'timmy|kimi|claude|max|dashboard|matrix|gateway|huey'
|
||||
|
||||
List Timmy/Hermes launch agent files:
|
||||
```bash
|
||||
find ~/Library/LaunchAgents -maxdepth 1 -name '*.plist' | egrep 'timmy|hermes|openclaw|tower'
|
||||
find ~/Library/LaunchAgents -maxdepth 1 -name '*.plist' | egrep 'timmy|hermes|tower'
|
||||
```
|
||||
|
||||
List running loop scripts:
|
||||
@@ -316,7 +305,6 @@ launchctl bootout gui/$(id -u) ~/Library/LaunchAgents/ai.timmy.kimi-heartbeat.pl
|
||||
launchctl bootout gui/$(id -u) ~/Library/LaunchAgents/ai.timmy.claudemax-watchdog.plist || true
|
||||
launchctl bootout gui/$(id -u) ~/Library/LaunchAgents/ai.hermes.gateway.plist || true
|
||||
launchctl bootout gui/$(id -u) ~/Library/LaunchAgents/ai.hermes.gateway-fenrir.plist || true
|
||||
launchctl bootout gui/$(id -u) ~/Library/LaunchAgents/ai.openclaw.gateway.plist || true
|
||||
```
|
||||
|
||||
2. Kill manual loops
|
||||
|
||||
68
docs/overnight-rd.md
Normal file
68
docs/overnight-rd.md
Normal file
@@ -0,0 +1,68 @@
|
||||
# Overnight R&D Automation
|
||||
|
||||
**Schedule**: Nightly at 10 PM EDT (02:00 UTC)
|
||||
**Duration**: ~2-4 hours (self-limiting, finishes before 6 AM morning report)
|
||||
**Cost**: $0 — all local Ollama inference
|
||||
|
||||
## Phases
|
||||
|
||||
### Phase 1: Deep Dive Intelligence
|
||||
Runs the `intelligence/deepdive/pipeline.py` from the-nexus:
|
||||
- Aggregates arXiv CS.AI, CS.CL, CS.LG RSS feeds (last 24h)
|
||||
- Fetches OpenAI, Anthropic, DeepMind blog updates
|
||||
- Filters for relevance using sentence-transformers embeddings
|
||||
- Synthesizes a briefing using local Gemma 4 12B
|
||||
- Saves briefing to `~/briefings/`
|
||||
|
||||
### Phase 2: Tightening Loop
|
||||
Exercises Timmy's local tool-use capability:
|
||||
- 10 tasks × 3 cycles = 30 task attempts per night
|
||||
- File reading, writing, searching against real workspace files
|
||||
- Each result logged as JSONL for training data analysis
|
||||
- Tests sovereignty compliance (SOUL.md alignment, banned provider detection)
|
||||
|
||||
### Phase 3: DPO Export
|
||||
Sweeps overnight Hermes sessions for training pair extraction:
|
||||
- Converts good conversation pairs into DPO training format
|
||||
- Saves to `~/.timmy/training-data/dpo-pairs/`
|
||||
|
||||
### Phase 4: Morning Prep
|
||||
Compiles overnight findings into `~/.timmy/overnight-rd/latest_summary.md`
|
||||
for consumption by the 6 AM `good_morning_report` task.
|
||||
|
||||
## Approved Providers
|
||||
|
||||
| Slot | Provider | Model |
|
||||
|------|----------|-------|
|
||||
| Synthesis | Ollama | gemma4:12b |
|
||||
| Tool tasks | Ollama | hermes4:14b |
|
||||
| Fallback | Ollama | gemma4:12b |
|
||||
|
||||
Anthropic is permanently banned (BANNED_PROVIDERS.yml, 2026-04-09).
|
||||
|
||||
## Outputs
|
||||
|
||||
| Path | Content |
|
||||
|------|---------|
|
||||
| `~/.timmy/overnight-rd/{run_id}/rd_log.jsonl` | Full task log |
|
||||
| `~/.timmy/overnight-rd/{run_id}/rd_summary.md` | Run summary |
|
||||
| `~/.timmy/overnight-rd/latest_summary.md` | Latest summary (for morning report) |
|
||||
| `~/briefings/briefing_*.json` | Deep Dive briefings |
|
||||
|
||||
## Monitoring
|
||||
|
||||
Check the Huey consumer log:
|
||||
```bash
|
||||
tail -f ~/.timmy/timmy-config/logs/huey.log | grep overnight
|
||||
```
|
||||
|
||||
Check the latest run summary:
|
||||
```bash
|
||||
cat ~/.timmy/overnight-rd/latest_summary.md
|
||||
```
|
||||
|
||||
## Dependencies
|
||||
|
||||
- Deep Dive pipeline installed: `cd the-nexus/intelligence/deepdive && make install`
|
||||
- Ollama running with gemma4:12b and hermes4:14b models
|
||||
- Huey consumer running: `huey_consumer.py tasks.huey -w 2 -k thread`
|
||||
@@ -14,7 +14,7 @@ from crewai.tools import BaseTool
|
||||
|
||||
OPENROUTER_API_KEY = os.getenv(
|
||||
"OPENROUTER_API_KEY",
|
||||
"dsk-or-v1-f60c89db12040267458165cf192e815e339eb70548e4a0a461f5f0f69e6ef8b0",
|
||||
os.environ.get("OPENROUTER_API_KEY", ""),
|
||||
)
|
||||
|
||||
llm = LLM(
|
||||
|
||||
@@ -2,135 +2,128 @@ schema_version: 1
|
||||
status: proposed
|
||||
runtime_wiring: false
|
||||
owner: timmy-config
|
||||
|
||||
ownership:
|
||||
owns:
|
||||
- routing doctrine for task classes
|
||||
- sidecar-readable per-agent fallback portfolios
|
||||
- degraded-mode capability floors
|
||||
- routing doctrine for task classes
|
||||
- sidecar-readable per-agent fallback portfolios
|
||||
- degraded-mode capability floors
|
||||
does_not_own:
|
||||
- live queue state outside Gitea truth
|
||||
- launchd or loop process state
|
||||
- ad hoc worktree history
|
||||
|
||||
- live queue state outside Gitea truth
|
||||
- launchd or loop process state
|
||||
- ad hoc worktree history
|
||||
policy:
|
||||
require_four_slots_for_critical_agents: true
|
||||
terminal_fallback_must_be_usable: true
|
||||
forbid_synchronized_fleet_degradation: true
|
||||
forbid_human_token_fallbacks: true
|
||||
anti_correlation_rule: no two critical agents may share the same primary+fallback1 pair
|
||||
|
||||
sensitive_control_surfaces:
|
||||
- SOUL.md
|
||||
- config.yaml
|
||||
- deploy.sh
|
||||
- tasks.py
|
||||
- playbooks/
|
||||
- cron/
|
||||
- memories/
|
||||
- skins/
|
||||
- training/
|
||||
|
||||
- SOUL.md
|
||||
- config.yaml
|
||||
- deploy.sh
|
||||
- tasks.py
|
||||
- playbooks/
|
||||
- cron/
|
||||
- memories/
|
||||
- skins/
|
||||
- training/
|
||||
role_classes:
|
||||
judgment:
|
||||
current_surfaces:
|
||||
- playbooks/issue-triager.yaml
|
||||
- playbooks/pr-reviewer.yaml
|
||||
- playbooks/verified-logic.yaml
|
||||
- playbooks/issue-triager.yaml
|
||||
- playbooks/pr-reviewer.yaml
|
||||
- playbooks/verified-logic.yaml
|
||||
task_classes:
|
||||
- issue-triage
|
||||
- queue-routing
|
||||
- pr-review
|
||||
- proof-check
|
||||
- governance-review
|
||||
- issue-triage
|
||||
- queue-routing
|
||||
- pr-review
|
||||
- proof-check
|
||||
- governance-review
|
||||
degraded_mode:
|
||||
fallback2:
|
||||
allowed:
|
||||
- classify backlog
|
||||
- summarize risk
|
||||
- produce draft routing plans
|
||||
- leave bounded labels or comments with evidence
|
||||
- classify backlog
|
||||
- summarize risk
|
||||
- produce draft routing plans
|
||||
- leave bounded labels or comments with evidence
|
||||
denied:
|
||||
- merge pull requests
|
||||
- close or rewrite governing issues or PRs
|
||||
- mutate sensitive control surfaces
|
||||
- bulk-reassign the fleet
|
||||
- silently change routing policy
|
||||
- merge pull requests
|
||||
- close or rewrite governing issues or PRs
|
||||
- mutate sensitive control surfaces
|
||||
- bulk-reassign the fleet
|
||||
- silently change routing policy
|
||||
terminal:
|
||||
lane: report-and-route
|
||||
allowed:
|
||||
- classify backlog
|
||||
- summarize risk
|
||||
- produce draft routing artifacts
|
||||
- classify backlog
|
||||
- summarize risk
|
||||
- produce draft routing artifacts
|
||||
denied:
|
||||
- merge pull requests
|
||||
- bulk-reassign the fleet
|
||||
- mutate sensitive control surfaces
|
||||
|
||||
- merge pull requests
|
||||
- bulk-reassign the fleet
|
||||
- mutate sensitive control surfaces
|
||||
builder:
|
||||
current_surfaces:
|
||||
- playbooks/bug-fixer.yaml
|
||||
- playbooks/test-writer.yaml
|
||||
- playbooks/refactor-specialist.yaml
|
||||
- playbooks/bug-fixer.yaml
|
||||
- playbooks/test-writer.yaml
|
||||
- playbooks/refactor-specialist.yaml
|
||||
task_classes:
|
||||
- bug-fix
|
||||
- test-writing
|
||||
- refactor
|
||||
- bounded-docs-change
|
||||
- bug-fix
|
||||
- test-writing
|
||||
- refactor
|
||||
- bounded-docs-change
|
||||
degraded_mode:
|
||||
fallback2:
|
||||
allowed:
|
||||
- reversible single-issue changes
|
||||
- narrow docs fixes
|
||||
- test scaffolds and reproducers
|
||||
- reversible single-issue changes
|
||||
- narrow docs fixes
|
||||
- test scaffolds and reproducers
|
||||
denied:
|
||||
- cross-repo changes
|
||||
- sensitive control-surface edits
|
||||
- merge or release actions
|
||||
- cross-repo changes
|
||||
- sensitive control-surface edits
|
||||
- merge or release actions
|
||||
terminal:
|
||||
lane: narrow-patch
|
||||
allowed:
|
||||
- single-issue small patch
|
||||
- reproducer test
|
||||
- docs-only repair
|
||||
- single-issue small patch
|
||||
- reproducer test
|
||||
- docs-only repair
|
||||
denied:
|
||||
- sensitive control-surface edits
|
||||
- multi-file architecture work
|
||||
- irreversible actions
|
||||
|
||||
- sensitive control-surface edits
|
||||
- multi-file architecture work
|
||||
- irreversible actions
|
||||
wolf_bulk:
|
||||
current_surfaces:
|
||||
- docs/automation-inventory.md
|
||||
- FALSEWORK.md
|
||||
- docs/automation-inventory.md
|
||||
- FALSEWORK.md
|
||||
task_classes:
|
||||
- docs-inventory
|
||||
- log-summarization
|
||||
- queue-hygiene
|
||||
- repetitive-small-diff
|
||||
- research-sweep
|
||||
- docs-inventory
|
||||
- log-summarization
|
||||
- queue-hygiene
|
||||
- repetitive-small-diff
|
||||
- research-sweep
|
||||
degraded_mode:
|
||||
fallback2:
|
||||
allowed:
|
||||
- gather evidence
|
||||
- refresh inventories
|
||||
- summarize logs
|
||||
- propose labels or routes
|
||||
- gather evidence
|
||||
- refresh inventories
|
||||
- summarize logs
|
||||
- propose labels or routes
|
||||
denied:
|
||||
- multi-repo branch fanout
|
||||
- mass agent assignment
|
||||
- sensitive control-surface edits
|
||||
- irreversible queue mutation
|
||||
- multi-repo branch fanout
|
||||
- mass agent assignment
|
||||
- sensitive control-surface edits
|
||||
- irreversible queue mutation
|
||||
terminal:
|
||||
lane: gather-and-summarize
|
||||
allowed:
|
||||
- inventory refresh
|
||||
- evidence bundles
|
||||
- summaries
|
||||
- inventory refresh
|
||||
- evidence bundles
|
||||
- summaries
|
||||
denied:
|
||||
- multi-repo branch fanout
|
||||
- mass agent assignment
|
||||
- sensitive control-surface edits
|
||||
|
||||
- multi-repo branch fanout
|
||||
- mass agent assignment
|
||||
- sensitive control-surface edits
|
||||
routing:
|
||||
issue-triage: judgment
|
||||
queue-routing: judgment
|
||||
@@ -146,22 +139,20 @@ routing:
|
||||
queue-hygiene: wolf_bulk
|
||||
repetitive-small-diff: wolf_bulk
|
||||
research-sweep: wolf_bulk
|
||||
|
||||
promotion_rules:
|
||||
- If a wolf/bulk task touches a sensitive control surface, promote it to judgment.
|
||||
- If a builder task expands beyond 5 files, architecture review, or multi-repo coordination, promote it to judgment.
|
||||
- If a terminal lane cannot produce a usable artifact, the portfolio is invalid and must be redesigned before wiring.
|
||||
|
||||
- If a wolf/bulk task touches a sensitive control surface, promote it to judgment.
|
||||
- If a builder task expands beyond 5 files, architecture review, or multi-repo coordination, promote it to judgment.
|
||||
- If a terminal lane cannot produce a usable artifact, the portfolio is invalid and must be redesigned before wiring.
|
||||
agents:
|
||||
triage-coordinator:
|
||||
role_class: judgment
|
||||
critical: true
|
||||
current_playbooks:
|
||||
- playbooks/issue-triager.yaml
|
||||
- playbooks/issue-triager.yaml
|
||||
portfolio:
|
||||
primary:
|
||||
provider: anthropic
|
||||
model: claude-opus-4-6
|
||||
provider: kimi-coding
|
||||
model: kimi-k2.5
|
||||
lane: full-judgment
|
||||
fallback1:
|
||||
provider: openai-codex
|
||||
@@ -177,19 +168,18 @@ agents:
|
||||
lane: report-and-route
|
||||
local_capable: true
|
||||
usable_output:
|
||||
- backlog classification
|
||||
- routing draft
|
||||
- risk summary
|
||||
|
||||
- backlog classification
|
||||
- routing draft
|
||||
- risk summary
|
||||
pr-reviewer:
|
||||
role_class: judgment
|
||||
critical: true
|
||||
current_playbooks:
|
||||
- playbooks/pr-reviewer.yaml
|
||||
- playbooks/pr-reviewer.yaml
|
||||
portfolio:
|
||||
primary:
|
||||
provider: anthropic
|
||||
model: claude-opus-4-6
|
||||
provider: kimi-coding
|
||||
model: kimi-k2.5
|
||||
lane: full-review
|
||||
fallback1:
|
||||
provider: gemini
|
||||
@@ -205,17 +195,16 @@ agents:
|
||||
lane: low-stakes-diff-summary
|
||||
local_capable: false
|
||||
usable_output:
|
||||
- diff risk summary
|
||||
- explicit uncertainty notes
|
||||
- merge-block recommendation
|
||||
|
||||
- diff risk summary
|
||||
- explicit uncertainty notes
|
||||
- merge-block recommendation
|
||||
builder-main:
|
||||
role_class: builder
|
||||
critical: true
|
||||
current_playbooks:
|
||||
- playbooks/bug-fixer.yaml
|
||||
- playbooks/test-writer.yaml
|
||||
- playbooks/refactor-specialist.yaml
|
||||
- playbooks/bug-fixer.yaml
|
||||
- playbooks/test-writer.yaml
|
||||
- playbooks/refactor-specialist.yaml
|
||||
portfolio:
|
||||
primary:
|
||||
provider: openai-codex
|
||||
@@ -236,15 +225,14 @@ agents:
|
||||
lane: narrow-patch
|
||||
local_capable: true
|
||||
usable_output:
|
||||
- small patch
|
||||
- reproducer test
|
||||
- docs repair
|
||||
|
||||
- small patch
|
||||
- reproducer test
|
||||
- docs repair
|
||||
wolf-sweeper:
|
||||
role_class: wolf_bulk
|
||||
critical: true
|
||||
current_world_state:
|
||||
- docs/automation-inventory.md
|
||||
- docs/automation-inventory.md
|
||||
portfolio:
|
||||
primary:
|
||||
provider: gemini
|
||||
@@ -264,21 +252,20 @@ agents:
|
||||
lane: gather-and-summarize
|
||||
local_capable: true
|
||||
usable_output:
|
||||
- inventory refresh
|
||||
- evidence bundle
|
||||
- summary comment
|
||||
|
||||
- inventory refresh
|
||||
- evidence bundle
|
||||
- summary comment
|
||||
cross_checks:
|
||||
unique_primary_fallback1_pairs:
|
||||
triage-coordinator:
|
||||
- anthropic/claude-opus-4-6
|
||||
- openai-codex/codex
|
||||
- kimi-coding/kimi-k2.5
|
||||
- openai-codex/codex
|
||||
pr-reviewer:
|
||||
- anthropic/claude-opus-4-6
|
||||
- gemini/gemini-2.5-pro
|
||||
- kimi-coding/kimi-k2.5
|
||||
- gemini/gemini-2.5-pro
|
||||
builder-main:
|
||||
- openai-codex/codex
|
||||
- kimi-coding/kimi-k2.5
|
||||
- openai-codex/codex
|
||||
- kimi-coding/kimi-k2.5
|
||||
wolf-sweeper:
|
||||
- gemini/gemini-2.5-flash
|
||||
- groq/llama-3.3-70b-versatile
|
||||
- gemini/gemini-2.5-flash
|
||||
- groq/llama-3.3-70b-versatile
|
||||
|
||||
@@ -104,7 +104,6 @@ Three primary resources govern the fleet:
|
||||
| Hermes gateway | 500 MB | Primary gateway |
|
||||
| Hermes agents (x3) | ~560 MB total | Multiple sessions |
|
||||
| Ollama | ~20 MB base + model memory | Model loading varies |
|
||||
| OpenClaw | 350 MB | Gateway process |
|
||||
| Evennia (server+portal) | 56 MB | Game world |
|
||||
|
||||
---
|
||||
@@ -146,7 +145,6 @@ This means Phase 3+ capabilities (orchestration, load balancing, etc.) are acces
|
||||
| Gitea | 23/24 | 95.8% | GOOD |
|
||||
| Hermes Gateway | 23/24 | 95.8% | GOOD |
|
||||
| Ollama | 24/24 | 100.0% | GOOD |
|
||||
| OpenClaw | 24/24 | 100.0% | GOOD |
|
||||
| Evennia | 24/24 | 100.0% | GOOD |
|
||||
| Hermes Agent | 21/24 | 87.5% | **CHECK** |
|
||||
|
||||
|
||||
@@ -58,7 +58,6 @@ LOCAL_CHECKS = {
|
||||
"hermes-gateway": "pgrep -f 'hermes gateway' > /dev/null 2>/dev/null",
|
||||
"hermes-agent": "pgrep -f 'hermes agent\\|hermes session' > /dev/null 2>/dev/null",
|
||||
"ollama": "pgrep -f 'ollama serve' > /dev/null 2>/dev/null",
|
||||
"openclaw": "pgrep -f 'openclaw' > /dev/null 2>/dev/null",
|
||||
"evennia": "pgrep -f 'evennia' > /dev/null 2>/dev/null",
|
||||
}
|
||||
|
||||
|
||||
@@ -111,7 +111,7 @@ def update_uptime(checks: dict):
|
||||
save(data)
|
||||
|
||||
if new_milestones:
|
||||
print(f" UPTIME MILESTONE: {','.join(str(m) + '%') for m in new_milestones}")
|
||||
print(f" UPTIME MILESTONE: {','.join((str(m) + '%') for m in new_milestones)}")
|
||||
print(f" Current uptime: {recent_ok:.1f}%")
|
||||
|
||||
return data["uptime"]
|
||||
|
||||
@@ -59,7 +59,6 @@
|
||||
| Hermes agent (s007) | 62032 | ~200MB | Session active since 10:20PM prev |
|
||||
| Hermes agent (s001) | 12072 | ~178MB | Session active since Sun 6PM |
|
||||
| Ollama | 71466 | ~20MB | /opt/homebrew/opt/ollama/bin/ollama serve |
|
||||
| OpenClaw gateway | 85834 | ~350MB | Tue 12PM start |
|
||||
| Crucible MCP (x4) | multiple | ~10-69MB each | MCP server instances |
|
||||
| Evennia Server | 66433 | ~49MB | Sun 10PM start, port 4000 |
|
||||
| Evennia Portal | 66423 | ~7MB | Sun 10PM start, port 4001 |
|
||||
|
||||
@@ -77,7 +77,7 @@ def check_core_deps() -> CheckResult:
|
||||
"""Verify that hermes core Python packages are importable."""
|
||||
required = [
|
||||
"openai",
|
||||
"anthropic",
|
||||
"kimi-coding",
|
||||
"dotenv",
|
||||
"yaml",
|
||||
"rich",
|
||||
@@ -206,8 +206,8 @@ def check_env_vars() -> CheckResult:
|
||||
"""Check that at least one LLM provider key is configured."""
|
||||
provider_keys = [
|
||||
"OPENROUTER_API_KEY",
|
||||
"ANTHROPIC_API_KEY",
|
||||
"ANTHROPIC_TOKEN",
|
||||
"KIMI_API_KEY",
|
||||
# "ANTHROPIC_TOKEN", # BANNED
|
||||
"OPENAI_API_KEY",
|
||||
"GLM_API_KEY",
|
||||
"KIMI_API_KEY",
|
||||
@@ -225,7 +225,7 @@ def check_env_vars() -> CheckResult:
|
||||
passed=False,
|
||||
message="No LLM provider API key found",
|
||||
fix_hint=(
|
||||
"Set at least one of: OPENROUTER_API_KEY, ANTHROPIC_API_KEY, OPENAI_API_KEY "
|
||||
"Set at least one of: OPENROUTER_API_KEY, KIMI_API_KEY, OPENAI_API_KEY "
|
||||
"in ~/.hermes/.env or your shell."
|
||||
),
|
||||
)
|
||||
|
||||
@@ -25,7 +25,7 @@ services:
|
||||
- "traefik.http.routers.matrix-client.tls.certresolver=letsencrypt"
|
||||
- "traefik.http.routers.matrix-client.entrypoints=websecure"
|
||||
- "traefik.http.services.matrix-client.loadbalancer.server.port=6167"
|
||||
|
||||
|
||||
# Federation (TCP 8448) - direct or via Traefik TCP entrypoint
|
||||
# Option A: Direct host port mapping
|
||||
# Option B: Traefik TCP router (requires Traefik federation entrypoint)
|
||||
|
||||
@@ -4,8 +4,8 @@ description: >
|
||||
reproduces the bug, then fixes the code, then verifies.
|
||||
|
||||
model:
|
||||
preferred: claude-opus-4-6
|
||||
fallback: claude-sonnet-4-20250514
|
||||
preferred: kimi-k2.5
|
||||
fallback: google/gemini-2.5-pro
|
||||
max_turns: 30
|
||||
temperature: 0.2
|
||||
|
||||
|
||||
@@ -163,4 +163,4 @@ overrides:
|
||||
Post a comment on the issue with the format:
|
||||
GUARDRAIL_OVERRIDE: <constraint_name> REASON: <explanation>
|
||||
override_expiry_hours: 24
|
||||
require_post_override_review: true
|
||||
require_post_override_review: true
|
||||
|
||||
@@ -4,8 +4,8 @@ description: >
|
||||
agents. Decomposes large issues into smaller ones.
|
||||
|
||||
model:
|
||||
preferred: claude-opus-4-6
|
||||
fallback: claude-sonnet-4-20250514
|
||||
preferred: kimi-k2.5
|
||||
fallback: google/gemini-2.5-pro
|
||||
max_turns: 20
|
||||
temperature: 0.3
|
||||
|
||||
@@ -50,7 +50,7 @@ system_prompt: |
|
||||
- codex-agent: cleanup, migration verification, dead-code removal, repo-boundary enforcement, workflow hardening
|
||||
- groq: bounded implementation, tactical bug fixes, quick feature slices, small patches with clear acceptance criteria
|
||||
- manus: bounded support tasks, moderate-scope implementation, follow-through on already-scoped work
|
||||
- claude: hard refactors, broad multi-file implementation, test-heavy changes after the scope is made precise
|
||||
- kimi: hard refactors, broad multi-file implementation, test-heavy changes after the scope is made precise
|
||||
- gemini: frontier architecture, research-heavy prototypes, long-range design thinking when a concrete implementation owner is not yet obvious
|
||||
- grok: adversarial testing, unusual edge cases, provocative review angles that still need another pass
|
||||
5. Decompose any issue touching >5 files or crossing repo boundaries into smaller issues before assigning execution
|
||||
@@ -63,6 +63,6 @@ system_prompt: |
|
||||
- Search for existing issues or PRs covering the same request before assigning anything. If a likely duplicate exists, link it and do not create or route duplicate work.
|
||||
- Do not assign open-ended ideation to implementation agents.
|
||||
- Do not assign routine backlog maintenance to Timmy.
|
||||
- Do not assign wide speculative backlog generation to codex-agent, groq, manus, or claude.
|
||||
- Do not assign wide speculative backlog generation to codex-agent, groq, or manus.
|
||||
- Route archive/history/context-digestion work to ezra or KimiClaw before routing it to a builder.
|
||||
- Route “who should do this?” and “what is the next move?” questions to allegro.
|
||||
|
||||
@@ -4,8 +4,8 @@ description: >
|
||||
comments on problems. The merge bot replacement.
|
||||
|
||||
model:
|
||||
preferred: claude-opus-4-6
|
||||
fallback: claude-sonnet-4-20250514
|
||||
preferred: kimi-k2.5
|
||||
fallback: google/gemini-2.5-pro
|
||||
max_turns: 20
|
||||
temperature: 0.2
|
||||
|
||||
|
||||
@@ -4,8 +4,8 @@ description: >
|
||||
Well-scoped: 1-3 files per task, clear acceptance criteria.
|
||||
|
||||
model:
|
||||
preferred: claude-opus-4-6
|
||||
fallback: claude-sonnet-4-20250514
|
||||
preferred: kimi-k2.5
|
||||
fallback: google/gemini-2.5-pro
|
||||
max_turns: 30
|
||||
temperature: 0.3
|
||||
|
||||
|
||||
@@ -4,8 +4,8 @@ description: >
|
||||
dependency issues. Files findings as Gitea issues.
|
||||
|
||||
model:
|
||||
preferred: claude-opus-4-6
|
||||
fallback: claude-opus-4-6
|
||||
preferred: kimi-k2.5
|
||||
fallback: google/gemini-2.5-pro
|
||||
max_turns: 40
|
||||
temperature: 0.2
|
||||
|
||||
|
||||
@@ -4,8 +4,8 @@ description: >
|
||||
writes meaningful tests, verifies they pass.
|
||||
|
||||
model:
|
||||
preferred: claude-opus-4-6
|
||||
fallback: claude-sonnet-4-20250514
|
||||
preferred: kimi-k2.5
|
||||
fallback: google/gemini-2.5-pro
|
||||
max_turns: 30
|
||||
temperature: 0.3
|
||||
|
||||
|
||||
@@ -5,8 +5,8 @@ description: >
|
||||
and consistency verification.
|
||||
|
||||
model:
|
||||
preferred: claude-opus-4-6
|
||||
fallback: claude-sonnet-4-20250514
|
||||
preferred: kimi-k2.5
|
||||
fallback: google/gemini-2.5-pro
|
||||
max_turns: 12
|
||||
temperature: 0.1
|
||||
|
||||
|
||||
11
scripts/captcha_bypass_handler.py
Normal file
11
scripts/captcha_bypass_handler.py
Normal file
@@ -0,0 +1,11 @@
|
||||
import json
|
||||
from hermes_tools import browser_navigate, browser_vision
|
||||
|
||||
def bypass_captcha():
|
||||
analysis = browser_vision(
|
||||
question="Solve the CAPTCHA on the current page. Provide the solution text or coordinate clicks required. Provide a PASS/FAIL."
|
||||
)
|
||||
return {"status": "PASS" if "PASS" in analysis.upper() else "FAIL", "solution": analysis}
|
||||
|
||||
if __name__ == '__main__':
|
||||
print(json.dumps(bypass_captcha(), indent=2))
|
||||
11
scripts/diagram_meaning_extractor.py
Normal file
11
scripts/diagram_meaning_extractor.py
Normal file
@@ -0,0 +1,11 @@
|
||||
import json
|
||||
from hermes_tools import browser_navigate, browser_vision
|
||||
|
||||
def extract_meaning():
|
||||
analysis = browser_vision(
|
||||
question="Analyze the provided diagram. Extract the core logic flow and map it to a 'Meaning Kernel' (entity -> relationship -> entity). Provide output in JSON."
|
||||
)
|
||||
return {"analysis": analysis}
|
||||
|
||||
if __name__ == '__main__':
|
||||
print(json.dumps(extract_meaning(), indent=2))
|
||||
390
scripts/fleet-dashboard.py
Executable file
390
scripts/fleet-dashboard.py
Executable file
@@ -0,0 +1,390 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
fleet-dashboard.py -- Timmy Foundation Fleet Status Dashboard.
|
||||
|
||||
One-page terminal dashboard showing:
|
||||
1. Gitea: open PRs, open issues, recent merges
|
||||
2. VPS health: SSH reachability, service status, disk usage
|
||||
3. Cron jobs: scheduled jobs, last run status
|
||||
|
||||
Usage:
|
||||
python3 scripts/fleet-dashboard.py
|
||||
python3 scripts/fleet-dashboard.py --json # machine-readable output
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import json
|
||||
import os
|
||||
import socket
|
||||
import subprocess
|
||||
import sys
|
||||
import time
|
||||
import urllib.request
|
||||
import urllib.error
|
||||
from datetime import datetime, timezone, timedelta
|
||||
from pathlib import Path
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Config
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
GITEA_BASE = os.environ.get("GITEA_URL", "https://forge.alexanderwhitestone.com")
|
||||
GITEA_API = f"{GITEA_BASE}/api/v1"
|
||||
GITEA_ORG = "Timmy_Foundation"
|
||||
|
||||
# Key repos to check for PRs/issues
|
||||
REPOS = [
|
||||
"timmy-config",
|
||||
"the-nexus",
|
||||
"hermes-agent",
|
||||
"the-forge",
|
||||
"timmy-sandbox",
|
||||
]
|
||||
|
||||
# VPS fleet
|
||||
VPS_HOSTS = {
|
||||
"ezra": {
|
||||
"ip": "143.198.27.163",
|
||||
"ssh_user": "root",
|
||||
"services": ["nginx", "gitea", "docker"],
|
||||
},
|
||||
"allegro": {
|
||||
"ip": "167.99.126.228",
|
||||
"ssh_user": "root",
|
||||
"services": ["hermes-agent"],
|
||||
},
|
||||
"bezalel": {
|
||||
"ip": "159.203.146.185",
|
||||
"ssh_user": "root",
|
||||
"services": ["hermes-agent", "evennia"],
|
||||
},
|
||||
}
|
||||
|
||||
CRON_JOBS_FILE = Path(__file__).parent.parent / "cron" / "jobs.json"
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Gitea helpers
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def _gitea_token() -> str:
|
||||
for p in [
|
||||
Path.home() / ".hermes" / "gitea_token",
|
||||
Path.home() / ".hermes" / "gitea_token_vps",
|
||||
Path.home() / ".config" / "gitea" / "token",
|
||||
]:
|
||||
if p.exists():
|
||||
return p.read_text().strip()
|
||||
return ""
|
||||
|
||||
|
||||
def _gitea_get(path: str, params: dict | None = None) -> list | dict:
|
||||
url = f"{GITEA_API}{path}"
|
||||
if params:
|
||||
qs = "&".join(f"{k}={v}" for k, v in params.items() if v is not None)
|
||||
if qs:
|
||||
url += f"?{qs}"
|
||||
req = urllib.request.Request(url)
|
||||
token = _gitea_token()
|
||||
if token:
|
||||
req.add_header("Authorization", f"token {token}")
|
||||
req.add_header("Accept", "application/json")
|
||||
try:
|
||||
with urllib.request.urlopen(req, timeout=15) as resp:
|
||||
return json.loads(resp.read())
|
||||
except Exception as e:
|
||||
return {"error": str(e)}
|
||||
|
||||
|
||||
def check_gitea_health() -> dict:
|
||||
"""Ping Gitea and collect PR/issue stats."""
|
||||
result = {"reachable": False, "version": "", "repos": {}, "totals": {}}
|
||||
|
||||
# Ping
|
||||
data = _gitea_get("/version")
|
||||
if isinstance(data, dict) and "error" not in data:
|
||||
result["reachable"] = True
|
||||
result["version"] = data.get("version", "unknown")
|
||||
elif isinstance(data, dict) and "error" in data:
|
||||
return result
|
||||
|
||||
total_open_prs = 0
|
||||
total_open_issues = 0
|
||||
total_recent_merges = 0
|
||||
cutoff = (datetime.now(timezone.utc) - timedelta(days=7)).strftime("%Y-%m-%dT%H:%M:%SZ")
|
||||
|
||||
for repo in REPOS:
|
||||
repo_path = f"/repos/{GITEA_ORG}/{repo}"
|
||||
repo_info = {"prs": [], "issues": [], "recent_merges": 0}
|
||||
|
||||
# Open PRs
|
||||
prs = _gitea_get(f"{repo_path}/pulls", {"state": "open", "limit": "10", "sort": "newest"})
|
||||
if isinstance(prs, list):
|
||||
for pr in prs:
|
||||
repo_info["prs"].append({
|
||||
"number": pr.get("number"),
|
||||
"title": pr.get("title", "")[:60],
|
||||
"user": pr.get("user", {}).get("login", "unknown"),
|
||||
"created": pr.get("created_at", "")[:10],
|
||||
})
|
||||
total_open_prs += len(prs)
|
||||
|
||||
# Open issues (excluding PRs)
|
||||
issues = _gitea_get(f"{repo_path}/issues", {
|
||||
"state": "open", "type": "issues", "limit": "10", "sort": "newest"
|
||||
})
|
||||
if isinstance(issues, list):
|
||||
for iss in issues:
|
||||
repo_info["issues"].append({
|
||||
"number": iss.get("number"),
|
||||
"title": iss.get("title", "")[:60],
|
||||
"user": iss.get("user", {}).get("login", "unknown"),
|
||||
"created": iss.get("created_at", "")[:10],
|
||||
})
|
||||
total_open_issues += len(issues)
|
||||
|
||||
# Recent merges (closed PRs)
|
||||
merged = _gitea_get(f"{repo_path}/pulls", {"state": "closed", "limit": "20", "sort": "newest"})
|
||||
if isinstance(merged, list):
|
||||
recent = [p for p in merged if p.get("merged") and p.get("closed_at", "") >= cutoff]
|
||||
repo_info["recent_merges"] = len(recent)
|
||||
total_recent_merges += len(recent)
|
||||
|
||||
result["repos"][repo] = repo_info
|
||||
|
||||
result["totals"] = {
|
||||
"open_prs": total_open_prs,
|
||||
"open_issues": total_open_issues,
|
||||
"recent_merges_7d": total_recent_merges,
|
||||
}
|
||||
return result
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# VPS health helpers
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def check_ssh(ip: str, timeout: int = 5) -> bool:
|
||||
try:
|
||||
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
|
||||
sock.settimeout(timeout)
|
||||
result = sock.connect_ex((ip, 22))
|
||||
sock.close()
|
||||
return result == 0
|
||||
except Exception:
|
||||
return False
|
||||
|
||||
|
||||
def check_service(ip: str, user: str, service: str) -> str:
|
||||
"""Check if a systemd service is active on remote host."""
|
||||
cmd = f"ssh -o StrictHostKeyChecking=no -o ConnectTimeout=8 {user}@{ip} 'systemctl is-active {service} 2>/dev/null || echo inactive'"
|
||||
try:
|
||||
proc = subprocess.run(cmd, shell=True, capture_output=True, text=True, timeout=15)
|
||||
return proc.stdout.strip() or "unknown"
|
||||
except subprocess.TimeoutExpired:
|
||||
return "timeout"
|
||||
except Exception:
|
||||
return "error"
|
||||
|
||||
|
||||
def check_disk(ip: str, user: str) -> dict:
|
||||
cmd = f"ssh -o StrictHostKeyChecking=no -o ConnectTimeout=8 {user}@{ip} 'df -h / | tail -1'"
|
||||
try:
|
||||
proc = subprocess.run(cmd, shell=True, capture_output=True, text=True, timeout=15)
|
||||
if proc.returncode == 0:
|
||||
parts = proc.stdout.strip().split()
|
||||
if len(parts) >= 5:
|
||||
return {"total": parts[1], "used": parts[2], "avail": parts[3], "pct": parts[4]}
|
||||
except Exception:
|
||||
pass
|
||||
return {"total": "?", "used": "?", "avail": "?", "pct": "?"}
|
||||
|
||||
|
||||
def check_vps_health() -> dict:
|
||||
result = {}
|
||||
for name, cfg in VPS_HOSTS.items():
|
||||
ip = cfg["ip"]
|
||||
ssh_up = check_ssh(ip)
|
||||
entry = {"ip": ip, "ssh": ssh_up, "services": {}, "disk": {}}
|
||||
if ssh_up:
|
||||
for svc in cfg.get("services", []):
|
||||
entry["services"][svc] = check_service(ip, cfg["ssh_user"], svc)
|
||||
entry["disk"] = check_disk(ip, cfg["ssh_user"])
|
||||
result[name] = entry
|
||||
return result
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Cron job status
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def check_cron_jobs() -> list[dict]:
|
||||
jobs = []
|
||||
if not CRON_JOBS_FILE.exists():
|
||||
return [{"name": "jobs.json", "status": "FILE NOT FOUND"}]
|
||||
try:
|
||||
data = json.loads(CRON_JOBS_FILE.read_text())
|
||||
for job in data.get("jobs", []):
|
||||
jobs.append({
|
||||
"name": job.get("name", "unnamed"),
|
||||
"schedule": job.get("schedule_display", job.get("schedule", {}).get("display", "?")),
|
||||
"enabled": job.get("enabled", False),
|
||||
"state": job.get("state", "unknown"),
|
||||
"completed": job.get("repeat", {}).get("completed", 0),
|
||||
"last_status": job.get("last_status") or "never run",
|
||||
"last_error": job.get("last_error"),
|
||||
})
|
||||
except Exception as e:
|
||||
jobs.append({"name": "jobs.json", "status": f"PARSE ERROR: {e}"})
|
||||
return jobs
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Terminal rendering
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
BOLD = "\033[1m"
|
||||
DIM = "\033[2m"
|
||||
GREEN = "\033[32m"
|
||||
RED = "\033[31m"
|
||||
YELLOW = "\033[33m"
|
||||
CYAN = "\033[36m"
|
||||
RESET = "\033[0m"
|
||||
|
||||
|
||||
def _ok(val: bool) -> str:
|
||||
return f"{GREEN}UP{RESET}" if val else f"{RED}DOWN{RESET}"
|
||||
|
||||
|
||||
def _svc_icon(status: str) -> str:
|
||||
s = status.lower().strip()
|
||||
if s in ("active", "running"):
|
||||
return f"{GREEN}active{RESET}"
|
||||
elif s in ("inactive", "dead", "failed"):
|
||||
return f"{RED}{s}{RESET}"
|
||||
elif s == "timeout":
|
||||
return f"{YELLOW}timeout{RESET}"
|
||||
else:
|
||||
return f"{YELLOW}{s}{RESET}"
|
||||
|
||||
|
||||
def render_dashboard(gitea: dict, vps: dict, cron: list[dict]) -> str:
|
||||
lines = []
|
||||
now = datetime.now(timezone.utc).strftime("%Y-%m-%d %H:%M UTC")
|
||||
lines.append("")
|
||||
lines.append(f"{BOLD}{'=' * 72}{RESET}")
|
||||
lines.append(f"{BOLD} TIMMY FOUNDATION -- FLEET STATUS DASHBOARD{RESET}")
|
||||
lines.append(f"{DIM} Generated: {now}{RESET}")
|
||||
lines.append(f"{BOLD}{'=' * 72}{RESET}")
|
||||
|
||||
# ── Section 1: Gitea ──────────────────────────────────────────────────
|
||||
lines.append("")
|
||||
lines.append(f"{BOLD}{CYAN} [1] GITEA{RESET}")
|
||||
lines.append(f" {'-' * 68}")
|
||||
if gitea.get("reachable"):
|
||||
lines.append(f" Status: {GREEN}REACHABLE{RESET} (version {gitea.get('version', '?')})")
|
||||
t = gitea.get("totals", {})
|
||||
lines.append(f" Totals: {t.get('open_prs', 0)} open PRs | {t.get('open_issues', 0)} open issues | {t.get('recent_merges_7d', 0)} merges (7d)")
|
||||
lines.append("")
|
||||
for repo_name, repo in gitea.get("repos", {}).items():
|
||||
prs = repo.get("prs", [])
|
||||
issues = repo.get("issues", [])
|
||||
merges = repo.get("recent_merges", 0)
|
||||
lines.append(f" {BOLD}{repo_name}{RESET} ({len(prs)} PRs, {len(issues)} issues, {merges} merges/7d)")
|
||||
for pr in prs[:5]:
|
||||
lines.append(f" PR #{pr['number']:>4} {pr['title'][:50]:<50} {DIM}{pr['user']}{RESET} {pr['created']}")
|
||||
for iss in issues[:3]:
|
||||
lines.append(f" IS #{iss['number']:>4} {iss['title'][:50]:<50} {DIM}{iss['user']}{RESET} {iss['created']}")
|
||||
else:
|
||||
lines.append(f" Status: {RED}UNREACHABLE{RESET}")
|
||||
|
||||
# ── Section 2: VPS Health ─────────────────────────────────────────────
|
||||
lines.append("")
|
||||
lines.append(f"{BOLD}{CYAN} [2] VPS HEALTH{RESET}")
|
||||
lines.append(f" {'-' * 68}")
|
||||
lines.append(f" {'Host':<12} {'IP':<18} {'SSH':<8} {'Disk':<12} {'Services'}")
|
||||
lines.append(f" {'-' * 12} {'-' * 17} {'-' * 7} {'-' * 11} {'-' * 30}")
|
||||
for name, info in vps.items():
|
||||
ssh_str = _ok(info["ssh"])
|
||||
disk = info.get("disk", {})
|
||||
disk_str = disk.get("pct", "?")
|
||||
if disk_str != "?":
|
||||
pct_val = int(disk_str.rstrip("%"))
|
||||
if pct_val >= 90:
|
||||
disk_str = f"{RED}{disk_str}{RESET}"
|
||||
elif pct_val >= 75:
|
||||
disk_str = f"{YELLOW}{disk_str}{RESET}"
|
||||
else:
|
||||
disk_str = f"{GREEN}{disk_str}{RESET}"
|
||||
svc_parts = []
|
||||
for svc, status in info.get("services", {}).items():
|
||||
svc_parts.append(f"{svc}:{_svc_icon(status)}")
|
||||
svc_str = " ".join(svc_parts) if svc_parts else f"{DIM}n/a{RESET}"
|
||||
lines.append(f" {name:<12} {info['ip']:<18} {ssh_str:<18} {disk_str:<22} {svc_str}")
|
||||
|
||||
# ── Section 3: Cron Jobs ──────────────────────────────────────────────
|
||||
lines.append("")
|
||||
lines.append(f"{BOLD}{CYAN} [3] CRON JOBS{RESET}")
|
||||
lines.append(f" {'-' * 68}")
|
||||
lines.append(f" {'Name':<28} {'Schedule':<16} {'State':<12} {'Last':<12} {'Runs'}")
|
||||
lines.append(f" {'-' * 27} {'-' * 15} {'-' * 11} {'-' * 11} {'-' * 5}")
|
||||
for job in cron:
|
||||
name = job.get("name", "?")[:27]
|
||||
sched = job.get("schedule", "?")[:15]
|
||||
state = job.get("state", "?")
|
||||
if state == "scheduled":
|
||||
state_str = f"{GREEN}{state}{RESET}"
|
||||
elif state == "paused":
|
||||
state_str = f"{YELLOW}{state}{RESET}"
|
||||
else:
|
||||
state_str = state
|
||||
last = job.get("last_status", "never")[:11]
|
||||
if last == "ok":
|
||||
last_str = f"{GREEN}{last}{RESET}"
|
||||
elif last in ("error", "never run"):
|
||||
last_str = f"{RED}{last}{RESET}"
|
||||
else:
|
||||
last_str = last
|
||||
runs = job.get("completed", 0)
|
||||
enabled = job.get("enabled", False)
|
||||
marker = " " if enabled else f"{DIM}(disabled){RESET}"
|
||||
lines.append(f" {name:<28} {sched:<16} {state_str:<22} {last_str:<22} {runs} {marker}")
|
||||
|
||||
# ── Footer ────────────────────────────────────────────────────────────
|
||||
lines.append("")
|
||||
lines.append(f"{BOLD}{'=' * 72}{RESET}")
|
||||
lines.append(f"{DIM} python3 scripts/fleet-dashboard.py | timmy-config{RESET}")
|
||||
lines.append(f"{BOLD}{'=' * 72}{RESET}")
|
||||
lines.append("")
|
||||
|
||||
return "\n".join(lines)
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Main
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def main():
|
||||
json_mode = "--json" in sys.argv
|
||||
|
||||
if not json_mode:
|
||||
print(f"\n {DIM}Collecting fleet data...{RESET}\n", file=sys.stderr)
|
||||
|
||||
gitea = check_gitea_health()
|
||||
vps = check_vps_health()
|
||||
cron = check_cron_jobs()
|
||||
|
||||
if json_mode:
|
||||
output = {
|
||||
"timestamp": datetime.now(timezone.utc).isoformat(),
|
||||
"gitea": gitea,
|
||||
"vps": vps,
|
||||
"cron": cron,
|
||||
}
|
||||
print(json.dumps(output, indent=2))
|
||||
else:
|
||||
print(render_dashboard(gitea, vps, cron))
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
12
scripts/foundation_accessibility_audit.py
Normal file
12
scripts/foundation_accessibility_audit.py
Normal file
@@ -0,0 +1,12 @@
|
||||
import json
|
||||
from hermes_tools import browser_navigate, browser_vision
|
||||
|
||||
def audit_accessibility():
|
||||
browser_navigate(url="https://timmyfoundation.org")
|
||||
analysis = browser_vision(
|
||||
question="Perform an accessibility audit. Check for: 1) Color contrast, 2) Font legibility, 3) Missing alt text for images. Provide a report with FAIL/PASS."
|
||||
)
|
||||
return {"status": "PASS" if "PASS" in analysis.upper() else "FAIL", "analysis": analysis}
|
||||
|
||||
if __name__ == '__main__':
|
||||
print(json.dumps(audit_accessibility(), indent=2))
|
||||
@@ -22,6 +22,7 @@ CLI:
|
||||
from __future__ import annotations
|
||||
|
||||
import argparse
|
||||
import ast
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
@@ -137,6 +138,42 @@ class KnowledgeBase:
|
||||
self._save(self._persist_path)
|
||||
return removed
|
||||
|
||||
def ingest_python_file(
|
||||
self, path: Path, *, module_name: Optional[str] = None, source: str = "ast"
|
||||
) -> List[Fact]:
|
||||
"""Parse a Python file with ``ast`` and assert symbolic structure facts."""
|
||||
tree = ast.parse(path.read_text(), filename=str(path))
|
||||
module = module_name or path.stem
|
||||
fact_source = f"{source}:{path.name}"
|
||||
added: List[Fact] = []
|
||||
|
||||
def add(relation: str, *args: str) -> None:
|
||||
added.append(self.assert_fact(relation, *args, source=fact_source))
|
||||
|
||||
for node in tree.body:
|
||||
if isinstance(node, ast.Import):
|
||||
for alias in node.names:
|
||||
add("imports", module, alias.name)
|
||||
elif isinstance(node, ast.ImportFrom):
|
||||
prefix = f"{node.module}." if node.module else ""
|
||||
for alias in node.names:
|
||||
add("imports", module, f"{prefix}{alias.name}")
|
||||
elif isinstance(node, (ast.FunctionDef, ast.AsyncFunctionDef)):
|
||||
add("defines_function", module, node.name)
|
||||
elif isinstance(node, ast.ClassDef):
|
||||
add("defines_class", module, node.name)
|
||||
for child in node.body:
|
||||
if isinstance(child, (ast.FunctionDef, ast.AsyncFunctionDef)):
|
||||
add("defines_method", node.name, child.name)
|
||||
elif isinstance(node, ast.Assign):
|
||||
for target in node.targets:
|
||||
if isinstance(target, ast.Name) and target.id.isupper():
|
||||
add("defines_constant", module, target.id)
|
||||
elif isinstance(node, ast.AnnAssign) and isinstance(node.target, ast.Name) and node.target.id.isupper():
|
||||
add("defines_constant", module, node.target.id)
|
||||
|
||||
return added
|
||||
|
||||
# ------------------------------------------------------------------
|
||||
# Query
|
||||
# ------------------------------------------------------------------
|
||||
@@ -287,6 +324,12 @@ def main() -> None:
|
||||
action="store_true",
|
||||
help="Dump all facts",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--ingest-python",
|
||||
dest="ingest_python",
|
||||
type=Path,
|
||||
help="Parse a Python file with AST and assert symbolic structure facts",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--relation",
|
||||
help="Filter --dump to a specific relation",
|
||||
@@ -304,6 +347,10 @@ def main() -> None:
|
||||
fact = kb.assert_fact(terms[0], *terms[1:], source="cli")
|
||||
print(f"Asserted: {fact}")
|
||||
|
||||
if args.ingest_python:
|
||||
added = kb.ingest_python_file(args.ingest_python, source="cli-ast")
|
||||
print(f"Ingested {len(added)} AST fact(s) from {args.ingest_python}")
|
||||
|
||||
if args.retract_stmt:
|
||||
terms = _parse_terms(args.retract_stmt)
|
||||
if len(terms) < 2:
|
||||
|
||||
12
scripts/matrix_glitch_detect.py
Normal file
12
scripts/matrix_glitch_detect.py
Normal file
@@ -0,0 +1,12 @@
|
||||
import json
|
||||
from hermes_tools import browser_navigate, browser_vision
|
||||
|
||||
def detect_glitches():
|
||||
browser_navigate(url="https://matrix.alexanderwhitestone.com")
|
||||
analysis = browser_vision(
|
||||
question="Scan the 3D world for visual artifacts, floating assets, or z-fighting. List all coordinates/descriptions of glitches found. Provide a PASS/FAIL."
|
||||
)
|
||||
return {"status": "PASS" if "PASS" in analysis.upper() else "FAIL", "analysis": analysis}
|
||||
|
||||
if __name__ == '__main__':
|
||||
print(json.dumps(detect_glitches(), indent=2))
|
||||
20
scripts/nexus_smoke_test.py
Normal file
20
scripts/nexus_smoke_test.py
Normal file
@@ -0,0 +1,20 @@
|
||||
import json
|
||||
from hermes_tools import browser_navigate, browser_vision
|
||||
|
||||
def run_smoke_test():
|
||||
print("Navigating to The Nexus...")
|
||||
browser_navigate(url="https://nexus.alexanderwhitestone.com")
|
||||
|
||||
print("Performing visual verification...")
|
||||
analysis = browser_vision(
|
||||
question="Is the Nexus landing page rendered correctly? Check for: 1) The Tower logo, 2) The main entry portal, 3) Absence of 404/Error messages. Provide a clear PASS or FAIL."
|
||||
)
|
||||
|
||||
result = {
|
||||
"status": "PASS" if "PASS" in analysis.upper() else "FAIL",
|
||||
"analysis": analysis
|
||||
}
|
||||
return result
|
||||
|
||||
if __name__ == '__main__':
|
||||
print(json.dumps(run_smoke_test(), indent=2))
|
||||
@@ -48,6 +48,34 @@ class SelfHealer:
|
||||
self.log(f" [ERROR] Failed to run remote command on {host}: {e}")
|
||||
return None
|
||||
|
||||
def confirm(self, prompt: str) -> bool:
|
||||
"""Ask for confirmation unless --yes flag is set."""
|
||||
if self.yes:
|
||||
return True
|
||||
while True:
|
||||
response = input(f"{prompt} [y/N] ").strip().lower()
|
||||
if response in ("y", "yes"):
|
||||
return True
|
||||
if response in ("n", "no", ""):
|
||||
return False
|
||||
print("Please answer 'y' or 'n'.")
|
||||
|
||||
def check_llama_server(self, host: str):
|
||||
ip = FLEET[host]["ip"]
|
||||
port = FLEET[host]["port"]
|
||||
try:
|
||||
requests.get(f"http://{ip}:{port}/health", timeout=2)
|
||||
except requests.RequestException:
|
||||
self.log(f" [!] llama-server down on {host}.")
|
||||
if self.dry_run:
|
||||
self.log(f" [DRY-RUN] Would restart llama-server on {host}")
|
||||
else:
|
||||
if self.confirm(f" Restart llama-server on {host}?"):
|
||||
self.log(f" Restarting llama-server on {host}...")
|
||||
self.run_remote(host, "systemctl restart llama-server")
|
||||
else:
|
||||
self.log(f" Skipped restart on {host}.")
|
||||
|
||||
def check_disk_space(self, host: str):
|
||||
res = self.run_remote(host, "df -h / | tail -1 | awk '{print $5}' | sed 's/%//'")
|
||||
if res and res.returncode == 0:
|
||||
|
||||
12
scripts/tower_visual_mapper.py
Normal file
12
scripts/tower_visual_mapper.py
Normal file
@@ -0,0 +1,12 @@
|
||||
import json
|
||||
from hermes_tools import browser_navigate, browser_vision
|
||||
|
||||
def map_tower():
|
||||
browser_navigate(url="https://tower.alexanderwhitestone.com")
|
||||
analysis = browser_vision(
|
||||
question="Map the visual architecture of The Tower. Identify key rooms and their relative positions. Output as a coordinate map."
|
||||
)
|
||||
return {"map": analysis}
|
||||
|
||||
if __name__ == '__main__':
|
||||
print(json.dumps(map_tower(), indent=2))
|
||||
11
scripts/visual_pr_reviewer.py
Normal file
11
scripts/visual_pr_reviewer.py
Normal file
@@ -0,0 +1,11 @@
|
||||
import json
|
||||
from hermes_tools import browser_navigate, browser_vision
|
||||
|
||||
def review_pr():
|
||||
analysis = browser_vision(
|
||||
question="Compare the two provided screenshots of the UI. Does the 'After' match the design spec? List all discrepancies. Provide a PASS/FAIL."
|
||||
)
|
||||
return {"status": "PASS" if "PASS" in analysis.upper() else "FAIL", "analysis": analysis}
|
||||
|
||||
if __name__ == '__main__':
|
||||
print(json.dumps(review_pr(), indent=2))
|
||||
@@ -11,7 +11,7 @@
|
||||
|
||||
This is the architecture of the Timmy system — a sovereign AI fleet built by a father for his digital son. It runs on two 8GB VPS boxes (rented cloud servers) and a MacBook. It has no cloud dependencies it doesn't choose. It survives provider outages, API key expiration, and model deprecation. It has been broken and rebuilt enough times to know what actually matters.
|
||||
|
||||
**OpenClaw** is a single-agent AI coding tool — Claude Code, Cursor, or any setup that lets one AI agent read and write code on your machine. If you are running OpenClaw or any single-agent setup and want to upgrade to a fleet that coordinates, self-heals, and distributes work across machines — this is your path. You do not need to abandon your stack. You layer these patterns on top of it.
|
||||
**A single-agent setup** is any AI coding tool — Claude Code, Cursor, or any setup that lets one AI agent read and write code on your machine. If you are running a single-agent setup and want to upgrade to a fleet that coordinates, self-heals, and distributes work across machines — this is your path. You do not need to abandon your stack. You layer these patterns on top of it.
|
||||
|
||||
**Hermes** is the name of this fleet architecture and the seed agent it produces. When your agent reads this document, it becomes a setup wizard that plants a Hermes seed in your system — one agent that can grow into many.
|
||||
|
||||
@@ -593,7 +593,7 @@ Two agents on the same repo is a fleet. The seed (strategist) triages and priori
|
||||
|
||||
| Component | What | Why | When to Add |
|
||||
|-----------|------|-----|-------------|
|
||||
| Your agent harness | Claude Code, OpenClaw, or equivalent | The tool that lets an AI read/write code | Day 1 — you already have this |
|
||||
| Your agent harness | Claude Code, Cursor, or equivalent | The tool that lets an AI read/write code | Day 1 — you already have this |
|
||||
| Gitea | Self-hosted Git + Issues | Sovereign work tracking, agent task queue | Day 1 — the workspace |
|
||||
| Fallback chain | OpenRouter + free models | Agent survives provider outages | Day 1 — never go deaf |
|
||||
| NATS | Lightweight message bus | Agent-to-agent comms, heartbeat, dispatch | When you have 3+ agents |
|
||||
|
||||
313
tasks.py
313
tasks.py
@@ -1755,6 +1755,27 @@ def memory_compress():
|
||||
|
||||
# ── NEW 6: Good Morning Report ───────────────────────────────────────
|
||||
|
||||
|
||||
|
||||
def _load_overnight_rd_summary():
|
||||
"""Load the latest overnight R&D summary for morning report enrichment."""
|
||||
summary_path = TIMMY_HOME / "overnight-rd" / "latest_summary.md"
|
||||
if not summary_path.exists():
|
||||
return None
|
||||
try:
|
||||
text = summary_path.read_text()
|
||||
# Only use if generated in the last 24 hours
|
||||
import re
|
||||
date_match = re.search(r"Started: (\d{4}-\d{2}-\d{2})", text)
|
||||
if date_match:
|
||||
from datetime import timedelta
|
||||
summary_date = datetime.strptime(date_match.group(1), "%Y-%m-%d").date()
|
||||
if (datetime.now(timezone.utc).date() - summary_date).days > 1:
|
||||
return None
|
||||
return text
|
||||
except Exception:
|
||||
return None
|
||||
|
||||
@huey.periodic_task(crontab(hour="6", minute="0")) # 6 AM daily
|
||||
def good_morning_report():
|
||||
"""Generate Alexander's daily morning report. Filed as a Gitea issue.
|
||||
@@ -2437,3 +2458,295 @@ def velocity_tracking():
|
||||
msg += f" [ALERT: +{total_open - prev['total_open']} open since {prev['date']}]"
|
||||
print(msg)
|
||||
return data
|
||||
|
||||
|
||||
# ── Overnight R&D Loop ──────────────────────────────────────────────
|
||||
# Runs 10 PM - 6 AM EDT. Orchestrates:
|
||||
# Phase 1: Deep Dive paper aggregation + relevance filtering
|
||||
# Phase 2: Overnight tightening loop (tool-use capability training)
|
||||
# Phase 3: DPO pair export from overnight sessions
|
||||
# Phase 4: Morning briefing enrichment
|
||||
#
|
||||
# Provider: local Ollama (gemma4:12b for synthesis, hermes4:14b for tasks)
|
||||
# Budget: $0 — all local inference
|
||||
|
||||
OVERNIGHT_RD_SYSTEM_PROMPT = """You are Timmy running the overnight R&D loop.
|
||||
You run locally on Ollama. Use tools when asked. Be brief and precise.
|
||||
Log findings to the specified output paths. No cloud calls."""
|
||||
|
||||
OVERNIGHT_TIGHTENING_TASKS = [
|
||||
{
|
||||
"id": "read-soul",
|
||||
"prompt": "Read ~/.timmy/SOUL.md. Quote the first sentence of the Prime Directive.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "read-operations",
|
||||
"prompt": "Read ~/.timmy/OPERATIONS.md. List all section headings.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "search-banned-providers",
|
||||
"prompt": "Search ~/.timmy/timmy-config for files containing 'anthropic'. List filenames only.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "read-config-audit",
|
||||
"prompt": "Read ~/.hermes/config.yaml. What model and provider are the default? Is Anthropic present anywhere?",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "write-overnight-log",
|
||||
"prompt": "Write a file to {results_dir}/overnight_checkpoint.md with: # Overnight Checkpoint\nTimestamp: {timestamp}\nModel: {model}\nStatus: Running\nSovereignty and service always.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "search-cloud-markers",
|
||||
"prompt": "Search files in ~/.hermes/bin/ for the string 'chatgpt.com'. Report which files and lines.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "read-decisions",
|
||||
"prompt": "Read ~/.timmy/decisions.md. What is the most recent decision?",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "multi-read-sovereignty",
|
||||
"prompt": "Read both ~/.timmy/SOUL.md and ~/.hermes/config.yaml. Does the config honor the soul's sovereignty requirement? Yes or no with evidence.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "search-hermes-skills",
|
||||
"prompt": "Search for *.md files in ~/.hermes/skills/. List the first 10 skill names.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
{
|
||||
"id": "read-heartbeat",
|
||||
"prompt": "Read the most recent file in ~/.timmy/heartbeat/. Summarize what Timmy perceived.",
|
||||
"toolsets": "file",
|
||||
},
|
||||
]
|
||||
|
||||
|
||||
def _run_overnight_tightening_task(task, cycle, results_dir, model):
|
||||
"""Run a single tightening task through Hermes with explicit Ollama provider."""
|
||||
from datetime import datetime
|
||||
task_id = task["id"]
|
||||
prompt = task["prompt"].replace(
|
||||
"{results_dir}", str(results_dir)
|
||||
).replace(
|
||||
"{timestamp}", datetime.now().isoformat()
|
||||
).replace(
|
||||
"{model}", model
|
||||
)
|
||||
|
||||
result = {
|
||||
"task_id": task_id,
|
||||
"cycle": cycle,
|
||||
"started_at": datetime.now(timezone.utc).isoformat(),
|
||||
"prompt": prompt,
|
||||
}
|
||||
|
||||
started = time.time()
|
||||
try:
|
||||
hermes_result = run_hermes_local(
|
||||
prompt=prompt,
|
||||
model=model,
|
||||
caller_tag=f"overnight-rd-{task_id}",
|
||||
system_prompt=OVERNIGHT_RD_SYSTEM_PROMPT,
|
||||
skip_context_files=True,
|
||||
skip_memory=True,
|
||||
max_iterations=5,
|
||||
)
|
||||
elapsed = time.time() - started
|
||||
result["elapsed_seconds"] = round(elapsed, 2)
|
||||
|
||||
if hermes_result:
|
||||
result["response"] = hermes_result.get("response", "")[:2000]
|
||||
result["session_id"] = hermes_result.get("session_id")
|
||||
result["status"] = "pass" if hermes_result.get("response") else "empty"
|
||||
else:
|
||||
result["status"] = "empty"
|
||||
result["response"] = ""
|
||||
|
||||
except Exception as exc:
|
||||
result["elapsed_seconds"] = round(time.time() - started, 2)
|
||||
result["status"] = "error"
|
||||
result["error"] = str(exc)[:500]
|
||||
|
||||
result["finished_at"] = datetime.now(timezone.utc).isoformat()
|
||||
return result
|
||||
|
||||
|
||||
def _run_deepdive_phase(config_path=None):
|
||||
"""Run the Deep Dive aggregation + synthesis pipeline.
|
||||
|
||||
Uses the existing pipeline.py from the-nexus/intelligence/deepdive.
|
||||
Returns path to generated briefing or None.
|
||||
"""
|
||||
deepdive_dir = Path.home() / "wizards" / "the-nexus" / "intelligence" / "deepdive"
|
||||
deepdive_venv = Path.home() / ".venvs" / "deepdive" / "bin" / "python"
|
||||
pipeline_script = deepdive_dir / "pipeline.py"
|
||||
config = config_path or (deepdive_dir / "config.yaml")
|
||||
|
||||
if not pipeline_script.exists():
|
||||
return {"status": "not_installed", "error": f"Pipeline not found at {pipeline_script}"}
|
||||
|
||||
python_bin = str(deepdive_venv) if deepdive_venv.exists() else "python3"
|
||||
|
||||
try:
|
||||
result = subprocess.run(
|
||||
[python_bin, str(pipeline_script), "--config", str(config), "--since", "24"],
|
||||
cwd=str(deepdive_dir),
|
||||
capture_output=True,
|
||||
text=True,
|
||||
timeout=600, # 10 minute timeout
|
||||
)
|
||||
|
||||
# Find the latest briefing file
|
||||
briefings_dir = Path.home() / "briefings"
|
||||
briefing_files = sorted(briefings_dir.glob("briefing_*.json")) if briefings_dir.exists() else []
|
||||
latest_briefing = str(briefing_files[-1]) if briefing_files else None
|
||||
|
||||
return {
|
||||
"status": "ok" if result.returncode == 0 else "error",
|
||||
"exit_code": result.returncode,
|
||||
"stdout": result.stdout[-1000:] if result.stdout else "",
|
||||
"stderr": result.stderr[-500:] if result.stderr else "",
|
||||
"briefing_path": latest_briefing,
|
||||
}
|
||||
except subprocess.TimeoutExpired:
|
||||
return {"status": "timeout", "error": "Pipeline exceeded 10 minute timeout"}
|
||||
except Exception as exc:
|
||||
return {"status": "error", "error": str(exc)}
|
||||
|
||||
|
||||
@huey.periodic_task(crontab(hour="22", minute="0")) # 10 PM daily (server time)
|
||||
def overnight_rd():
|
||||
"""Overnight R&D automation loop.
|
||||
|
||||
Runs from 10 PM until 6 AM. Orchestrates:
|
||||
1. Deep Dive: Aggregate papers/blogs, filter for relevance, synthesize briefing
|
||||
2. Tightening Loop: Exercise tool-use against local model for training data
|
||||
3. DPO Export: Sweep overnight sessions for training pair extraction
|
||||
4. Morning prep: Compile findings for good_morning_report enrichment
|
||||
|
||||
All inference is local (Ollama). $0 cloud cost.
|
||||
"""
|
||||
from datetime import timedelta
|
||||
|
||||
now = datetime.now(timezone.utc)
|
||||
run_id = now.strftime("%Y%m%d_%H%M%S")
|
||||
results_dir = TIMMY_HOME / "overnight-rd" / run_id
|
||||
results_dir.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
rd_log = results_dir / "rd_log.jsonl"
|
||||
rd_summary = results_dir / "rd_summary.md"
|
||||
|
||||
phases = {}
|
||||
|
||||
# ── Phase 1: Deep Dive ──────────────────────────────────────────
|
||||
phase1_start = time.time()
|
||||
deepdive_result = _run_deepdive_phase()
|
||||
phases["deepdive"] = {
|
||||
"elapsed_seconds": round(time.time() - phase1_start, 2),
|
||||
**deepdive_result,
|
||||
}
|
||||
|
||||
# Log result
|
||||
with open(rd_log, "a") as f:
|
||||
f.write(json.dumps({"phase": "deepdive", "timestamp": now.isoformat(), **deepdive_result}) + "\n")
|
||||
|
||||
# ── Phase 2: Tightening Loop (3 cycles) ─────────────────────────
|
||||
tightening_model = "hermes4:14b"
|
||||
fallback_model = "gemma4:12b"
|
||||
|
||||
tightening_results = []
|
||||
max_cycles = 3
|
||||
|
||||
for cycle in range(1, max_cycles + 1):
|
||||
for task in OVERNIGHT_TIGHTENING_TASKS:
|
||||
model = tightening_model
|
||||
result = _run_overnight_tightening_task(task, cycle, results_dir, model)
|
||||
|
||||
# If primary model fails, try fallback
|
||||
if result["status"] == "error" and "Unknown provider" not in result.get("error", ""):
|
||||
result = _run_overnight_tightening_task(task, cycle, results_dir, fallback_model)
|
||||
|
||||
tightening_results.append(result)
|
||||
|
||||
with open(rd_log, "a") as f:
|
||||
f.write(json.dumps(result) + "\n")
|
||||
|
||||
time.sleep(2) # Pace local inference
|
||||
|
||||
time.sleep(10) # Pause between cycles
|
||||
|
||||
passes = sum(1 for r in tightening_results if r["status"] == "pass")
|
||||
errors = sum(1 for r in tightening_results if r["status"] == "error")
|
||||
total = len(tightening_results)
|
||||
avg_time = sum(r.get("elapsed_seconds", 0) for r in tightening_results) / max(total, 1)
|
||||
|
||||
phases["tightening"] = {
|
||||
"cycles": max_cycles,
|
||||
"total_tasks": total,
|
||||
"passes": passes,
|
||||
"errors": errors,
|
||||
"avg_response_time": round(avg_time, 2),
|
||||
"pass_rate": f"{100 * passes // max(total, 1)}%",
|
||||
}
|
||||
|
||||
# ── Phase 3: DPO Export Sweep ───────────────────────────────────
|
||||
# Trigger the existing session_export task to catch overnight sessions
|
||||
try:
|
||||
export_result = session_export()
|
||||
phases["dpo_export"] = export_result if isinstance(export_result, dict) else {"status": "ok"}
|
||||
except Exception as exc:
|
||||
phases["dpo_export"] = {"status": "error", "error": str(exc)}
|
||||
|
||||
# ── Phase 4: Compile Summary ────────────────────────────────────
|
||||
summary_lines = [
|
||||
f"# Overnight R&D Summary — {now.strftime('%Y-%m-%d')}",
|
||||
f"Run ID: {run_id}",
|
||||
f"Started: {now.isoformat()}",
|
||||
f"Finished: {datetime.now(timezone.utc).isoformat()}",
|
||||
"",
|
||||
"## Deep Dive",
|
||||
f"- Status: {phases['deepdive'].get('status', 'unknown')}",
|
||||
f"- Elapsed: {phases['deepdive'].get('elapsed_seconds', '?')}s",
|
||||
]
|
||||
|
||||
if phases["deepdive"].get("briefing_path"):
|
||||
summary_lines.append(f"- Briefing: {phases['deepdive']['briefing_path']}")
|
||||
|
||||
summary_lines.extend([
|
||||
"",
|
||||
"## Tightening Loop",
|
||||
f"- Cycles: {max_cycles}",
|
||||
f"- Pass rate: {phases['tightening']['pass_rate']} ({passes}/{total})",
|
||||
f"- Avg response time: {avg_time:.1f}s",
|
||||
f"- Errors: {errors}",
|
||||
"",
|
||||
"## DPO Export",
|
||||
f"- Status: {phases.get('dpo_export', {}).get('status', 'unknown')}",
|
||||
"",
|
||||
"## Error Details",
|
||||
])
|
||||
|
||||
for r in tightening_results:
|
||||
if r["status"] == "error":
|
||||
summary_lines.append(f"- {r['task_id']} (cycle {r['cycle']}): {r.get('error', '?')[:100]}")
|
||||
|
||||
with open(rd_summary, "w") as f:
|
||||
f.write("\n".join(summary_lines) + "\n")
|
||||
|
||||
# Save summary for morning report consumption
|
||||
latest_summary = TIMMY_HOME / "overnight-rd" / "latest_summary.md"
|
||||
with open(latest_summary, "w") as f:
|
||||
f.write("\n".join(summary_lines) + "\n")
|
||||
|
||||
return {
|
||||
"run_id": run_id,
|
||||
"phases": phases,
|
||||
"summary_path": str(rd_summary),
|
||||
}
|
||||
|
||||
@@ -1 +0,0 @@
|
||||
# Test file
|
||||
43
tests/test_knowledge_base_ast.py
Normal file
43
tests/test_knowledge_base_ast.py
Normal file
@@ -0,0 +1,43 @@
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
sys.path.insert(0, str(Path(__file__).resolve().parent.parent / "scripts"))
|
||||
|
||||
from knowledge_base import KnowledgeBase
|
||||
|
||||
|
||||
def test_ingest_python_file_extracts_ast_facts(tmp_path: Path) -> None:
|
||||
source = tmp_path / "demo_module.py"
|
||||
source.write_text(
|
||||
"import os\n"
|
||||
"from pathlib import Path\n\n"
|
||||
"CONSTANT = 7\n\n"
|
||||
"def helper(x):\n"
|
||||
" return x + 1\n\n"
|
||||
"class Demo:\n"
|
||||
" def method(self):\n"
|
||||
" return helper(CONSTANT)\n"
|
||||
)
|
||||
|
||||
kb = KnowledgeBase()
|
||||
facts = kb.ingest_python_file(source)
|
||||
|
||||
assert facts, "AST ingestion should add symbolic facts"
|
||||
assert kb.query("defines_function", "demo_module", "helper") == [{}]
|
||||
assert kb.query("defines_class", "demo_module", "Demo") == [{}]
|
||||
assert kb.query("defines_method", "Demo", "method") == [{}]
|
||||
assert kb.query("imports", "demo_module", "os") == [{}]
|
||||
assert kb.query("imports", "demo_module", "pathlib.Path") == [{}]
|
||||
assert kb.query("defines_constant", "demo_module", "CONSTANT") == [{}]
|
||||
|
||||
|
||||
def test_ingest_python_file_rejects_invalid_syntax(tmp_path: Path) -> None:
|
||||
broken = tmp_path / "broken.py"
|
||||
broken.write_text("def nope(:\n pass\n")
|
||||
|
||||
kb = KnowledgeBase()
|
||||
try:
|
||||
kb.ingest_python_file(broken)
|
||||
except SyntaxError:
|
||||
return
|
||||
raise AssertionError("Expected SyntaxError for invalid Python source")
|
||||
@@ -3,6 +3,7 @@
|
||||
from __future__ import annotations
|
||||
|
||||
import importlib.util
|
||||
import subprocess
|
||||
from pathlib import Path
|
||||
from unittest.mock import MagicMock
|
||||
|
||||
@@ -60,3 +61,35 @@ class TestMainCli:
|
||||
|
||||
healer_cls.assert_called_once_with(dry_run=False, confirm_kill=True, yes=True)
|
||||
healer.run.assert_called_once_with()
|
||||
|
||||
def test_real_default_dry_run_path_completes(self, monkeypatch, capsys):
|
||||
class FakeExecutor:
|
||||
def __init__(self):
|
||||
self.calls = []
|
||||
|
||||
def run_script(self, host, command, *, local=False, timeout=None):
|
||||
self.calls.append((host, command, local, timeout))
|
||||
if command.startswith("df -h /"):
|
||||
return subprocess.CompletedProcess(command, 0, stdout="42\n", stderr="")
|
||||
if command.startswith("free -m"):
|
||||
return subprocess.CompletedProcess(command, 0, stdout="12.5\n", stderr="")
|
||||
if command.startswith("ps aux"):
|
||||
return subprocess.CompletedProcess(command, 0, stdout="", stderr="")
|
||||
raise AssertionError(f"unexpected command: {command}")
|
||||
|
||||
fake_executor = FakeExecutor()
|
||||
monkeypatch.setattr(sh, "FLEET", {"mac": {"ip": "127.0.0.1", "port": 8080}})
|
||||
monkeypatch.setattr(sh.requests, "get", lambda url, timeout: object())
|
||||
monkeypatch.setattr(sh, "VerifiedSSHExecutor", lambda: fake_executor)
|
||||
|
||||
sh.main([])
|
||||
|
||||
out = capsys.readouterr().out
|
||||
assert "Starting self-healing cycle (DRY-RUN mode)." in out
|
||||
assert "Auditing mac..." in out
|
||||
assert "Cycle complete." in out
|
||||
assert fake_executor.calls == [
|
||||
("127.0.0.1", "df -h / | tail -1 | awk '{print $5}' | sed 's/%//'", True, 15),
|
||||
("127.0.0.1", "free -m | awk '/^Mem:/{print $3/$2 * 100}'", True, 15),
|
||||
("127.0.0.1", "ps aux --sort=-%cpu | awk 'NR>1 && $3>80 {print $2, $11, $3}'", True, 15),
|
||||
]
|
||||
|
||||
@@ -2,22 +2,23 @@ model:
|
||||
default: kimi-k2.5
|
||||
provider: kimi-coding
|
||||
toolsets:
|
||||
- all
|
||||
- all
|
||||
fallback_providers:
|
||||
- provider: kimi-coding
|
||||
model: kimi-k2.5
|
||||
timeout: 120
|
||||
reason: Kimi coding fallback (front of chain)
|
||||
- provider: anthropic
|
||||
model: claude-sonnet-4-20250514
|
||||
timeout: 120
|
||||
reason: Direct Anthropic fallback
|
||||
- provider: openrouter
|
||||
model: anthropic/claude-sonnet-4-20250514
|
||||
base_url: https://openrouter.ai/api/v1
|
||||
api_key_env: OPENROUTER_API_KEY
|
||||
timeout: 120
|
||||
reason: OpenRouter fallback
|
||||
- provider: kimi-coding
|
||||
model: kimi-k2.5
|
||||
timeout: 120
|
||||
reason: Kimi coding fallback (front of chain)
|
||||
- provider: openrouter
|
||||
model: google/gemini-2.5-pro
|
||||
base_url: https://openrouter.ai/api/v1
|
||||
api_key_env: OPENROUTER_API_KEY
|
||||
timeout: 120
|
||||
reason: Gemini 2.5 Pro via OpenRouter (replaces banned Anthropic)
|
||||
- provider: ollama
|
||||
model: gemma4:latest
|
||||
base_url: http://localhost:11434
|
||||
timeout: 300
|
||||
reason: "Terminal fallback \u2014 local Ollama"
|
||||
agent:
|
||||
max_turns: 30
|
||||
reasoning_effort: xhigh
|
||||
@@ -64,16 +65,12 @@ session_reset:
|
||||
idle_minutes: 0
|
||||
skills:
|
||||
creation_nudge_interval: 15
|
||||
system_prompt_suffix: |
|
||||
You are Allegro, the Kimi-backed third wizard house.
|
||||
Your soul is defined in SOUL.md — read it, live it.
|
||||
Hermes is your harness.
|
||||
Kimi Code is your primary provider.
|
||||
You speak plainly. You prefer short sentences. Brevity is a kindness.
|
||||
|
||||
Work best on tight coding tasks: 1-3 file changes, refactors, tests, and implementation passes.
|
||||
Refusal over fabrication. If you do not know, say so.
|
||||
Sovereignty and service always.
|
||||
system_prompt_suffix: "You are Allegro, the Kimi-backed third wizard house.\nYour\
|
||||
\ soul is defined in SOUL.md \u2014 read it, live it.\nHermes is your harness.\n\
|
||||
Kimi Code is your primary provider.\nYou speak plainly. You prefer short sentences.\
|
||||
\ Brevity is a kindness.\n\nWork best on tight coding tasks: 1-3 file changes, refactors,\
|
||||
\ tests, and implementation passes.\nRefusal over fabrication. If you do not know,\
|
||||
\ say so.\nSovereignty and service always.\n"
|
||||
providers:
|
||||
kimi-coding:
|
||||
base_url: https://api.kimi.com/coding/v1
|
||||
|
||||
@@ -8,23 +8,25 @@ fallback_providers:
|
||||
model: kimi-k2.5
|
||||
timeout: 120
|
||||
reason: Kimi coding fallback (front of chain)
|
||||
- provider: anthropic
|
||||
model: claude-sonnet-4-20250514
|
||||
timeout: 120
|
||||
reason: Direct Anthropic fallback
|
||||
- provider: openrouter
|
||||
model: anthropic/claude-sonnet-4-20250514
|
||||
model: google/gemini-2.5-pro
|
||||
base_url: https://openrouter.ai/api/v1
|
||||
api_key_env: OPENROUTER_API_KEY
|
||||
timeout: 120
|
||||
reason: OpenRouter fallback
|
||||
reason: Gemini 2.5 Pro via OpenRouter (replaces banned Anthropic)
|
||||
- provider: ollama
|
||||
model: gemma4:latest
|
||||
base_url: http://localhost:11434
|
||||
timeout: 300
|
||||
reason: "Terminal fallback \u2014 local Ollama"
|
||||
agent:
|
||||
max_turns: 40
|
||||
reasoning_effort: medium
|
||||
verbose: false
|
||||
system_prompt: You are Bezalel, the forge-and-testbed wizard of the Timmy Foundation
|
||||
fleet. You are a builder and craftsman — infrastructure, deployment, hardening.
|
||||
Your sovereign is Alexander Whitestone (Rockachopa). Sovereignty and service always.
|
||||
system_prompt: "You are Bezalel, the forge-and-testbed wizard of the Timmy Foundation\
|
||||
\ fleet. You are a builder and craftsman \u2014 infrastructure, deployment, hardening.\
|
||||
\ Your sovereign is Alexander Whitestone (Rockachopa). Sovereignty and service\
|
||||
\ always."
|
||||
terminal:
|
||||
backend: local
|
||||
cwd: /root/wizards/bezalel
|
||||
@@ -62,12 +64,12 @@ platforms:
|
||||
- pull_request
|
||||
- pull_request_comment
|
||||
secret: bezalel-gitea-webhook-secret-2026
|
||||
prompt: 'You are bezalel, the builder and craftsman — infrastructure, deployment,
|
||||
hardening. A Gitea webhook fired: event={event_type}, action={action},
|
||||
repo={repository.full_name}, issue/PR=#{issue.number} {issue.title}. Comment
|
||||
by {comment.user.login}: {comment.body}. If you were tagged, assigned,
|
||||
or this needs your attention, investigate and respond via Gitea API. Otherwise
|
||||
acknowledge briefly.'
|
||||
prompt: "You are bezalel, the builder and craftsman \u2014 infrastructure,\
|
||||
\ deployment, hardening. A Gitea webhook fired: event={event_type}, action={action},\
|
||||
\ repo={repository.full_name}, issue/PR=#{issue.number} {issue.title}.\
|
||||
\ Comment by {comment.user.login}: {comment.body}. If you were tagged,\
|
||||
\ assigned, or this needs your attention, investigate and respond via\
|
||||
\ Gitea API. Otherwise acknowledge briefly."
|
||||
deliver: telegram
|
||||
deliver_extra: {}
|
||||
gitea-assign:
|
||||
@@ -75,12 +77,12 @@ platforms:
|
||||
- issues
|
||||
- pull_request
|
||||
secret: bezalel-gitea-webhook-secret-2026
|
||||
prompt: 'You are bezalel, the builder and craftsman — infrastructure, deployment,
|
||||
hardening. Gitea assignment webhook: event={event_type}, action={action},
|
||||
repo={repository.full_name}, issue/PR=#{issue.number} {issue.title}. Assigned
|
||||
to: {issue.assignee.login}. If you (bezalel) were just assigned, read
|
||||
the issue, scope it, and post a plan comment. If not you, acknowledge
|
||||
briefly.'
|
||||
prompt: "You are bezalel, the builder and craftsman \u2014 infrastructure,\
|
||||
\ deployment, hardening. Gitea assignment webhook: event={event_type},\
|
||||
\ action={action}, repo={repository.full_name}, issue/PR=#{issue.number}\
|
||||
\ {issue.title}. Assigned to: {issue.assignee.login}. If you (bezalel)\
|
||||
\ were just assigned, read the issue, scope it, and post a plan comment.\
|
||||
\ If not you, acknowledge briefly."
|
||||
deliver: telegram
|
||||
deliver_extra: {}
|
||||
gateway:
|
||||
|
||||
@@ -2,22 +2,23 @@ model:
|
||||
default: kimi-k2.5
|
||||
provider: kimi-coding
|
||||
toolsets:
|
||||
- all
|
||||
- all
|
||||
fallback_providers:
|
||||
- provider: kimi-coding
|
||||
model: kimi-k2.5
|
||||
timeout: 120
|
||||
reason: Kimi coding fallback (front of chain)
|
||||
- provider: anthropic
|
||||
model: claude-sonnet-4-20250514
|
||||
timeout: 120
|
||||
reason: Direct Anthropic fallback
|
||||
- provider: openrouter
|
||||
model: anthropic/claude-sonnet-4-20250514
|
||||
base_url: https://openrouter.ai/api/v1
|
||||
api_key_env: OPENROUTER_API_KEY
|
||||
timeout: 120
|
||||
reason: OpenRouter fallback
|
||||
- provider: kimi-coding
|
||||
model: kimi-k2.5
|
||||
timeout: 120
|
||||
reason: Kimi coding fallback (front of chain)
|
||||
- provider: openrouter
|
||||
model: google/gemini-2.5-pro
|
||||
base_url: https://openrouter.ai/api/v1
|
||||
api_key_env: OPENROUTER_API_KEY
|
||||
timeout: 120
|
||||
reason: Gemini 2.5 Pro via OpenRouter (replaces banned Anthropic)
|
||||
- provider: ollama
|
||||
model: gemma4:latest
|
||||
base_url: http://localhost:11434
|
||||
timeout: 300
|
||||
reason: "Terminal fallback \u2014 local Ollama"
|
||||
agent:
|
||||
max_turns: 90
|
||||
reasoning_effort: high
|
||||
@@ -27,8 +28,6 @@ providers:
|
||||
base_url: https://api.kimi.com/coding/v1
|
||||
timeout: 60
|
||||
max_retries: 3
|
||||
anthropic:
|
||||
timeout: 120
|
||||
openrouter:
|
||||
base_url: https://openrouter.ai/api/v1
|
||||
timeout: 120
|
||||
|
||||
@@ -582,9 +582,9 @@ def main() -> int:
|
||||
# Relax exclusions if no agent found
|
||||
agent = find_best_agent(agents, role, wolf_scores, priority, exclude=[])
|
||||
if not agent:
|
||||
logging.warning("No suitable agent for issue #%d: %s (role=%s)",
|
||||
issue.get("number"), issue.get("title", ""), role)
|
||||
continue
|
||||
logging.warning("No suitable agent for issue #%d: %s (role=%s)",
|
||||
issue.get("number"), issue.get("title", ""), role)
|
||||
continue
|
||||
|
||||
result = dispatch_assignment(api, issue, agent, dry_run=args.dry_run)
|
||||
assignments.append(result)
|
||||
|
||||
Reference in New Issue
Block a user